Qwen3_8B_no_thinking_sft
Checkpoint from the RoleVerse project — a multilingual benchmark for social reasoning through same-universe role-playing across 5 languages (EN, ZH, JA, FR, ES).
This checkpoint was produced by the RoleVerse training pipeline (SFT and/or DPO). See the project repo and dataset for details on how it was trained.
- Dataset: https://huggingface.co/datasets/tinyrolls/RoleVerse
- Repo ID:
tinyrolls/Qwen3_8B_no_thinking_sft
Usage
from transformers import AutoModelForCausalLM, AutoTokenizer
model_id = "tinyrolls/Qwen3_8B_no_thinking_sft"
tok = AutoTokenizer.from_pretrained(model_id)
model = AutoModelForCausalLM.from_pretrained(model_id, torch_dtype="auto", device_map="auto")
- Downloads last month
- 17
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support