Qwen3_8B_no_thinking_sft

Checkpoint from the RoleVerse project — a multilingual benchmark for social reasoning through same-universe role-playing across 5 languages (EN, ZH, JA, FR, ES).

This checkpoint was produced by the RoleVerse training pipeline (SFT and/or DPO). See the project repo and dataset for details on how it was trained.

Usage

from transformers import AutoModelForCausalLM, AutoTokenizer

model_id = "tinyrolls/Qwen3_8B_no_thinking_sft"
tok = AutoTokenizer.from_pretrained(model_id)
model = AutoModelForCausalLM.from_pretrained(model_id, torch_dtype="auto", device_map="auto")
Downloads last month
17
Safetensors
Model size
8B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Dataset used to train tinyrolls/Qwen3_8B_no_thinking_sft