Model Description

This is an EAGLE-3 draft model for Qwen3-235B-A22B-Instruct-2507, trained from scratch using LK losses — training objectives that directly target acceptance rate rather than using KL divergence as a proxy.

Training Details

Performance

Average acceptance length (τ) measured across MT-bench, HumanEval, and GSM8K with K = 7:

Configuration Temperature = 0 Temperature = 1
EAGLE-3 + KL 4.25 3.77
EAGLE-3 + LK (ours) 4.32 4.08

Comparison with Public Checkpoints

Model MT-bench (τ) HumanEval (τ) GSM8K (τ)
zhuyksir/EAGLE3-Qwen3-235B-A22B-Instruct-2507-FP8 2.64 4.03 4.43
RedHatAI/Qwen3-235B-A22B-Instruct-2507-speculator.eagle3 2.56 3.52 3.78
Ours 3.18 4.42 4.65

Measured at temperature = 1 with K = 7

Usage with vLLM

from vllm import LLM, SamplingParams

llm = LLM(
    model="Qwen/Qwen3-235B-A22B-Instruct-2507",
    speculative_config={
        "method": "eagle3",
        "model": "nebius/EAGLE3-Qwen3-235B-A22B-Instruct-2507",
        "num_speculative_tokens": 6,
    },
)

sampling_params = SamplingParams(temperature=0.7)
outputs = llm.generate(["Explain speculative decoding in simple terms."], sampling_params)

Note: The current vLLM implementation samples draft tokens greedily regardless of temperature settings, which can underestimate acceptance rates at temperature > 0. A community fix is under development (see vllm-project/vllm#20459). The acceptance metrics reported above were measured with proper rejection sampling.

License

CC BY 4.0

Citation

@misc{samarin2026lklosses,
  title     = {LK Losses: Direct Acceptance Rate Optimization for Speculative Decoding},
  author    = {Alexander Samarin and Sergei Krutikov and Anton Shevtsov and Sergei Skvortsov and Filipp Fisin and Alexander Golubev},
  year      = {2026},
  eprint    = {2602.23881},
  archivePrefix = {arXiv},
  primaryClass  = {cs.LG},
  url       = {https://arxiv.org/abs/2602.23881}
}
Downloads last month
76
Safetensors
Model size
0.6B params
Tensor type
I32
·
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for nebius/EAGLE3-Qwen3-235B-A22B-Instruct-2507

Finetuned
(15)
this model

Dataset used to train nebius/EAGLE3-Qwen3-235B-A22B-Instruct-2507

Collection including nebius/EAGLE3-Qwen3-235B-A22B-Instruct-2507

Paper for nebius/EAGLE3-Qwen3-235B-A22B-Instruct-2507

Evaluation results