Beyond Length Scaling: Synergizing Breadth and Depth for Generative Reward Models Paper • 2603.01571 • Published 9 days ago • 33
Olmo 3.1 Collection The latest members of the Olmo 3 family: another 3 weeks of RL for 32B Think, the 32B Instruct model, large post-training research datasets... • 9 items • Updated Dec 23, 2025 • 48
view article Article Transformers v5: Simple model definitions powering the AI ecosystem +2 Dec 1, 2025 • 305
GLM-4.5 Collection GLM-4.5: An open-source large language model designed for intelligent agents by Z.ai • 8 items • Updated 9 days ago • 252
Falcon Edge series Collection A series of powerful, universal and fine-tunable small Language Models • 7 items • Updated Nov 6, 2025 • 25
Cohere Labs Aya Vision Collection Aya Vision is a state-of-the-art family of vision models that brings multimodal capabilities to 23 languages. • 5 items • Updated Jul 31, 2025 • 72
Tulu 3 Models Collection All models released with Tulu 3 -- state of the art open post-training recipes. • 11 items • Updated Dec 23, 2025 • 104
Qwen2.5-1M Collection The long-context version of Qwen2.5, supporting 1M-token context lengths • 3 items • Updated Dec 31, 2025 • 126
Qwen2.5-VL Collection Vision-language model series based on Qwen2.5 • 10 items • Updated 9 days ago • 557