OLMo 3
model7B and 32B dense Transformers with 65K native context (up from 4K). 3-stage training: main pretraining (~95%), mid-training (100-200B tokens of code/math/QA/thinking), and long-context extension. Trained on Dolma 3 (~9.3T token pool including olmOCR-processed science PDFs). 5.5-5.9T tokens used.
32B Think: MATH 96.1, AIME 2024 76.8, AIME 2025 72.5, HumanEvalPlus 91.4, MMLU 85.4. OLMo 3.1 (Dec 2025) extends RL training (+3 weeks RLVR): AIME 2025 78.1, IFEval 93.8. AA Intelligence: 14 (3.1 Think). Strongest fully open thinking model at release. Apache 2.0.
Paper (arXiv)HuggingFace (32B Think)HuggingFace (3.1 Think)Artificial Analysis (3.1 Think)OpenRouter (3.1 Think)
Model Details
Architecture DENSE
Parameters 32B
Context window 65,536
Variants
| Name | Parameters | Notes |
|---|---|---|
| OLMo 3 7B | 7B | — |
| OLMo 3 32B | 32B | — |
Paper
arXiv: 2512.13961