
OLMo-2-0425-1B
Text
OLMo 2 1B (≈1 B params, Apache-2.0)
Open-science transformer you can run on a laptop.
Spec sheet. 16-layer decoder-only stack, 2 048 hidden size, 16 attention heads, 4 096-token window; pre-trained on 4 T tokens of mixed web/code text.
Punches above its class. OLMo 2 1B posts an average benchmark score of 42.7, edging out peer 1 B models like Gemma-3 and Llama-3.1 on MMLU, GSM-8K, and other tasks.
Runs light. FP16 weights need ≈ 6 GB VRAM; 8-bit ≈ 3 GB; 4-bit ≈ 1 GB—laptops or single A10G are fine.
ab-notebook release. AI2 ships the full training logs, intermediate checkpoints, and code so you can audit, rewind, or fine-tune without guesswork.
Why pick it for Norman AI?
Apache license, tiny VRAM needs, and full transparency make OLMo 2 1B a perfect sandbox: spin up edge demos, run interpretability experiments, or ship a “budget” inference tier without changing our Llama-friendly toolchain.
