Parameter Golf v3: ~1.151 BPB in 16MB — Int6 QAT, BigramHash, SmearGate, SWA, zstd-22
-
Updated
Mar 29, 2026 - Jupyter Notebook
Parameter Golf v3: ~1.151 BPB in 16MB — Int6 QAT, BigramHash, SmearGate, SWA, zstd-22
OpenAI Parameter Golf — SOTA 0.3958 BPB (3-seed mean). Beat prior best (0.4416) by 0.0458. Causal BackoffNgramMixer. PR #1094.
OLYMPUS: The Agora — Community companion for OpenAI Parameter Golf. Live leaderboard, compliance engine, technique map, funding transparency, compute guide.
Exploratory compact LM research branch derived from LUMI-Arch, focused on structure-sensitive generalization under tight model constraints.
Can a language model learn algorithms instead of memorizing patterns? Exploring extreme parameter efficiency in vocabulary space — 350K params, no embedding, no output projection, every intermediate state readable as words.
Sparse I/O bottleneck language model using C. elegans connectome wiring + JEPA + SigREG. Parameter Golf submission: 1.79 BPB in a 2MB artifact.
Byte-level adaptation-native 16MB language model for OpenAI Parameter Golf
A practical starter repo and working setup for the OpenAI Parameter Golf competition
Add a description, image, and links to the parameter-golf topic page so that developers can more easily learn about it.
To associate your repository with the parameter-golf topic, visit your repo's landing page and select "manage topics."