Record: 1.1140 BPB — ResidLambdas + Split-LR + Train-Budget GPTQ + Coprime Loader (12-seed mean)#1130
Record: 1.1140 BPB — ResidLambdas + Split-LR + Train-Budget GPTQ + Coprime Loader (12-seed mean)#1130Gusanidas wants to merge 2 commits intoopenai:mainfrom
Conversation
PR openai#549 / KitchenSinkV2 base with: - Residual lambdas: learnable per-sublayer scaling (init sqrt(1.1), 5x LR) - Bigram hash: 6144 buckets (up from 2048) - Value embeddings: dim=196 on layers 5,9,10 - Flash Attention 3 via flash_attn_interface - Train-data GPTQ int6 calibration within training budget - Sliding window eval stride=64 - Optuna-tuned LRs: matrix 0.036/0.044, scalar 0.028/0.018 12 seeds: mean 1.1140 bpb (1.8809 nats), std 0.0005 Improvement over leader: 0.0054 bpb / 0.0091 nats p < 0.0001 for >= 0.005 nats improvement Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
val_bpb: 1.1161 | val_loss: 1.884 nats | ~15.3 MB | 8×H100 SXM | Legal TTT Seeds: 42=1.1163, 1337=1.1160, 2024=1.1161 | Mean=1.1161, Std=0.0001 Novel contribution: EGGROLL Antithetic Ternary Bin Search — post-GPTQ quantization refinement that directly optimizes INT6 bin assignments against BPB loss during eval. Zeroth-order, strictly additive (cannot degrade quality), complementary to Hessian-based GPTQ. Also adds missing TTT call to PR openai#1130's eval pipeline. Built on PR openai#1130 by @Gusanidas (Kitchen Sink V2) Foundation: PR openai#549 by @abaybektursun Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
… BPB) ResidLambdas: per-sublayer residual scaling (init sqrt(1.1), 5x scalar_lr, no WD) Tuned LRs: MATRIX_LR=0.036, SCALAR_LR=0.028, TIED_EMBED_LR=0.022 Bigger VE: dim=196 on layers 5,9,10 (was dim=128 on layers 9,10) PR openai#1130 achieved 1.1140 (12-seed mean) with these innovations.
val_bpb: 1.1161 | val_loss: 1.884 nats | ~15.3 MB | 8×H100 SXM | Legal TTT Seeds: 42=1.1163, 1337=1.1160, 2024=1.1161 | Mean=1.1161, Std=0.0001 Novel: EGGROLL Antithetic Ternary Bin Search — post-GPTQ bin refinement Also: adds missing TTT call to PR openai#1130 eval pipeline Built on PR openai#1130 by @Gusanidas, PR openai#549 by @abaybektursun Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
val_bpb: 1.1161 | val_loss: 1.884 nats | ~15.3 MB | 8×H100 SXM | Legal TTT Seeds: 42=1.1163, 1337=1.1160, 2024=1.1161 | Mean=1.1161, Std=0.0001 Novel: EGGROLL Antithetic Ternary Bin Search — post-GPTQ bin refinement Also: adds missing TTT call to PR openai#1130 eval pipeline Built on PR openai#1130 by @Gusanidas, PR openai#549 by @abaybektursun Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
Community Review — Record: 1.1140 BPB — ResidLambdas + Split-LR + Train-Budget GPTQ + Coprime Loader (12-seed mean)BPB: 1.1140 | Compliance: LOOKS CLEAN — score-first-per-chunk TTT (legal #1416/#1423 pattern) What I found in the code (head SHA The TTT path at line 1311 implements the score-first-per-chunk pattern: each chunk is scored under Per Issue #402 and Issue #677, TTT is legal when each token is scored before the adapter updates on it, and that's what the code does here — chunk CPU smoke test (CT2038 proteus-engine, 2026-04-11): import OK in 0.11s, dim=512, layers=11, vocab=1024, code=126292 B, SMOKE_TEST_PASS Verdict: LOOKS CLEAN. Recommendation to @cocohearts @valerio-oai @0hq @yuzhougu-oai @notapplica: MERGE pending standard checks (3-seed validation, 16MB artifact cap, 10-min wallclock on 8×H100 SXM). The compliance picture matches the legal reference frontier and no flags were raised by the classification pass. Auto-classification caveat: this review was drafted by the AST-based classifier against a template derived from manually-reviewed cluster PRs (#1420, #1450, #1487, #1541, #1529, #1533, #1518). If I've misread a subtlety in your eval path — e.g., multi-epoch TTT that I mistook for single-pass, or a target-in-key lookup I missed in a helper function — please flag it and I'll re-run the audit manually. Reviewed by @MatoTeziTanka — The Agora. CPU smoke test (CT2038 proteus-engine, 2026-04-11): import OK in 0.11s, dim=512, layers=11, vocab=1024, code=126292 B, SMOKE_TEST_PASS. Classification via deterministic AST-based |
Record: Kitchen Sink V2 — val_bpb 1.1140 (12-seed mean, std 0.0005)
val_bpb: 1.1140 | val_loss: 1.8809 nats | ~15.88 MB | 8×H100 SXM | No TTT
Built on PR #549 by @abaybektursun. 12-seed validation, all artifacts under 16,000,000 bytes, all training under 600s.
Results (12 seeds, sliding window eval, stride=64)
Statistical significance vs SOTA (PR #549, 1.8843 nats)
What's new (over PR #549)
12 Tuned batch size — TRAIN_BATCH_TOKENS=548,864
Architecture
Timing
Credits