{ "author": "sam", "github_id": "samacqua", "name": "LoRA TTT", "blurb": "Naive baseline + per-document LoRA test-time training at eval. Rank-8 LoRA on lm_head/Q/V with Adam lr=0.01, overlapping 256-token chunks in 1024-token context windows. Same training, smarter eval.", "date": "2026-03-19T10:00:00Z", "val_loss": 2.0142, "val_bpb": 1.1929, "bytes_total": 15882446, "bytes_code": 58509 }