| Age | Commit message (Collapse) | Author | |
|---|---|---|---|
| 2026-04-10 | Fix two bugs: PEFT cleanup model corruption and K=16 OOM | YurenHao0426 | |
| Bug 1: PEFTBaseline.cleanup() corrupted wrapper.model after LoRA unload, causing 'Qwen2Model has no attribute prepare_inputs_for_generation' for subsequent methods. Fix: save reference to original model before wrapping, restore it directly in cleanup() instead of relying on unload(). Bug 2: fit_theta OOM at K=16 due to large logit chunks (128 × 151936 vocab). Fix: reduce CHUNK_SIZE from 128 to 32 (~4x less memory per chunk). Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com> | |||
| 2026-04-03 | Initial commit: UPH project codebase and experiment results | YurenHao0426 | |
| Includes model code, evaluation scripts, configs, analysis outputs, and experiment results for the User Prior Head personalization method. Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com> | |||
