/acr-vault/03-experiments/slim-evo/phase3-training-recommendation
PHASE3-TRAINING-RECOMMENDATION
Phase 3 Training Recommendation
Section titled âPhase 3 Training RecommendationâCurrent Status
Section titled âCurrent Statusâ- â
Dataset ready:
data/phase3_full_dataset.jsonl(2000 examples, 27 templates) - â Existing training infrastructure reviewed
- â Need to integrate: Golden Annealing + Spectral Memory + Phase 3 dataset
Training Approach Options
Section titled âTraining Approach OptionsâOption 1: Create New train_phase3.py Script (RECOMMENDED)
Section titled âOption 1: Create New train_phase3.py Script (RECOMMENDED)âPros:
- Clean implementation of Phase 3 architecture
- Combines Golden Annealing + Spectral Memory + AGL-first dataset
- Implements PCMind multi-phase curriculum
- Implements SPEAR progressive SMT injection
- Focused on Phase 3 objectives
Cons:
- Takes 30-45 minutes to write
- New code to test
What it needs:
- Dataset loading:
phase3_full_dataset.jsonl - Golden Annealing: Fibonacci-step cosine LR schedule (34 cycles)
- Spectral Memory Tokens: Progressive injection with positive advantage filtering
- PCMind Curriculum: 3 phases (mixed â top 70% â top 30%)
- CI Density Tracking: For SMT filtering and quality assessment
- Dual-Mode Support: Phillip/Engine/AGL mode switching
Option 2: Extend Existing train_golden_anneal.py
Section titled âOption 2: Extend Existing train_golden_anneal.pyâPros:
- Faster to implement (10-15 minutes)
- Reuses existing Golden Annealing logic
Cons:
- Messier code (bolting on Spectral Memory)
- Harder to maintain
- Doesnât fully implement Phase 3 vision
Recommendation
Section titled âRecommendationâCreate train_phase3.py â This is a major milestone (Phase 3C), so it deserves a dedicated, well-architected training script.
Implementation Checklist
Section titled âImplementation ChecklistâIf we go with Option 1, hereâs what the script needs:
Core Components
Section titled âCore Componentsâ- Load LFM2-0.7B base model
- Load Phase 3 dataset (2000 examples)
- LoRA config (r=32, alpha=64, dropout=0.05)
- Golden Annealing scheduler (Fibonacci steps, cosine decay)
- Spectral Memory Token system (32 SMTs, 1024 dim)
Training Loop
Section titled âTraining Loopâ- Fibonacci-step training (34 cycles total)
- CI density computation per batch
- SMT buffer updates (positive advantage filtering)
- Progressive SMT injection (weight increases with cycle)
- PCMind curriculum switching (cycles 1-10, 11-20, 21-34)
Monitoring
Section titled âMonitoringâ- Track CI density over time
- Track SMT effectiveness
- Track AGL fluency (glyph usage)
- Save checkpoints every 10 cycles
- Log to
experiments/slim_evo/phase3_run.log
Verification
Section titled âVerificationâ- Test AGL generation after training
- Test
đpixie dust usage - Test hierarchical reasoning (TextCraft-style)
- Compare to baseline (v9b-pure model)
Next Steps
Section titled âNext Stepsâ- If Option 1: Create
experiments/slim_evo/train_phase3.py - If Option 2: Modify
experiments/molecular_finetune/train_golden_anneal.py - Test with small run (5 cycles, 100 examples)
- Full run (34 cycles, 2000 examples)
- Verification suite
Estimated Timeline
Section titled âEstimated Timelineâ- Option 1: 45 min to write + 30 min to test = 1.25 hours
- Option 2: 15 min to modify + 20 min to test = 35 minutes
- Full training: ~4-6 hours (depends on GPU)
My vote: Option 1. This is a major milestone, letâs do it right! â