Adaptive Weighting in Knowledge Distillation: An Axiomatic Framework for Multi-Scale Teacher Ensemble Optimization Paper • 2601.17910 • Published Jan 25
Recursive Meta-Distillation: An Axiomatic Framework for Iterative Knowledge Refinement Paper • 2601.13100 • Published Jan 19
Multi-Teacher Ensemble Distillation: A Mathematical Framework for Probability-Domain Knowledge Aggregation Paper • 2601.09165 • Published Jan 14
Sparse Knowledge Distillation: A Mathematical Framework for Probability-Domain Temperature Scaling and Multi-Stage Compression Paper • 2601.03195 • Published Jan 6
Post-Training Probability Manifold Correction via Structured SVD Pruning and Self-Referential Distillation Paper • 2602.00372 • Published 30 days ago