Generative Diffusion Prior Distillation for Time-Series Classification
A new arXiv paper (2605.11414) introduces Generative Diffusion Prior Distillation (GDPD) to address the challenge of classifying time-series data when only partial prefixes are available due to latency and cost constraints. Traditional classifiers assume full sequences at inference, but partial data often lacks class-discriminative patterns, hindering generalization. The authors propose using knowledge distillation (KD) from a full-sequence teacher model to a partial-sequence student model. However, direct feature matching fails when the generalization gap stems from data differences (full vs. partial) rather than parameter capacity. GDPD employs a generative diffusion prior to provide progressive, diverse, and collective teacher supervision, enabling the student to learn from the teacher's full-context features more effectively. The method is validated on multiple time-series benchmarks, showing improved classification accuracy on partial inputs.
Key facts
- Paper ID: arXiv:2605.11414
- Published on arXiv
- Addresses time-series classification with partial prefixes
- Uses knowledge distillation from full-sequence teacher to partial-sequence student
- Proposes Generative Diffusion Prior Distillation (GDPD)
- GDPD provides progressive, diverse, and collective teacher supervision
- Validated on multiple time-series benchmarks
- Improves classification accuracy on partial inputs
Entities
Institutions
- arXiv