ARTFEED — Contemporary Art Intelligence

Generative Diffusion Prior Distillation for Time-Series Classification

other · 2026-05-13

A new arXiv paper (2605.11414) introduces Generative Diffusion Prior Distillation (GDPD) to address the challenge of classifying time-series data when only partial prefixes are available due to latency and cost constraints. Traditional classifiers assume full sequences at inference, but partial data often lacks class-discriminative patterns, hindering generalization. The authors propose using knowledge distillation (KD) from a full-sequence teacher model to a partial-sequence student model. However, direct feature matching fails when the generalization gap stems from data differences (full vs. partial) rather than parameter capacity. GDPD employs a generative diffusion prior to provide progressive, diverse, and collective teacher supervision, enabling the student to learn from the teacher's full-context features more effectively. The method is validated on multiple time-series benchmarks, showing improved classification accuracy on partial inputs.

Key facts

  • Paper ID: arXiv:2605.11414
  • Published on arXiv
  • Addresses time-series classification with partial prefixes
  • Uses knowledge distillation from full-sequence teacher to partial-sequence student
  • Proposes Generative Diffusion Prior Distillation (GDPD)
  • GDPD provides progressive, diverse, and collective teacher supervision
  • Validated on multiple time-series benchmarks
  • Improves classification accuracy on partial inputs

Entities

Institutions

  • arXiv

Sources