If you're curious about how knowledge distillation can boost model performance using real and generated data, you're in the right place — this project is made for you.
We are a research team at UMass Boston. We are Can Do Crew Team. We are exploring how knowledge distillation helps student models perform better when trained on a mix of real images and AI-generated images. Our study focuses on 10 object classes and evaluates how training size, data sources, and augmentation impact model performance.
Our goal is to investigate how model distillation can improve learning efficiency for image classification. In simple words, we train a student model to mimic the behavior of a powerful teacher model, hoping that the student can learn faster and perform better—even with less data.
This main idea is important when we want to:
We built a teacher model trained to classify 10 visual classes using CIFAR datasets. This model serves as the ground truth for guiding all student models. Please note: While we use the CIFAR dataset to train the teacher model, we intentionally avoid using CIFAR for the student models to prevent bias and ensure a more realistic evaluation.
We Asked Ourselves:
👉 How well can student models learn from the teacher under different dataset conditions?
High-quality, real-world images across 10 categories.
Web-crawled images with diverse styles and content.
AI-generated images based on creative prompts across 10 categories.
800 Google train + 200 Real test
Scratch: 15.61% | KD: 40%
800 Bing train + 200 Real test
Scratch: 17% | KD: 26.5%
800 Web train + 200 Real test
Scratch: 16.5% | KD: 38%
400 Web + 400 Augmented (Color) + 200 Real test
Scratch: 14.07% | KD: 61.5%
200 Google + 200 Bing + 400 Augmented + 200 test
Scratch: 10.0% | KD: 35.4%
800 Generated train + 200 Real test
Scratch: 15.5% | KD: 14.0%
800 Generated train + 200 Real test
Scratch: 10.85% | KD: 31.53%
2000 Google + 2000 Bing train
Test: 500 Google + 500 Bing
Scratch: 39.1% | KD: 38.1%
1000 Google + 1000 Bing + 1000 V2 + 1000 Diffusers
Test: 500 Google + 500 Bing
Scratch: 30.8% | KD: 36.19%
8000 Web train + 2000 Real test
Scratch: 45.85% | KD: 70.6%
8000 Web train + 2000 Real test
Scratch: 36.4% | KD: 36.9%
8000 Generated train + 2000 Real test
Scratch: 24.25% | KD: 21.6%
8000 Generated train + 2000 Real test
Scratch: 31.5% | KD: 32.9%