Distilling Convolutional Neural Network Performance from Multiple Data Sources

✨ CAN DO CREW ✨
Rami Huu Nguyen · Avanith Kanamarlapudi · Lakshmi Pranathi Vutla · Shauna Murray
Team Photo

If you're curious about how knowledge distillation can boost model performance using real and generated data, you're in the right place — this project is made for you.

We are a research team at UMass Boston. We are Can Do Crew Team. We are exploring how knowledge distillation helps student models perform better when trained on a mix of real images and AI-generated images. Our study focuses on 10 object classes and evaluates how training size, data sources, and augmentation impact model performance.

🟦 What’s This Project About?

Our goal is to investigate how model distillation can improve learning efficiency for image classification. In simple words, we train a student model to mimic the behavior of a powerful teacher model, hoping that the student can learn faster and perform better—even with less data.

This main idea is important when we want to:

  • Reduce model size for deployment
  • Train on limited or mixed-quality datasets
  • Save computation but still keep high accuracy

🟦 Before We Started

We built a teacher model trained to classify 10 visual classes using CIFAR datasets. This model serves as the ground truth for guiding all student models. Please note: While we use the CIFAR dataset to train the teacher model, we intentionally avoid using CIFAR for the student models to prevent bias and ensure a more realistic evaluation.

We Asked Ourselves:

👉 How well can student models learn from the teacher under different dataset conditions?

Explore Project ↓

Dataset Sources

Google Images

High-quality, real-world images across 10 categories.

Google Bird Google Dog Google Cat

Bing Images

Web-crawled images with diverse styles and content.

Bing 1 Bing 2 Bing 3

Stable Diffusion (V2 & Diffusers)

AI-generated images based on creative prompts across 10 categories.

SD V2 1 SD Diffuser 1 SD Diffuser 2

Search Queries & Prompts

Google/Bing Keywords

  • frog
  • horse
  • truck
  • cat
  • bird
  • dog
  • airplane
  • deer
  • ship
  • automobile

Stable Diffusion Prompts

  • 🐕 golden retriever
  • 🚗 sports car
  • 🐸 frog on branch
  • 🦌 snowy deer
  • ✈ sunset airplane
  • 🚙 classic car
  • ⛵ pirate ship
  • 🐎 wild horse
  • 🦜 parrot on branch
  • 🐶 police dog
  • 🌲 misty forest deer
  • 🚛 racing truck
  • 🛫 night cockpit
  • 😺 sunlit cat
  • 🐦 evening bird

Our Experiments

Google

800 Google train + 200 Real test

Scratch: 15.61% | KD: 40%

Bing

800 Bing train + 200 Real test

Scratch: 17% | KD: 26.5%

Google + Bing

800 Web train + 200 Real test

Scratch: 16.5% | KD: 38%

Google + Bing + Color Aug

400 Web + 400 Augmented (Color) + 200 Real test

Scratch: 14.07% | KD: 61.5%

Google + Bing + Rotation Aug

200 Google + 200 Bing + 400 Augmented + 200 test

Scratch: 10.0% | KD: 35.4%

Stable Diffusion V2

800 Generated train + 200 Real test

Scratch: 15.5% | KD: 14.0%

Stable Diffusion 3 Medium

800 Generated train + 200 Real test

Scratch: 10.85% | KD: 31.53%

Google + Bing

2000 Google + 2000 Bing train
Test: 500 Google + 500 Bing

Scratch: 39.1% | KD: 38.1%

Google + Bing + Stable Diffusion (V2 + Diffusers)

1000 Google + 1000 Bing + 1000 V2 + 1000 Diffusers
Test: 500 Google + 500 Bing

Scratch: 30.8% | KD: 36.19%

Google Dataset

8000 Web train + 2000 Real test

Scratch: 45.85% | KD: 70.6%

Bing Dataset

8000 Web train + 2000 Real test

Scratch: 36.4% | KD: 36.9%

Stable Diffusion V2

8000 Generated train + 2000 Real test

Scratch: 24.25% | KD: 21.6%

Stable Diffusion 3 Medium

8000 Generated train + 2000 Real test

Scratch: 31.5% | KD: 32.9%

Results & Key Findings

Results Chart

View full chart

Key Insights

Limitations & Future Work

Limitations

  • Focused only on 10 classes; real-world cybersecurity includes more.
  • Evaluated only Google, Bing, and Stable Diffusion sources.
  • Stable Diffusion V2 sometimes generates low-quality images.

Future Work

  • Include additional categories from Firefox, Kaggle, and beyond.
  • Try more fine-tuned generative models.
  • Expand to real-world cybersecurity applications.