8000 GitHub - liwei-group/CCD
[go: up one dir, main page]
More Web Proxy on the site http://driver.im/
Skip to content

liwei-group/CCD

Repository files navigation

What Role Does Data Augmentation Play in Knowledge Distillation ?

  • This project provides source code for our Joint Data Augmentation and Cosine Confidence Distillation (JDA and CCD).
  • This paper is publicly available at the official ACCV proceedings.
    • 🏆 __SOTA of Knowledge Distillation for student ResNet-18 trained by teacher ResNet-34 on ImageNet.

Installation

Requirements

  • pytorch
  • numpy
  • Ubuntu 20.04 LTS
  1. please install create environment first.
conda create -n torch python==3.7.10
  1. then activate this environment.
source activate torch
  1. then install pytorch
conda install pytorch torchvision torchaudio cudatoolkit=11.3 -c pytorch

Perform Offline KD experiments on CIFAR-100 dataset

Dataset

CIFAR-100 : download

  • download to your folder, named [cifar100_folder]

Teacher checkpoint

CKPT: download

  • download to your floder, name [ckpt_folder]

Train KD+JDA and CCD+JDA

  1. Modify the paths of the teacher pre-training files and datasets in these two files : multirunc100_jda.sh, multirunc100_jda_ccd.sh

  2. run them

bash multirunc100_jda.sh
bash multirunc100_jda_ccd.sh

Perform Offline KD experiments on ImageNet dataset

Dataset preparation

  • Download the ImageNet dataset to YOUR_IMAGENET_PATH and move validation images to labeled subfolders

  • Create a datasets subfolder and a symlink to the ImageNet dataset

Folder of ImageNet Dataset:

data/ImageNet
├── train
├── val

Teacher checkpoint

The pre-trained backbone weights of ResNet-34 follow the resnet34-333f7ec4.pth downloaded from the official PyTorch.

Train KD+JDA and CCD+JDA

  1. Modify the paths of the teacher pre-training files and datasets in this file : train_student_imagenet.py

  2. run this py format file

python train_student_imagenet.py
python train_student_imagenet.py --ccd

Code References

Our code is based on torchdistill (e.g., SPKD, KD, SSKD, CRD) and HSAKD (e.g., the training framework).

Contact us

email : shaoshitong@seu.edu.cn name : shaoshitong

Should know

  1. The code is only provided for academic research. Any commercial use of this code is prohibited.
  2. The use of our code requires references to this paper.

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Contributors 3

  •  
  •  
  •  
0