Ecosyste.ms: Awesome
An open API service indexing awesome lists of open source software.
https://github.com/bhheo/ab_distillation
Knowledge Transfer via Distillation of Activation Boundaries Formed by Hidden Neurons (AAAI 2019)
https://github.com/bhheo/ab_distillation
knowledge-distillation knowledge-transfer network-compression teacher-student-learning transfer-learning
Last synced: 3 days ago
JSON representation
Knowledge Transfer via Distillation of Activation Boundaries Formed by Hidden Neurons (AAAI 2019)
- Host: GitHub
- URL: https://github.com/bhheo/ab_distillation
- Owner: bhheo
- License: mit
- Created: 2018-11-21T05:21:58.000Z (almost 6 years ago)
- Default Branch: master
- Last Pushed: 2019-09-09T11:04:03.000Z (about 5 years ago)
- Last Synced: 2024-05-22T19:35:56.123Z (6 months ago)
- Topics: knowledge-distillation, knowledge-transfer, network-compression, teacher-student-learning, transfer-learning
- Language: Python
- Homepage:
- Size: 15 MB
- Stars: 103
- Watchers: 5
- Forks: 18
- Open Issues: 0
-
Metadata Files:
- Readme: README.md
- License: LICENSE
Awesome Lists containing this project
README
# Knowledge Transfer via Distillation of Activation Boundaries Formed by Hidden Neurons
Official Pytorch implementation of paper:
[Knowledge Transfer via Distillation of Activation Boundaries Formed by Hidden Neurons](https://arxiv.org/abs/1811.03233) (AAAI 2019).
Slides and poster are available on [homepage](https://sites.google.com/view/byeongho-heo/home)
## Environment
Python 3.6, Pytorch 0.4.1, Torchvision## Knowledge distillation [(CIFAR-10)](https://www.cs.toronto.edu/~kriz/cifar.html)
cifar10_AB_distillation.py
\
Distillation from WRN 22-4 (teacher) to WRN 16-2 (student) on CIFAR-10 dataset.Pre-trained teacher network (WRN 22-4) is included. Just run the code.
## Transfer learning [(MIT_scenes)](http://web.mit.edu/torralba/www/indoor.html)
MITscenes_AB_distillation.py
\
Transfer learning from ImageNet pre-trained model (teacher) to randomly initialized model (student).Teacher : ImageNet pre-trained ResNet 50
Student : MobileNet or MobileNetV2 (randomly initialized model)
Please change base learning rate to 0.1 for MobileNetV2.
\
MIT_scenes dataset should be arranged for Torchvision ImageFolder function.Train set :
`$dataset_path / train / $class_name / $image_name `Test set :
`$dataset_path / test / $class_name / $image name`and run with dataset path.
MobileNet
```
python MITscenes_AB_distillation.py --data_root $dataset_path
```MobileNet V2
```
python MITscenes_AB_distillation.py --data_root $dataset_path --network mobilenetV2
```
## Other implementations
Tensorflow: https://github.com/sseung0703/Knowledge_distillation_methods_wtih_Tensorflow## Citation
```
@inproceedings{ABdistill,
title = {Knowledge Transfer via Distillation of Activation Boundaries Formed by Hidden Neurons},
author = {Byeongho Heo, Minsik Lee, Sangdoo Yun, Jin Young Choi},
booktitle = {AAAI Conference on Artificial Intelligence (AAAI)},
year = {2019}
}
```