Ecosyste.ms: Awesome

An open API service indexing awesome lists of open source software.

Awesome Lists | Featured Topics | Projects

https://github.com/kakaobrain/autoclint

A specially designed light version of Fast AutoAugment
https://github.com/kakaobrain/autoclint

1st-place autocv-challenge autocv2-challenge autodl-challenge automated-machine-learning automl cnn computer-vision convolutional-neural-networks deep-learning fast-auto-augment image-classification pytorch

Last synced: 3 days ago
JSON representation

A specially designed light version of Fast AutoAugment

Awesome Lists containing this project

README

        

# AutoCLINT
[![KakaoBrain](https://img.shields.io/badge/kakao-brain-ffcd00.svg)](http://kakaobrain.com/)
[![pytorch](https://img.shields.io/badge/pytorch-1.0.1-%23ee4c2c.svg)](https://pytorch.org/)
[![tensorflow](https://img.shields.io/badge/tensorflow-1.13.1-ed6c20.svg)](https://www.tensorflow.org/)
[![autocv 1st place](https://img.shields.io/badge/autocv-1st_place-%235339D3.svg)](#)
[![autocv2 1st place](https://img.shields.io/badge/autocv2-1st_place-%235339D3.svg)](#)
[![HitCount](http://hits.dwyl.io/kakaobrain/autoclint.svg)](http://hits.dwyl.io/kakaobrain/autoclint)

## Automatic **C**omputationally **LI**ght **N**etwork **T**ransfer

A specially designed light version of **[Fast AutoAugment][]** is implemented to adapt to **various tasks** under **limited resources**.

This is our solution to [NeurIPS 2019 AutoDL Challenges][].
We won the **1st place** in the final learderboards in both [AutoCV][] and [AutoCV2][] Challenges.

## [AutoCV][]/[AutoCV2][] Challenge Introduction
#### Fully Automated Image (and Video) Classification without ANY human intervention
> Despite recent successes of deep learning and other machine learning techniques, practical experience and expertise is still required to select models and/or choose hyper-parameters when applying techniques to new datasets. This problem is drawing githincreasing interest, yielding progress towards fully automated solutions. In this challenge your machine learning code is trained and tested on this platform, without human intervention whatsoever, on image or video classification tasks you have never seen before, with time and memory limitations. All problems are multi-label classification problems, coming from various domains including medical imaging, satellite imaging, object recognition, character recognition, face recognition, etc. They lend themselves to deep learning solutions, but other methods may be used. Raw data is provided, but formatted in a uniform manner, to encourage you to submit generic algorithms.

## Methods

We employ a network transfer strategy and implement a light version of [Fast AutoAugment][] for the fast adaptation and the efficient search of data augmentation policies.

### Network Transfer

The [AutoCV][] Challenges are given limited memory and computational resources. Thus, we considered a small size of architecture that could use the pre-trained models that were transferred.

We have discovered the optimal hyperparameters and architectures to get the best performance in five minutes from five public data sets (Munster, Chuckey, Pedro, Decal and Hammer). In this process, no data augmentation is used.

Due to the variability in image size (median shape 28x28x1 for munster vs. 576x944x3 for decal) the input tensor size of network must be automatically adapted for each dataset to allow for adequate aggregation of spatial information and to keep the aspect ratio of original image.
We automatically adapt these parameters to the median size of each dataset, so that the network effectively trains on entire datasets. Due to time constraints, we do not increase the input tensor volume (without channels) beyond 64^2. If the median shape of the dataset is smaller than 64^2 then we use the median shape as original input.

| | Munster | Chucky | Pedro | Decal | Hammer | Kreatur | Katze | Kraut |
|:---|:-------:|:-------:|:-------:|:-------:|:-------:|:-------:|:-----:|:-----:|
| data type | image | image | image | image | image | video | video | video |
| original median shape | 28x28x1 | 32x32x1 | 203x74x3 | 576x944x3 | 300x400x3 | 46x60x80x3 | 46x120x160x1 | 46x120x160x1 |
| input tensor shape | 28x28x1 | 32x32x1 | 128x48x3 | 48x64x3 | 48x64x3 | 8x48x64x3 | 8x48x64x1 | 8x48x64x1 |

### [Fast AutoAugment][]

Fast AutoAugment learns augmentation policies using a more efficient search strategy based on density matching. Ideally, Fast AutoAugment should be performed automatically, allowing the training data to adapt to test data.

We modify the search space and implement a light version of Fast AutoAugment algorithm to surmount the restricted computational resources.

As Fast AutoAugment, we search the augmentation policies that match the density of train data with density of augmented valid data. We deviate from the original version in that we replace 5-fold with single-fold search and use random search (within subset of searched policy in original) instead of TPE algorithm.

## [AutoCV][] Results
### Public

#### V1.XLARGE
* experiment environment: [BrainCloud][] V1.XLARGE Type (NVIDIA Tesla V100 1GPU, 14CPU, 122GB)

| metrics | Munster | Chucky | Pedro | Decal | Hammer |
|:-------:|--------:|--------:|--------:|--------:|--------:|
| ALC | 0.9421 | 0.8082 | 0.7948 | 0.8647 | 0.8147 |
| 2*AUC-1 | 0.9992 | 0.9297 | 0.9241 | 0.9233 | 0.8863 |
| curves | ![](./assets/public_final_result_v1_munster.png) | ![](./assets/public_final_result_v1_Chuckey.png) | ![](./assets/public_final_result_v1_pedro.png) | ![](./assets/public_final_result_v1_Decal.png) | ![](./assets/public_final_result_v1_Hammer.png) |

#### P1.XLARGE
* experiment environment: [BrainCloud][] P1.XLARGE Type (NVIDIA Tesla P40 1GPU, 6CPU, 61GB)

| metrics | Munster | Chucky | Pedro | Decal | Hammer |
|:-------:|--------:|--------:|--------:|--------:|--------:|
| ALC | 0.9440 | 0.7835 | 0.7366 | 0.8353 | 0.8286 |
| 2*AUC-1 | 0.9977 | 0.9353 | 0.9214 | 0.9347 | 0.9142 |
| curves | ![](./assets/public_final_result_p1_munster.png) | ![](./assets/public_final_result_p1_Chuckey.png) | ![](./assets/public_final_result_p1_pedro.png) | ![](./assets/public_final_result_p1_Decal.png) | ![](./assets/public_final_result_p1_Hammer.png) |

### Private
* experiment environment: [CodaLab](https://autodl.lri.fr/) (NVIDIA Tesla P100 1GPU, 4vCPU, 26GB)

| metrics | beatriz | Caucase | Hippoc. | Saturn | ukulele |
|:-------:|--------:|--------:|--------:|--------:|--------:|
| ALC | 0.6756 | 0.7359 | 0.7744 | 0.8309 | 0.9075 |
| 2*AUC-1 | 0.8014 | 0.9411 | 0.9534 | 0.9884 | 0.9985 |
| curves | ![](./assets/private_final_result_beatriz.png) | ![](./assets/private_final_result_Caucase.png) | ![](./assets/private_final_result_Hippocrate.png) | ![](./assets/private_final_result_Saturn.png) | ![](./assets/private_final_result_ukulele.png) |

## [AutoCV2][] Results
### Public (video only)

#### V1.XLARGE
* experiment environment: [BrainCloud][] V1.XLARGE Type (NVIDIA Tesla V100 1GPU, 14CPU, 122GB)

| metrics | Kreature | Katze | Kraut |
|:-------:|---------:|------:|------:|
| ALC | 0.8677 | 0.8613 | 0.6678 |
| 2*AUC-1 | 0.9613 | 0.9588 | 0.7365 |
| curves | ![](./assets/public_final_result_v1_kreature.png) | ![](./assets/public_final_result_v1_katze.png) | ![](./assets/public_final_result_v1_kraut.png) |

#### P1.XLARGE
* experiment environment: [BrainCloud][] P1.XLARGE Type (NVIDIA Tesla P40 1GPU, 6CPU, 61GB)

| metrics | Kreature | Katze | Kraut |
|:-------:|---------:|------:|------:|
| ALC | 0.8675 | 0.8757 | 0.6883 |
| 2*AUC-1 | 0.9587 | 0.9572 | 0.7559 |
| curves | ![](./assets/public_final_result_p1_kreature.png) | ![](./assets/public_final_result_p1_katze.png) | ![](./assets/public_final_result_p1_kraut.png) |

### Private
* experiment environment: [CodaLab](https://autodl.lri.fr/) (NVIDIA Tesla P100 1GPU, 4vCPU, 26GB)

| metrics | Ideal | freddy | Homer | Isaac2 | Formula |
|:-------:|------:|-------:|------:|-------:|--------:|
| ALC | 0.8229 | 0.7516 | 0.3843 | 0.7064 | 0.7661 |
| 2*AUC-1 | 0.9605 | 0.9945 | 0.5500 | 0.9845 | 0.9661 |
| curves | ![](./assets/private_final_result_ideal.png) | ![](./assets/private_final_result_freddy.png) | ![](./assets/private_final_result_homer.png) | ![](./assets/private_final_result_isaac2.png) | ![](./assets/private_final_result_formula.png) |

### Final (blind)
* experiment environment: [CodaLab](https://autodl.lri.fr/) (NVIDIA Tesla P100 1GPU, 4vCPU, 26GB)

| metrics | Apollon | loukooum | Fiona | Monica1 | Kitsune |
|:-------:|------:|-------:|------:|-------:|--------:|
| ALC | 0.5593 | 0.9256 | 0.4074 | 0.4491 | 0.2132 |
| 2*AUC-1 | 0.8022 | 0.9978 | 0.5312 | 0.8617 | 0.2467 |
| curves | ![](./assets/blind_final_result_apollon.png) | ![](./assets/blind_final_result_loukoum.png) | ![](./assets/blind_final_result_fiona.png) | ![](./assets/blind_final_result_monica1.png) | ![](./assets/blind_final_result_kitsune.png) |

## Environment Setup & Experiments
* base docker environment: https://hub.docker.com/r/evariste/autodl

* pre requirements
```bash
$ apt update
$ apt install python3-tk
```

* clone and init. the repository
```bash
$ git clone https://github.com/kakaobrain/autoclint.git && cd autoclint
$ # 3rd parties libarary
$ git submodule init
$ git submodule update
$ # download pretrained models
$ wget https://download.pytorch.org/models/resnet18-5c106cde.pth -O ./models/resnet18-5c106cde.pth
$ # download public datasets
$ cd autodl && python download_public_datasets.py && cd ..
```

* run public datasets
```bash
$ # images
$ python autodl/run_local_test.py -time_budget=1200 -code_dir='./' -dataset_dir='autodl/AutoDL_public_data/Munster/'; cp autodl/AutoDL_scoring_output/learning-curve-*.png ./results
$ python autodl/run_local_test.py -time_budget=1200 -code_dir='./' -dataset_dir='autodl/AutoDL_public_data/Chucky/'; cp autodl/AutoDL_scoring_output/learning-curve-*.png ./results
$ python autodl/run_local_test.py -time_budget=1200 -code_dir='./' -dataset_dir='autodl/AutoDL_public_data/Pedro/'; cp autodl/AutoDL_scoring_output/learning-curve-*.png ./results
$ python autodl/run_local_test.py -time_budget=1200 -code_dir='./' -dataset_dir='autodl/AutoDL_public_data/Decal/'; cp autodl/AutoDL_scoring_output/learning-curve-*.png ./results
$ python autodl/run_local_test.py -time_budget=1200 -code_dir='./' -dataset_dir='autodl/AutoDL_public_data/Hammer/'; cp autodl/AutoDL_scoring_output/learning-curve-*.png ./results
$ # videos
$ python autodl/run_local_test.py -time_budget=1200 -code_dir='./' -dataset_dir='autodl/AutoDL_public_data/Kreatur/'; cp autodl/AutoDL_scoring_output/learning-curve-*.png ./results
$ python autodl/run_local_test.py -time_budget=1200 -code_dir='./' -dataset_dir='autodl/AutoDL_public_data/Katze/'; cp autodl/AutoDL_scoring_output/learning-curve-*.png ./results
$ python autodl/run_local_test.py -time_budget=1200 -code_dir='./' -dataset_dir='autodl/AutoDL_public_data/Kraut/'; cp autodl/AutoDL_scoring_output/learning-curve-*.png ./results
```

* (optional) display learning curve
```bash
$ # item2 utils to visualize learning curve
$ wget https://www.iterm2.com/utilities/imgcat -O bin/imgcat; chmod 0677 bin/imgcat
$ bin/imgcat ./results/learning-curve-*.png
```

## Authors and Licensing
This project is developed by [Woonhyuk Baek][], [Ildoo Kim][] and [Sungbin Lim][] at
[Kakao Brain][]. It is distributed under [Apache License
2.0](LICENSE).

## Citation

If you apply this library to any project and research, please cite our code:

```
@article{baek2020autoclint,
title = {AutoCLINT: The Winning Method in AutoCV Challenge 2019}
author = {Woonhyuk Baek and Ildoo Kim and Sungwoong Kim and Sungbin Lim},
year = {2020},
eprint = {2005.04373},
archivePrefix = {arXiv}
}
```

## References & Open sources
1. [Fast AutoAugment][]
- paper: https://arxiv.org/abs/1905.00397
- codes: https://github.com/kakaobrain/fast-autoaugment
2. Pretrained models for Pytorch
- codes: https://github.com/Cadene/pretrained-models.pytorch
3. TorchVision models
- pages: https://pytorch.org/docs/stable/torchvision/models.html
3. TQDM: Progress Bar for Python and CLI
- codes: https://github.com/tqdm/tqdm
4. AutoCV/AutoDL startking kit
- codes: https://github.com/zhengying-liu/autodl_starting_kit_stable

[Kakao Brain]: https://kakaobrain.com/
[BrainCloud]: https://cloud.kakaobrain.com/
[Sungbin Lim]: https://github.com/sungbinlim
[Ildoo Kim]: https://github.com/ildoonet
[Woonhyuk Baek]: https://github.com/wbaek
[Fast AutoAugment]: https://arxiv.org/abs/1905.00397
[AutoCV]: https://autodl.lri.fr/competitions/118#home
[AutoCV2]: https://autodl.lri.fr/competitions/3#home
[NeurIPS 2019 AutoDL Challenges]: https://autodl.chalearn.org/