Ecosyste.ms: Awesome
An open API service indexing awesome lists of open source software.
https://github.com/NeroLoh/darts-tensorflow
Tensorflow code for Differentiable architecture search
https://github.com/NeroLoh/darts-tensorflow
Last synced: about 1 month ago
JSON representation
Tensorflow code for Differentiable architecture search
- Host: GitHub
- URL: https://github.com/NeroLoh/darts-tensorflow
- Owner: NeroLoh
- Created: 2018-11-30T12:25:17.000Z (about 6 years ago)
- Default Branch: master
- Last Pushed: 2018-12-29T08:39:50.000Z (almost 6 years ago)
- Last Synced: 2024-08-01T22:50:11.567Z (4 months ago)
- Language: Python
- Size: 506 KB
- Stars: 74
- Watchers: 4
- Forks: 18
- Open Issues: 4
-
Metadata Files:
- Readme: README.md
Awesome Lists containing this project
- awesome-image-classification - unofficial-tensorflow : https://github.com/NeroLoh/darts-tensorflow
- awesome-image-classification - unofficial-tensorflow : https://github.com/NeroLoh/darts-tensorflow
README
## Paper
This code implement the DARTS paper in Tensorflow
>[DARTS: Differentiable Architecture Search](https://arxiv.org/abs/1806.09055)
>Hanxiao Liu, Karen Simonyan, Yiming Yang.
>arXiv:1806.09055.
## Architecture Search
To carry out architecture search using 2nd-order approximation, run
```
python cnn/train_search.py
```
NOTE:
* Only implement the Cifar10 experiments
* Bacht size is set to only 16 (64 in the pytorch Ver) for the OOM issue using one single 1080 Ti GPU
* Only implement the 2nd-order approximation, but you can simply modify the code if you want 1st-order one
* Didn't implement the PATH DROP operation on test code for the unclear motivationHere shows the details training pocess
Fig. The Train loss and Valid loss of the searching processThe train loss is decreased steadly during the searching pocess, which is more stable than the RL based method. Note that the valid loss refer to the 2nd-order loss for architecture params.
Fig. The final normal cell
Fig. The final reduction cell
## Architecture evaluation
The test error finally decreased to around 5.8 after training 415 epoches, while the best results in pytorch version is 2.76 but trained with 600 epoches. Training more epoches can narrow the gap of performance drop. Besides that, repeating the architecture search process with different seed to choose the best structure can avoid local minimun.
Fig. The test accuracy of the searched architecture## Acknowledgement
* This implementation is based on the original Torch implementation [quark0/darts](https://github.com/quark0/darts)