Ecosyste.ms: Awesome
An open API service indexing awesome lists of open source software.
https://github.com/ayulockin/losslandscape
Explores the ideas presented in Deep Ensembles: A Loss Landscape Perspective (https://arxiv.org/abs/1912.02757) by Stanislav Fort, Huiyi Hu, and Balaji Lakshminarayanan.
https://github.com/ayulockin/losslandscape
deep-neural-networks keras loss-landscape neural-networks tensorflow
Last synced: 2 months ago
JSON representation
Explores the ideas presented in Deep Ensembles: A Loss Landscape Perspective (https://arxiv.org/abs/1912.02757) by Stanislav Fort, Huiyi Hu, and Balaji Lakshminarayanan.
- Host: GitHub
- URL: https://github.com/ayulockin/losslandscape
- Owner: ayulockin
- License: apache-2.0
- Created: 2020-07-13T15:19:41.000Z (over 4 years ago)
- Default Branch: master
- Last Pushed: 2020-07-29T02:59:53.000Z (over 4 years ago)
- Last Synced: 2024-10-03T12:16:20.190Z (3 months ago)
- Topics: deep-neural-networks, keras, loss-landscape, neural-networks, tensorflow
- Language: Jupyter Notebook
- Homepage:
- Size: 2.52 MB
- Stars: 61
- Watchers: 4
- Forks: 4
- Open Issues: 0
-
Metadata Files:
- Readme: README.md
- License: LICENSE
Awesome Lists containing this project
README
# LossLandscape
Explores the ideas presented in [Deep Ensembles: A Loss Landscape Perspective](https://arxiv.org/abs/1912.02757) by Stanislav Fort, Huiyi Hu, and Balaji Lakshminarayanan.
In the paper, the authors investigate the question - ***why deep ensembles work better than single deep neural networks?***
In their investigation the authors figure out:
* Different snapshots (i.e. model from epoch 1, model from epoch 2, and so on) of a same model exhibit functional similarity. Hence, their ensemble is **less likely** to explore the different modes of local minimas in the optimization space.
* Different solutions of a same model (i.e. trained with different random initializations each time) exhibit functional dissimilarity. Hence, their ensemble is **more likely** to explore the different modes of local minimas in the optimization space.Along with these fascinating findings, they present a number of different things that are useful to understand the dynamics of deep neural networks in general. To know more about them check out our report - [Understanding the Effectivity of Ensembles in Deep Learning](https://app.wandb.ai/authors/loss-landscape/reports/Understanding-the-effectivity-of-ensembles-in-deep-learning-(tentative)--VmlldzoxODAxNjA).
## About the notebooks
- `*_CIFAR10.ipynb`: Shows the training process with three different architectures (SmallCNN, MediumCNN, and ResNet20v1) as per the paper (with minor modifications).
- `*_Aug_Val_Acc_Ensembles.ipynb`: Investigates how accuracy can be represented as a function of ensemble size.
- `Visualizing_Function_Space_Similarity_*.ipynb`: Investigates cosine similarity between weights collected from different snapshots and trajectories, prediction disagreement between different snapshots and trajectories, and presents tSNE visualizations of a how particular solution travels along the optimization landscape (not available for ResNet20v1).## Model weights
Available [here](https://github.com/ayulockin/LossLandscape/releases/tag/v0.1.0).## Acknowledgements
Thanks to **Yannic Kilcher** for [his amazing explanation video](https://www.youtube.com/watch?v=5IRlUVrEVL8) of the paper which helped us pursue our experiments.
Thanks to **Balaji Lakshminarayanan** for providing feedback on the initial draft of the report and also for rectifying our mistake on the tSNE projections.