Ecosyste.ms: Awesome
An open API service indexing awesome lists of open source software.
https://github.com/lrunaways/awesome-deep-learning-mustreads
Must-read papers and must-known concepts
https://github.com/lrunaways/awesome-deep-learning-mustreads
List: awesome-deep-learning-mustreads
Last synced: 16 days ago
JSON representation
Must-read papers and must-known concepts
- Host: GitHub
- URL: https://github.com/lrunaways/awesome-deep-learning-mustreads
- Owner: lrunaways
- Created: 2020-06-10T23:35:09.000Z (over 4 years ago)
- Default Branch: master
- Last Pushed: 2021-05-30T19:20:11.000Z (over 3 years ago)
- Last Synced: 2024-04-10T13:55:28.145Z (9 months ago)
- Homepage:
- Size: 108 KB
- Stars: 0
- Watchers: 1
- Forks: 0
- Open Issues: 0
-
Metadata Files:
- Readme: README.md
Awesome Lists containing this project
- ultimate-awesome - awesome-deep-learning-mustreads - Must-read papers and must-known concepts. (Other Lists / PowerShell Lists)
README
# Awesome deep learning papers
### Must-read papers and must-known concepts## [Architectures](https://github.com/lrunaways/awesome-deep-learning-mustreads/blob/master/topics/Architectures.rst)
Neural networks evolution
[Review: Inception-v3 Sep 10 2018](https://medium.com/@sh.tsang/review-inception-v3-1st-runner-up-image-classification-in-ilsvrc-2015-17915421f77c)
[Dilated Residual Networks 28 May 2017](https://arxiv.org/pdf/1705.09914.pdf)
[Attention Is All You Need [6 Dec 2017]](https://arxiv.org/pdf/1706.03762.pdf)
(https://habr.com/ru/post/486158/)
[EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks 28 May 2019](https://arxiv.org/abs/1905.11946)
## [Training](https://github.com/lrunaways/awesome-deep-learning-mustreads/blob/master/topics/Training.rst)
Optimization techniques, regularization, tips & tricks, distribution
[A Recipe for Training Neural Networks Apr 25 2019](http://karpathy.github.io/2019/04/25/recipe/)
[Multi-GPU Training of ConvNets 18 Feb 2014](https://arxiv.org/pdf/1312.5853.pdf)
[The Effectiveness of Data Augmentation in Image Classification using Deep Learning 13 Dec 2017](https://arxiv.org/pdf/1712.04621.pdf)
[Parallel and Distributed Deep Learning](https://web.stanford.edu/~rezab/classes/cme323/S16/projects_reports/hedge_usmani.pdf)
[Demystifying Parallel and Distributed Deep Learning: An In-Depth Concurrency Analysis 15 Sep 2018](https://arxiv.org/pdf/1802.09941.pdf)
[Accurate, Large Minibatch SGD: Training ImageNet in 1 Hour 30 Apr 2018](https://arxiv.org/pdf/1706.02677.pdf)
[Highly Scalable Deep Learning Training System with Mixed-Precision: Training ImageNet in Four Minutes 30 Jul 2018](https://arxiv.org/pdf/1807.11205.pdf)
[MIXED PRECISION TRAINING 15 Feb 2018](https://arxiv.org/pdf/1710.03740.pdf)
[A Survey on Distributed Machine Learning 20 Dec 2019](https://arxiv.org/ftp/arxiv/papers/1912/1912.09789.pdf)
[NVIDIA Deep Learning Performance](https://docs.nvidia.com/deeplearning/performance/index.html)
[Stochastic Weight Averaging](https://towardsdatascience.com/stochastic-weight-averaging-a-new-way-to-get-state-of-the-art-results-in-deep-learning-c639ccf36a)
[Bag of Tricks for Image Classification with Convolutional Neural Networks 5 Dec 2018](https://arxiv.org/pdf/1812.01187.pdf)
[Training Neural Nets on Larger Batches: Practical Tips for 1-GPU, Multi-GPU & Distributed setups](https://medium.com/huggingface/training-larger-batches-practical-tips-on-1-gpu-multi-gpu-distributed-setups-ec88c3e51255)
[Training Generative Adversarial Networks with Limited Data 7 Oct 2020](https://arxiv.org/pdf/2006.06676.pdf)
## [Theory](https://github.com/lrunaways/awesome-deep-learning-mustreads/blob/master/topics/Theory.rst)
Understanding how neural networks work on a deeper level
[Deconvolution and Checkerboard Artifacts 2016](http://doi.org/10.23915/distill.00003)
[UNDERSTANDING DEEP LEARNING REQUIRES RETHINKING GENERALIZATION 26 Feb 2017](https://arxiv.org/pdf/1611.03530.pdf)
[Emergence of Invariance and Disentanglement in Deep Representations 28 Jun 2018](https://arxiv.org/pdf/1706.01350.pdf)
[Averaging Weights Leads to Wider Optima and Better Generalization 25 Feb 2019](https://arxiv.org/pdf/1803.05407.pdf)
[Towards a Mathematical Understanding of Neural Network-Based Machine Learning 1 Oct 2020](https://arxiv.org/pdf/2009.10713v2.pdf)
## [Computer vision](https://github.com/lrunaways/awesome-deep-learning-mustreads/blob/master/topics/Computer_vision.rst)
Semantic segmentation, object localization
[Hypercolumns for Object Segmentation and Fine-grained Localization](https://arxiv.org/pdf/1411.5752.pdf)
[On the Benefit of Adversarial Training for Monocular Depth Estimation 29 Oct 2019](https://arxiv.org/pdf/1910.13340.pdf)
## [Transfer learning](https://github.com/lrunaways/awesome-deep-learning-mustreads/blob/master/topics/Transfer_learning.rst)
Knowledge transfer, distilation, domain adaptation
[Data augmentation using generative adversarial networks (CycleGAN) to improve generalizability in CT segmentation tasks](https://www.nature.com/articles/s41598-019-52737-x)
## [Synthetic data](https://github.com/lrunaways/awesome-deep-learning-mustreads/blob/master/topics/Synthetic_data.rst)
Generating and using synthetic data to improve performance on real data
## [GANs](https://github.com/lrunaways/awesome-deep-learning-mustreads/blob/master/topics/GANs.rst)
Generative adversarial network evolution, tips on training
[Generative Adversarial Networks 10 Jun 2014](https://arxiv.org/abs/1406.2661)
[Conditional Generative Adversarial Nets 6 Nov 2014](https://arxiv.org/abs/1411.1784)
[Improved Techniques for Training GANs 10 Jun 2016](https://arxiv.org/pdf/1606.03498.pdf)
Feature matching - match intermediate discriminator layer statistics (mean) for real and generated examples
Minibatch discrimination - compute closeness of examples in a batch and use it as additional info for discriminator
Historical averaging - regularize weights to be not-very-far (L2) from previous weights of a network
One-sided label smoothing - smooth (change) labels of positive examples to alpha (0.9)
[Deeplearning.ai specialization: Generative Adversarial Networks](https://www.coursera.org/specializations/generative-adversarial-networks-gans)
[Conditional Generative Adversarial Nets 6 Nov 2014](https://arxiv.org/abs/1411.1784)
[Deep Convolutional Generative Adversarial Networks 19 Nov 2015](https://arxiv.org/abs/1511.06434)
[Wasserstein GAN 6 Dec 2017](https://arxiv.org/pdf/1701.07875.pdf)
[Improved Training of Wasserstein GANs 31 Mar 2017](https://arxiv.org/abs/1704.00028)
[From GAN to WGAN 20 Aug 2017](https://lilianweng.github.io/lil-log/2017/08/20/from-GAN-to-WGAN.html)
[Conditional Image Synthesis with Auxiliary Classifier GANs 20 Jul 2017](https://arxiv.org/pdf/1610.09585.pdf)
[GANs for Biological Image Synthesis 12 Sep 2017](https://arxiv.org/pdf/1708.04692.pdf)
[PROGRESSIVE GROWING OF GANS FOR IMPROVED QUALITY, STABILITY, AND VARIATION 26 Feb 2018](https://arxiv.org/pdf/1710.10196.pdf)
[Image-to-Image Translation with Conditional Adversarial Networks 26 Nov 2018](https://arxiv.org/pdf/1611.07004.pdf)
[GAN DISSECTION: VISUALIZING AND UNDERSTANDING GENERATIVE ADVERSARIAL NETWORKS 8 Dec 2018](https://arxiv.org/pdf/1811.10597.pdf)
[SPECTRAL NORMALIZATION FOR GENERATIVE ADVERSARIAL NETWORKS 16 Feb 2018](https://arxiv.org/pdf/1802.05957.pdf)
[Image-to-Image Translation with Conditional Adversarial Networks 26 Nov 2018](https://arxiv.org/pdf/1611.07004.pdf)
[DATA AUGMENTATION GENERATIVE ADVERSARIAL NETWORKS 21 Mar 2018](https://arxiv.org/pdf/1711.04340.pdf)
[High-Resolution Image Synthesis and Semantic Manipulation with Conditional GANs 20 Aug 2018](https://arxiv.org/pdf/1711.11585.pdf)
[Self-Attention Generative Adversarial Networks 14 Jun 2019](https://arxiv.org/pdf/1805.08318.pdf)
Self-attention modules + Spectral normalization + separate learning rates.
Self attention:
input: x
f(x) = Wf * x
g(x) = Wg * x
beta = row-wise softmax aplied to f(x).T * g(x)
h(x) = Wh * x
u(x) = Wu * x
attention map o = u(beta * h(x))
output: a * o + x
[LARGE SCALE GAN TRAINING FOR HIGH FIDELITY NATURAL IMAGE SYNTHESIS 25 Feb 2019](https://arxiv.org/pdf/1809.11096.pdf)
[A Large-Scale Study on Regularization and Normalization in GANs 14 May 2019](https://arxiv.org/pdf/1807.04720.pdf)
[A Style-Based Generator Architecture for Generative Adversarial Networks 29 Mar 2019](https://arxiv.org/pdf/1812.04948.pdf)
[Attributing Fake Images to GANs: Learning and Analyzing GAN Fingerprints 16 Aug 2019](https://arxiv.org/pdf/1811.08180.pdf)
[LARGE SCALE GAN TRAINING FOR HIGH FIDELITY NATURAL IMAGE SYNTHESIS 25 Feb 2019](https://arxiv.org/pdf/1809.11096.pdf)
[Analyzing and Improving the Image Quality of StyleGAN 23 Mar 2020](https://arxiv.org/pdf/1912.04958.pdf)
[Training Generative Adversarial Networks with Limited Data 7 Oct 2020](https://arxiv.org/abs/2006.06676)
[Interpreting the Latent Space of GANs for Semantic Face Editing 25 Jul 2019](https://arxiv.org/abs/1907.10786)
[Fast Fréchet Inception Distance 29 Sep 2020](https://arxiv.org/pdf/2009.14075.pdf)
[Pros and Cons of GAN Evaluation Measures 9 Feb 2018](https://arxiv.org/abs/1802.03446)
[Large Scale GAN Training for High Fidelity Natural Image Synthesis 28 Sep 2018](https://arxiv.org/abs/1809.11096)
[HYPE: A Benchmark for Human eYe Perceptual Evaluation of Generative Models 1 Apr 2019](https://arxiv.org/abs/1904.01121)
[Improved Precision and Recall Metric for Assessing Generative Models 15 Apr 2019](https://arxiv.org/abs/1904.06991)
[GANILLA: Generative Adversarial Networks for Image to Illustration Translation 13 Feb 2020](https://arxiv.org/abs/2002.05638)
(https://medium.com/@jonathan_hui/gan-how-to-measure-gan-performance-64b988c47732)
(https://machinelearningmastery.com/how-to-implement-the-frechet-inception-distance-fid-from-scratch/)
(https://github.com/sahilkhose/Generative-Adversarial-Networks-GANs-Specialization)
(https://arxiv.org/pdf/1807.10088.pdf)
(https://arxiv.org/pdf/1802.10560.pdf)
(http://wscg.zcu.cz/wscg2016/full/F71-full.pdf)
(https://arxiv.org/pdf/2008.02796.pdf)
(https://arxiv.org/pdf/1812.08352.pdf)
(https://arxiv.org/pdf/1905.01164.pdf)
(https://arxiv.org/pdf/1803.01229.pdf)
(https://arxiv.org/pdf/1905.08233.pdf)
(https://arxiv.org/pdf/1903.07291.pdf)
(https://deeppop.github.io/resources/robinson2017-deeppop.pdf)
(https://machinelearningmastery.com/practical-guide-to-gan-failure-modes/)
[Image Augmentations for GAN Training 4 Jun 2020](https://arxiv.org/abs/2006.02595#google)
[Navigating the GAN Parameter Space for Semantic Image Editing 1 Dec 2020](https://arxiv.org/pdf/2011.13786.pdf)
## [Other](https://github.com/lrunaways/awesome-deep-learning-mustreads/blob/master/topics/Other.rst)
[THE COST OF TRAINING NLP MODELS: A CONCISE OVERVIEW 19 Apr 2020](https://arxiv.org/pdf/2004.08900.pdf)
[On the Measure of Intelligence 25 Nov 2019](https://arxiv.org/pdf/1911.01547.pdf)
[Adversarial Robustness - Theory and Practice](https://adversarial-ml-tutorial.org/)
[Math as code](https://github.com/Jam3/math-as-code/)
[Math Snippets](https://github.com/terkelg/math)
## [Blogs](https://github.com/lrunaways/awesome-deep-learning-mustreads/blob/master/topics/blogs)
Great articles, blogs and awesome lists
[PEP 8 -- Style Guide for Python Code](https://www.python.org/dev/peps/pep-0008/)
[Distill blog](https://distill.pub/)
[OpenAI blog](https://openai.com/blog/)
[Google AI blog](https://ai.googleblog.com/)
[Neurohive](https://neurohive.io/)
[Towards data science](https://towardsdatascience.com)
--- add ---
Autoaugment: Learning augmentation strategies from data
Adversarial autoaugment
Towards principled methods for training generative adversarial networks.
Self-supervised gans via auxiliary rotation loss
Gans trained by a two time-scale update rule converge to a local nash equilibrium
On the "steerability" of generative adversarial networks
Are gans created equal? A large-scale study
Spectral normalization for generative adversarial networks
Self-attention generative adversarial networks.
Unsupervised data augmentation for consistency training
Amortised map inference for image super-resolution
A simple framework for contrastive learning of visual representations.
A holistic approach to semi-supervised learning
Foreground-aware Semantic Representations for Image Harmonization
Foreground-aware Semantic Representations for Image Harmonization
https://habr.com/ru/post/527860/
Freeze the discriminator: a simple baseline for fine-tuning GANs
Towards principled methods for training generative adversarial networks.
StarGAN v2: Diverse image synthesis for multiple domains.
Progressive growing of GANs for improved quality, stability, and variation.
Navigating the GAN Parameter Space for Semantic Image Editing
https://arxiv.org/pdf/1909.13719.pdf
[Implicit Neural Representations with Periodic Activation Functions](https://arxiv.org/pdf/2006.09661v1.pdf)
[Stylized Neural Painting](https://arxiv.org/pdf/2011.08114v1.pdf)
[A Neural Algorithm of Artistic Style](https://arxiv.org/pdf/1508.06576.pdf)