Ecosyste.ms: Awesome
An open API service indexing awesome lists of open source software.
https://github.com/surmenok/keras_lr_finder
Plots the change of the loss function of a Keras model when the learning rate is exponentially increasing.
https://github.com/surmenok/keras_lr_finder
Last synced: 8 days ago
JSON representation
Plots the change of the loss function of a Keras model when the learning rate is exponentially increasing.
- Host: GitHub
- URL: https://github.com/surmenok/keras_lr_finder
- Owner: surmenok
- License: mit
- Created: 2017-12-11T03:18:18.000Z (almost 7 years ago)
- Default Branch: master
- Last Pushed: 2020-07-04T17:57:29.000Z (over 4 years ago)
- Last Synced: 2024-09-18T18:08:46.805Z (about 2 months ago)
- Language: Python
- Size: 53.7 KB
- Stars: 254
- Watchers: 8
- Forks: 65
- Open Issues: 19
-
Metadata Files:
- Readme: README.md
- License: LICENSE
Awesome Lists containing this project
README
# keras_lr_finder
Plots the change of the loss function of a Keras model when the learning rate is exponentially increasing.
## Purpose
See details in ["Estimating an Optimal Learning Rate For a Deep Neural Network"](https://towardsdatascience.com/estimating-optimal-learning-rate-for-a-deep-neural-network-ce32f2556ce0).## Usage
Create and compile a Keras model, then execute this code:```python
# model is a Keras model
lr_finder = LRFinder(model)# Train a model with batch size 512 for 5 epochs
# with learning rate growing exponentially from 0.0001 to 1
lr_finder.find(x_train, y_train, start_lr=0.0001, end_lr=1, batch_size=512, epochs=5)
``````python
# Plot the loss, ignore 20 batches in the beginning and 5 in the end
lr_finder.plot_loss(n_skip_beginning=20, n_skip_end=5)
```![Loss function](https://cdn-images-1.medium.com/max/1600/1*HVj_4LWemjvOWv-cQO9y9g.png)
```python
# Plot rate of change of the loss
# Ignore 20 batches in the beginning and 5 in the end
# Smooth the curve using simple moving average of 20 batches
# Limit the range for y axis to (-0.02, 0.01)
lr_finder.plot_loss_change(sma=20, n_skip_beginning=20, n_skip_end=5, y_lim=(-0.01, 0.01))
```![Rate of change of the loss function](https://cdn-images-1.medium.com/max/1600/1*87mKq_XomYyJE29l91K0dw.png)
## Contributions
Contributions are welcome. Please, file issues and submit pull requests on GitHub, or contact me directly.## References
This code is based on:
- The method described in section 3.3 of the 2015 paper ["Cyclical Learning Rates for Training Neural Networks"](https://arxiv.org/abs/1506.01186) by Leslie N. Smith
- The implementation of the algorithm in [fastai library](https://github.com/fastai/fastai) by Jeremy Howard. See [fast.ai deep learning course](http://course.fast.ai/) for details.