Ecosyste.ms: Awesome
An open API service indexing awesome lists of open source software.
https://github.com/beresandras/clear-diffusion-keras
Implementation of denoising diffusion models with schedules, improved sampling, and other extensions using Keras.
https://github.com/beresandras/clear-diffusion-keras
Last synced: 14 days ago
JSON representation
Implementation of denoising diffusion models with schedules, improved sampling, and other extensions using Keras.
- Host: GitHub
- URL: https://github.com/beresandras/clear-diffusion-keras
- Owner: beresandras
- License: mit
- Created: 2022-05-31T18:05:28.000Z (over 2 years ago)
- Default Branch: master
- Last Pushed: 2024-01-22T10:19:05.000Z (10 months ago)
- Last Synced: 2024-08-01T16:52:59.021Z (3 months ago)
- Language: Python
- Homepage:
- Size: 12.7 MB
- Stars: 113
- Watchers: 5
- Forks: 15
- Open Issues: 9
-
Metadata Files:
- Readme: README.md
- License: LICENSE
Awesome Lists containing this project
README
# Modular and Readable Denoising Diffusion Models in Keras
![flowers stochastic generation](./assets/generation.webp)
Diffusion models are trained to denoise noisy images, and can generate images by iteratively denoising pure noise.
[![spaces badge](https://img.shields.io/badge/%F0%9F%A4%97%20Spaces-Demo-blue.svg)](https://huggingface.co/spaces/keras-io/denoising-diffusion-implicit-models)
[![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/beresandras/clear-diffusion-keras/blob/master/clear-diffusion-keras.ipynb)This repository contains:
* An implementation of [Denoising Diffusion Implicit Models (DDIM)](https://arxiv.org/abs/2010.02502) with continuous time. All variables are properly named and the code is densely commented. It was used for ablations and hyperparameter optimization for the corresponding [Keras code example](https://keras.io/examples/generative/ddim/).
* Stochastic sampling, with which the model becomes a [Denoising Diffusion Probabilistic Model (DDPM)](https://arxiv.org/abs/2006.11239). `Stochasticity` corresponds to *eta* in the DDIM paper, while the `variance_preserving` flag selects between the two sampling versions ([Equation 16 in DDIM](https://arxiv.org/abs/2010.02502)).
* [General second-order sampling](https://en.wikipedia.org/wiki/List_of_Runge%E2%80%93Kutta_methods#Generic_second-order_method), as proposed in [Elucidating the Design Space of Diffusion-Based Generative Models](https://arxiv.org/abs/2206.00364), controlled with `second_order_alpha`.
* [Multistep sampling](https://en.wikipedia.org/wiki/Linear_multistep_method#Adams%E2%80%93Bashforth_methods), similarly to [Pseudo Numerical Methods for Diffusion Models on Manifolds (PNDM)](https://arxiv.org/abs/2202.09778), supporting `num_multisteps` between 1 and 5. Note that in the initial steps I use lower order multistep sampling, instead of other higher-order methods, for simplicity.
* 3 diffusion schedules, selected with `schedule_type`, see below.
* 3 network parametrizations, selected with `prediction_type`. It can predict the unscaled random gaussian noise, the original image, or even the [diffusion velocity (v in Section 4)](https://arxiv.org/abs/2202.00512).
* 3 loss weightings, selected with `loss_type`, which correspond minimising the error of the predicted unscaled noise, predicted original image, or the [diffusion velocity](https://arxiv.org/abs/2202.00512).The network was optimized to offer reasonable performance with modest compute requirements (training time is below an hour on an A100). Other design choices are explained in detail in the corresponding [Keras code example](https://keras.io/examples/generative/ddim/).
## Sampling techniques
KID at different sampling steps with different sampling techniques, using cosine schedule. Note that I selected the sampling hyperparameters using DDIM sampling and 5 diffusion steps.
![sampling techniques](./assets/sampling.png)
For first order methods network evaluations = diffusion steps, and for second order methods network evaluations = 2 * diffusion steps.
## Diffusion schedules
![diffusion schedules](./assets/schedules.png)
For this plot I used 100 diffusion steps, and a `start_log_snr` and `end_log_snr` of 5.0 and -5.0 for symmetry, while their defaults are 2.5 and -7.5.
For implementation details, check out `diffusion_schedule()` in [model.py](model.py).
## Generation quality
[Kernel Inception Distance (KID)](https://arxiv.org/abs/1801.01401):
Dataset / Loss | mean absolute error (MAE) | mean squared error (MSE)
--- | --- | ---
**Oxford Flowers** | 0.282 | 0.399
**CelebA** | 0.148 | 0.104
**Caltech Birds** | 1.382 | 1.697
**CIFAR-10** | 0.217 | 0.175Network output / Loss weighting | noise | velocity | signal
--- | --- | --- | ---
**noise** | 0.282 | 0.327 | 0.348
**velocity** | 0.299 | 0.290 | 0.333
**signal** | 0.291 | 0.319 | 0.329Trained with default hyperparameters if not mentioned otherwise, tuned on Oxford Flowers.
* KID is a generative performance metric with a simple unbiased estimator, that is more suitable for limited amounts of images, and is also computationally cheaper to measure compared to the [Frechet Inception Distance (FID)](https://arxiv.org/abs/1706.08500).
* The Inceptionv3 network's pretrained weights are loaded from [Keras applications](https://keras.io/api/applications/inceptionv3/).
* For computational efficiency, the images are evaluated at the minimal possible resolution (75x75 instead of 299x299), therefore the exact values might not be comparable with other implementations.
* For computational efficiency, it is measured only on the validation splits of the datasets.
* For computational efficiency, it is measured on images generated with only 5 diffusion steps.## Visualizations
All visualizations below were generated using:
* 200 diffusion steps
* DDPM sampling with large variance (`stochasticity = 1.0, variance_preserving = False`)
* all other parameters left on default### Oxford Flowers 102
* 6500 training images (80% of every split)
* 64x64 resolution, center cropped![flowers generated images](./assets/flowers.png)
### CelebFaces Attributes (CelebA)
* 160.000 training images
* 64x64 resolution, center cropped![celeba generated images](./assets/celeba.png)
### Caltech Birds 2011 (CUB-200)
* 6000 training images
* 64x64 resolution, cropped on bounding boxes![birds generated images](./assets/birds.png)
### CIFAR-10
* 50.000 training images
* 32x32 resolution![cifar10 generated images](./assets/cifar10.png)
For a similar implementation of GANs and GAN losses, check out [this repository](https://github.com/beresandras/gan-flavours-keras).