Ecosyste.ms: Awesome
An open API service indexing awesome lists of open source software.
https://github.com/WangJY06/RDGAN
ICME 2019 paper "RDGAN: Retinex Decomposition based Adversarial Learning for Low-Light Enhancement"
https://github.com/WangJY06/RDGAN
Last synced: 2 months ago
JSON representation
ICME 2019 paper "RDGAN: Retinex Decomposition based Adversarial Learning for Low-Light Enhancement"
- Host: GitHub
- URL: https://github.com/WangJY06/RDGAN
- Owner: WangJY06
- Created: 2019-03-20T20:06:33.000Z (almost 6 years ago)
- Default Branch: master
- Last Pushed: 2019-11-22T16:05:57.000Z (about 5 years ago)
- Last Synced: 2024-05-21T15:19:00.912Z (8 months ago)
- Language: Python
- Homepage:
- Size: 66.2 MB
- Stars: 31
- Watchers: 2
- Forks: 10
- Open Issues: 4
-
Metadata Files:
- Readme: README.md
Awesome Lists containing this project
README
# RDGAN: Retinex Decomposition based Adversarial Learning for Low-Light Enhancement
[**ICME 2019 version**](https://ieeexplore.ieee.org/abstract/document/8785047)[**ICME 2019 supplement material**](RDGAN_Supp.pdf)
[**ICME 2019 poster**](ICME_Poster.pdf)
## 1. Abstract
Pictures taken under the low-light condition often suffer from low contrast and loss of image details, thus an approach that can effectively improve low-light images is demanded. Traditional Retinex-based methods assume that the reflectance components of low-light images keep unchanged, which neglect the color distortion and lost details. In this paper, we propose an end-to-end learning-based framework that first decomposes the low-light image and then learns to fuse the decomposed results to obtain the high quality enhanced result. Our framework can be divided into a RDNet (Retinex Decomposition Network) for decomposition and a FENet (Fusion Enhancement Network) for fusion. Specific multi-term losses are respectively designed for the two networks. We also present a new RDGAN (Retinex Decomposition based Generative Adversarial Network) loss, which is computed on the decomposed reflectance components of the enhanced and the reference images. Experiments demonstrate that our approach is good at color and detail restoration, which outperforms other state-of-the-art methods.## 2. Network Structure
![Image of Network](figs/model.png)## 3. Testing
Run [test.py](test.py) for results.## 4. Retinex Decomposition Comparison
![Image of RD](figs/rd.png)## 5. Low-Light Enhancement Comparison
![Image of Results](figs/result.png)## 6. Dataset
Following [SICE](https://github.com/csjcai/SICE).See more testing images in the folder [samples_sup](samples_sup).
## 7. Citation
```
@inproceedings{wang2019rdgan,
title={RDGAN: Retinex Decomposition Based Adversarial Learning for Low-Light Enhancement},
author={Wang, Junyi and Tan, Weimin and Niu, Xuejing and Yan, Bo},
booktitle={2019 IEEE International Conference on Multimedia and Expo (ICME)},
pages={1186--1191},
year={2019},
organization={IEEE}
}
```