Ecosyste.ms: Awesome
An open API service indexing awesome lists of open source software.
https://github.com/Chen-Yang-Liu/RSICC
Official LEVIR-CC dataset and Pytorch implementation for Remote Sensing Image Change Captioning With Dual-Branch Transformers: A New Method and a Large Scale Dataset
https://github.com/Chen-Yang-Liu/RSICC
remote-sensing-image-change-captioning
Last synced: 3 months ago
JSON representation
Official LEVIR-CC dataset and Pytorch implementation for Remote Sensing Image Change Captioning With Dual-Branch Transformers: A New Method and a Large Scale Dataset
- Host: GitHub
- URL: https://github.com/Chen-Yang-Liu/RSICC
- Owner: Chen-Yang-Liu
- Created: 2022-04-26T07:44:40.000Z (over 2 years ago)
- Default Branch: main
- Last Pushed: 2024-05-11T13:35:52.000Z (6 months ago)
- Last Synced: 2024-07-12T12:39:46.007Z (4 months ago)
- Topics: remote-sensing-image-change-captioning
- Language: Python
- Homepage:
- Size: 67.2 MB
- Stars: 99
- Watchers: 3
- Forks: 6
- Open Issues: 9
-
Metadata Files:
- Readme: README.md
Awesome Lists containing this project
- awesome-remote-sensing-change-detection - **LEVIR Change Captioning (LEVIR-CC) dataset**
README
Remote Sensing Image Change Captioning With Dual-Branch Transformers: A New Method and a Large Scale Dataset
**[Chenyang Liu](https://chen-yang-liu.github.io/), [Rui Zhao](https://ruizhaocv.github.io), [Hao Chen](http://chenhao.in/), [Zhengxia Zou](https://scholar.google.com.hk/citations?hl=en&user=DzwoyZsAAAAJ), and [Zhenwei Shi*✉](https://scholar.google.com.hk/citations?hl=en&user=kNhFWQIAAAAJ)**
![visitors](https://visitor-badge.glitch.me/badge?page_id=Chen-Yang-Liu.RSICC.visitor)
![GitHub stars](https://badgen.net/github/stars/Chen-Yang-Liu/RSICC)## Share us a :star: if this repo does help
## LEVIR-CC Dataset
**Download [Link](https://github.com/Chen-Yang-Liu/LEVIR-CC-Dataset)**## RSICCfromer
Here, we provide the pytorch implementation of the paper: "Remote Sensing Image Change Captioning With Dual-Branch Transformers: A New Method and a Large Scale Dataset".For more information, please see our published paper in [[IEEE](https://ieeexplore.ieee.org/document/9934924) | [Lab Server](http://levir.buaa.edu.cn/publications/ChangeCaptioning.pdf)] ***(Accepted by TGRS 2022)***
![RSICCformer_structure](Example/RSICCformer_structure.png)
### Installation and Dependencies
```python
git clone https://github.com/Chen-Yang-Liu/RSICC
cd RSICC
conda create -n RSICCformer_env python=3.6
conda activate RSICCformer_env
pip install -r requirements.txt
```### Data preparation
Firstly, put the downloaded dataset in `./LEVIR_CC_dataset/`.
Then preprocess dataset as follows:
```python
python create_input_files.py --min_word_freq 5
```
After that, you can find some resulted files in `./data/`.Besides, the resulted files can also be downloaded from here: [[Google Drive](https://drive.google.com/drive/folders/1cEv-BXISfWjw1RTzL39uBojH7atjLdCG?usp=sharing) | [Baidu Pan](https://pan.baidu.com/s/1YrWcz090kdqOZ0lrbqXJJA) (code:nq9y)]. Extract it to `./data/`.
!NOTE: For a fair comparison, we suggest that future researchers ensure `min_word_freq <= 5` or use our preprocessed data above with `min_word_freq = 5`.
### Inference Demo
You can download our RSICCformer pretrained model——by [[Google Drive](https://drive.google.com/drive/folders/1cEv-BXISfWjw1RTzL39uBojH7atjLdCG?usp=sharing) | [Baidu Pan](https://pan.baidu.com/s/1SBGjVS0yd2KHdK9t4NuiyA) (code:2fbc)]After downloaded the pretrained model, you can put it in `./models_checkpoint/`.
Then, run a demo to get started as follows:
```python
python caption.py --img_A ./Example/A/train_000016.png --img_B ./Example/B/train_000016.png --path ./models_checkpoint/
```
After that, you can find the generated caption in `./eval_results/`### Train
Make sure you performed the data preparation above. Then, start training as follows:
```python
python train.py --data_folder ./data/ --savepath ./models_checkpoint/
```### Evaluate
```python
python eval.py --data_folder ./data/ --path ./models_checkpoint/ --Split TEST
```
We recommend training 5 times to get an average score.## Citation:
```
@ARTICLE{9934924,
author={Liu, Chenyang and Zhao, Rui and Chen, Hao and Zou, Zhengxia and Shi, Zhenwei},
journal={IEEE Transactions on Geoscience and Remote Sensing},
title={Remote Sensing Image Change Captioning With Dual-Branch Transformers: A New Method and a Large Scale Dataset},
year={2022},
volume={60},
number={},
pages={1-20},
doi={10.1109/TGRS.2022.3218921}}
```
## Reference:
Thanks to the following repository:
[a-PyTorch-Tutorial-to-Image-Captioning](https://github.com/sgrvinod/a-PyTorch-Tutorial-to-Image-Captioning.git)