Ecosyste.ms: Awesome
An open API service indexing awesome lists of open source software.
https://github.com/vt-vl-lab/FGVC
[ECCV 2020] Flow-edge Guided Video Completion
https://github.com/vt-vl-lab/FGVC
Last synced: about 1 month ago
JSON representation
[ECCV 2020] Flow-edge Guided Video Completion
- Host: GitHub
- URL: https://github.com/vt-vl-lab/FGVC
- Owner: vt-vl-lab
- License: other
- Created: 2020-09-09T18:24:41.000Z (over 4 years ago)
- Default Branch: master
- Last Pushed: 2021-12-14T05:49:53.000Z (almost 3 years ago)
- Last Synced: 2024-08-01T13:29:46.948Z (4 months ago)
- Language: Python
- Size: 94.7 KB
- Stars: 1,550
- Watchers: 68
- Forks: 263
- Open Issues: 14
-
Metadata Files:
- Readme: README.md
- License: LICENSE
Awesome Lists containing this project
- awesome-video - vt-vl-lab/FGVC - [ECCV 2020] Flow-edge Guided Video Completion . (Encoding / Talks Presentations Podcasts)
README
# [ECCV 2020] Flow-edge Guided Video Completion
### [[Paper](https://arxiv.org/abs/2009.01835)] [[Project Website](http://chengao.vision/FGVC/)] [[Google Colab](https://colab.research.google.com/drive/1pb6FjWdwq_q445rG2NP0dubw7LKNUkqc?usp=sharing)]
We present a new flow-based video completion algorithm. Previous flow completion methods are often unable to retain the sharpness of motion boundaries. Our method first extracts and completes motion edges, and then uses them to guide piecewise-smooth flow completion with sharp edges. Existing methods propagate colors among local flow connections between adjacent frames. However, not all missing regions in a video can be reached in this way because the motion boundaries form impenetrable barriers. Our method alleviates this problem by introducing non-local flow connections to temporally distant frames, enabling propagating video content over motion boundaries. We validate our approach on the DAVIS dataset. Both visual and quantitative results show that our method compares favorably against the state-of-the-art algorithms.
**[ECCV 2020] Flow-edge Guided Video Completion**
[Chen Gao](http://chengao.vision), [Ayush Saraf](#), [Jia-Bin Huang](https://filebox.ece.vt.edu/~jbhuang/), and [Johannes Kopf](https://johanneskopf.de/)
In European Conference on Computer Vision (ECCV), 2020## Prerequisites
- Linux (tested on CentOS Linux release 7.4.1708)
- Anaconda
- Python 3.8 (tested on 3.8.5)
- PyTorch 1.6.0and the Python dependencies listed in requirements.txt
- To get started, please run the following commands:
```
conda create -n FGVC
conda activate FGVC
conda install pytorch=1.6.0 torchvision=0.7.0 cudatoolkit=10.1 matplotlib scipy opencv -c pytorch
pip install imageio imageio-ffmpeg scikit-image imutils
```- Next, please download the model weight and demo data using the following command:
```
chmod +x download_data_weights.sh
./download_data_weights.sh
```## Quick start
- Object removal:
```bash
cd tool
python video_completion.py \
--mode object_removal \
--path ../data/tennis \
--path_mask ../data/tennis_mask \
--outroot ../result/tennis_removal \
--seamless
```- FOV extrapolation:
```bash
cd tool
python video_completion.py \
--mode video_extrapolation \
--path ../data/tennis \
--outroot ../result/tennis_extrapolation \
--H_scale 2 \
--W_scale 2 \
--seamless
```You can remove the `--seamless` flag for a faster processing time.
## License
This work is licensed under MIT License. See [LICENSE](LICENSE) for details.If you find this code useful for your research, please consider citing the following paper:
@inproceedings{Gao-ECCV-FGVC,
author = {Gao, Chen and Saraf, Ayush and Huang, Jia-Bin and Kopf, Johannes},
title = {Flow-edge Guided Video Completion},
booktitle = {European Conference on Computer Vision},
year = {2020}
}## Acknowledgments
- Our flow edge completion network builds upon [EdgeConnect](https://github.com/knazeri/edge-connect).
- Our image inpainting network is modified from [DFVI](https://github.com/nbei/Deep-Flow-Guided-Video-Inpainting).