Ecosyste.ms: Awesome
An open API service indexing awesome lists of open source software.
https://github.com/shijx12/XNM-Net
Pytorch implementation of "Explainable and Explicit Visual Reasoning over Scene Graphs "
https://github.com/shijx12/XNM-Net
clevr cvpr2019 explainable-ai neural-module-networks scene-graph visual-reasoning
Last synced: 14 days ago
JSON representation
Pytorch implementation of "Explainable and Explicit Visual Reasoning over Scene Graphs "
- Host: GitHub
- URL: https://github.com/shijx12/XNM-Net
- Owner: shijx12
- License: mit
- Created: 2018-10-05T07:23:19.000Z (about 6 years ago)
- Default Branch: master
- Last Pushed: 2019-03-17T15:58:22.000Z (over 5 years ago)
- Last Synced: 2024-08-02T08:10:14.935Z (4 months ago)
- Topics: clevr, cvpr2019, explainable-ai, neural-module-networks, scene-graph, visual-reasoning
- Language: Python
- Homepage:
- Size: 14.4 MB
- Stars: 95
- Watchers: 3
- Forks: 18
- Open Issues: 4
-
Metadata Files:
- Readme: README.md
- License: LICENSE.txt
Awesome Lists containing this project
- awesome-gcn - shijx12/XNM-Net
README
# eXplainable and eXplicit Neural Modules (XNMs)
Pytorch implementation of paper
**[Explainable and Explicit Visual Reasoning over Scene Graphs](https://arxiv.org/abs/1812.01855)**
[Jiaxin Shi](https://shijx12.github.io), [Hanwang Zhang](http://www.ntu.edu.sg/home/hanwangzhang/#aboutme), Juanzi LiFlowchart of our model:
A visualization of our reasoning process:
If you find this code useful in your research, please cite
``` tex
@inproceedings{shi2019explainable,
title={Explainable and Explicit Visual Reasoning over Scene Graphs},
author={Jiaxin Shi, Hanwang Zhang, Juanzi Li},
booktitle={CVPR},
year={2019}
}
```## Requirements
- python==3.6
- pytorch==0.4.0
- h5py
- tqdm
- matplotlib## Experiments
We have 4 experiment settings:
- CLEVR dataset, Det setting (i.e., using detected scene graphs). Codes are in the directory `./exp_clevr_detected`.
- CLEVR dataset, GT setting (i.e., using ground truth scene graphs), attention is computed by softmax function over the label space. Codes are in `./exp_clevr_gt_softmax`.
- CLEVR dataset, GT setting, attention is computed by sigmoid function. Codes are in `./exp_clevr_gt_sigmoid`.
- VQA2.0 dataset, detected scene graphs. Codes are in `./exp_vqa`.We have a **separate README for each experiment setting** as an instruction to reimplement our reported results.
Feel free to contact me if you have any problems: [email protected]## Acknowledgement
- We refer to the repo [clevr-iep](https://github.com/facebookresearch/clevr-iep) for preprocessing codes.
- Our implementations of module model and dataloader are based on [tbd-net](https://github.com/davidmascharka/tbd-nets).
- Our stacked neural module implementation in `./exp_vqa` is based on Hu's [StackNMN](https://github.com/ronghanghu/snmn).