An open API service indexing awesome lists of open source software.

https://github.com/princeton-vl/spatialsense

An Adversarially Crowdsourced Benchmark for Spatial Relation Recognition
https://github.com/princeton-vl/spatialsense

computer-vision dataset iccv-2019 spatial-relation-recognition

Last synced: about 1 month ago
JSON representation

An Adversarially Crowdsourced Benchmark for Spatial Relation Recognition

Awesome Lists containing this project

README

        

# SpatialSense

[![Code style: black](https://img.shields.io/badge/code%20style-black-000000.svg)](https://github.com/psf/black)

![Samples](samples.jpg)

Dataset and code for the paper:

[SpatialSense: An Adversarially Crowdsourced Benchmark for Spatial Relation Recognition](https://arxiv.org/abs/1908.02660)
[Kaiyu Yang](https://yangky11.github.io/), [Olga Russakovsky](https://www.cs.princeton.edu/~olgarus/), and [Jia Deng](https://www.cs.princeton.edu/~jiadeng/)
International Conference on Computer Vision (ICCV) 2019

```bibtex
@inproceedings{yang2019spatialsense,
title={SpatialSense: An Adversarially Crowdsourced Benchmark for Spatial Relation Recognition},
author={Yang, Kaiyu and Russakovsky, Olga and Deng, Jia},
booktitle={International Conference on Computer Vision (ICCV)},
year={2019}
}
```

## Requirements

1. Download the SpatialSense dataset from [![DOI](https://zenodo.org/badge/DOI/10.5281/zenodo.8104370.svg)](https://doi.org/10.5281/zenodo.8104370) to the root of this repo. Unzip `images.tar.gz`.
1. Download and install [Miniconda Python 3](https://docs.conda.io/en/latest/miniconda.html) (Anaconda should also work).
1. Install Python dependencies using conda: `conda env create -f spatialsense.yaml && conda activate spatialsense`. If you have troubles with the aforementioned two steps, you may manually install the packages in [spatialsense.yaml](./spatialsense.yaml) in whatever way that works for you.
1. Download the pre-trained Word2Vec model [GoogleNews-vectors-negative300.bin.gz.](https://drive.google.com/file/d/0B7XkCwpI5KDYNlNUTTlSS21pQmM/edit?resourcekey=0-wjGZdNAUop6WykTtMip30g) to [./baselines](https://github.com/princeton-vl/SpatialSense/tree/master/baselines).

## Dataset

### Data Format

The `annotation.json` file contains a list in which each element contains the annotations for a single image.
For example, the first element is:

```
{
'url': 'https://farm4.staticflickr.com/3543/5704634119_8b8ccf3229.jpg', # URL for Flickr Images
'nsid': '10392797@N03', # Flickr NSID of the user
'height': 500,
'width': 281,
'split': 'train', # train/valid/test split
'annotations': [{ # a list of spatial relations
'_id': '59fbffe4f25c8070bb77ec42', # an unique identifier for the relation
'predicate': 'on',
'object': {
'y': 402, # (x, y) is a point on the object
'x': 148,
'name': 'ground',
'bbox': [196, 500, 3, 278] # bounding box
},
'subject': {
'y': 317,
'x': 157,
'name': 'cat',
'bbox': [230, 434, 31, 264]
},
'label': True # the relation is a positive example
}, {
'_id': '59ff0e910de0c80e4077c5f0',
'predicate': 'to the left of',
'object': {
'y': 213,
'x': 240,
'name': 'mirror',
'bbox': [0, 345, 160, 280]
},
'subject': {
'y': 303,
'x': 143,
'name': 'cat',
'bbox': [226, 449, 33, 271]
},
'label': True},
...
]}
```

**Conventions for coordinates and bounding boxes**: The origin is the upper-left corner of an image; the x-axis is along the width, and the y-axis is alone the height. A bounding box `[y0, y1, x0, x1]` has `(x0, y0)` as its upper-left corner and `(x1, y1)` as its bottom-right corner.

### Visualizations

To visualize the relations in SpatialSense: `python visualize.py`
Run `python visualize.py --help` to see the options.

## Baselines

Assuming you are in the `./baselines` directory, below are instructions for reproducing the baselines in the paper.

### Language-only

Training: `python main_L.py --train_split train_valid --exp_id language-only`

Predictions and model checkpoints will be saved in `./runs/language-only`.

### 2D-only

Training: `NO_WORD2VEC=1, python main_S.py --train_split train_valid --exp_id 2d-only`

Predictions and model checkpoints will be saved in `./runs/2d-only`.

### Vip-CNN

Training: `python main.py --train_split train_valid --exp_id vipcnn --model vipcnn --learning_rate 4e-4 --l2 5e-7 --n_epochs 40 --batchsize 16 --patience 18`

Predictions and model checkpoints will be saved in `./runs/vipcnn`.

### Peyre et al.

1. Extract the spatial features: `python unrel/spatial_features.py`
2. Extract the appearance features: `python unrel/appearance_features.py`
3. Train and test the model `python unrel/train.py --spatial --appr --no_val`

### PPR-FCN

Training: `python main.py --train_split train_valid --exp_id pprfcn --model pprfcn --backbone resnet101 --learning_rate 3e-4 --l2 6e-7 --batchsize 7 --patience 14`

Predictions and model checkpoints will be saved in `./runs/pprfcn`.

### DRNet

Training: `python main.py --train_split train_valid --exp_id drnet --model drnet --learning_rate 1.5e-4 --l2 2.8e-4`

Predictions and model checkpoints will be saved in `./runs/drnet`.

### VTransE

Training: `python main.py --train_split train_valid --exp_id vtranse --model vtranse --learning_rate 6e-4 --l2 3e-4 --feature_dim 128`

Predictions and model checkpoints will be saved in `./runs/vtranse`.

## Spatial Relations in 3D

In a more recent [paper](https://arxiv.org/abs/2012.01634), we constructed [**Rel3D**](https://github.com/princeton-vl/Rel3D): the first large-scale, human-annotated dataset for grounding spatial relations in 3D. It enables quantifying the effectiveness of 3D information in predicting spatial relations. Moreover, we propose minimally contrastive data collection—a novel crowdsourcing method for reducing dataset bias. The examples in Rel3D come in **minimally contrastive pairs**: two examples in a pair are almost identical but have different labels.

![Rel3D](./rel3d.gif)