Ecosyste.ms: Awesome
An open API service indexing awesome lists of open source software.
https://github.com/facebookresearch/simmc
With the aim of building next generation virtual assistants that can handle multimodal inputs and perform multimodal actions, we introduce two new datasets (both in the virtual shopping domain), the annotation schema, the core technical tasks, and the baseline models. The code for the baselines and the datasets will be opensourced.
https://github.com/facebookresearch/simmc
Last synced: 8 days ago
JSON representation
With the aim of building next generation virtual assistants that can handle multimodal inputs and perform multimodal actions, we introduce two new datasets (both in the virtual shopping domain), the annotation schema, the core technical tasks, and the baseline models. The code for the baselines and the datasets will be opensourced.
- Host: GitHub
- URL: https://github.com/facebookresearch/simmc
- Owner: facebookresearch
- License: other
- Created: 2020-06-11T19:21:40.000Z (over 4 years ago)
- Default Branch: main
- Last Pushed: 2023-10-21T16:24:39.000Z (about 1 year ago)
- Last Synced: 2024-04-09T03:08:17.130Z (7 months ago)
- Language: Python
- Size: 731 KB
- Stars: 130
- Watchers: 20
- Forks: 39
- Open Issues: 4
-
Metadata Files:
- Readme: README.md
- Contributing: CONTRIBUTING.md
- License: LICENSE
- Code of conduct: CODE_OF_CONDUCT.md
Awesome Lists containing this project
README
# Situated Interactive MultiModal Conversations (SIMMC) Challenge 2020
Welcome to the Situated Interactive Multimodal Conversations (SIMMC) Track for [DSTC9][dstc9] 2020.
The SIMMC challenge aims to lay the foundations for the real-world assistant agents that can handle multimodal inputs, and perform multimodal actions.
We thus focus on **task-oriented** dialogs that encompass a **situated** multimodal user context in the form of a co-observed image or virtual reality (VR) environment.
The context is **dynamically** updated on each turn based on the user input and the assistant action.
Our challenge focuses on our SIMMC datasets, both of which are shopping domains:
(a) furniture (grounded in a shared virtual environment) and,
(b) fashion (grounded in an evolving set of images).**Organizers**: Ahmad Beirami, Eunjoon Cho, Paul A. Crook, Ankita De, Alborz Geramifard, Satwik Kottur, Seungwhan Moon, Shivani Poddar, Rajen Subba
Example from SIMMC-Furniture Dataset### Latest News
* **[Apr 15, 2021]** Released screenshots for SIMMC-Furniture
([part 0][screenshot_link_0], [part 1][screenshot_link_1], [part 2][screenshot_link_2]).
Also released improved API calls with newer heuristics as SIMMC v1.2 ([PR][screenshot_pr]).
* **[Dec 29, 2020]** Fixed the errors in text spans for both SIMMC-Furniture and SIMMC-Fashion, released new JSON files as SIMMC v1.1 ([PR][span_fix_pr]).
* **[Sept 28, 2020]** Test-Std data released, End of Challenge Phase 1.
* **[July 8, 2020]** Evaluation scripts and code to train baselines for Sub-Task #1, Sub-Task #2 released.
* **[June 22, 2020]** Challenge announcement. Training / development datasets (SIMMC v1.0) are released.**Note:** DSTC9 SIMMC Challenge was conducted on SIMMC v1.0. Thus all the results and baseline performances are on SIMMC v1.0.
## Important Links
* [Task Description Paper][simmc_arxiv]
* [Challenge Registration](https://forms.gle/jdT79eBeySHVoa1QA)
* [Data Formats](data/README.md)
* **Baseline Details**: [MM Action Prediction](mm_action_prediction/README.md), [MM Response Generation](mm_response_generation/README.md), [MM-DST](mm_dst/README.md)
* [Challenge Instructions](#challenge-instructions)
* [Submission Instructions](SUBMISSION_INSTRUCTIONS.md)## Timeline
| **Date** | **Milestone** |
| :--: | :-- |
| June 22, 2020 | Training & development data released |
| Sept 28, 2020 | Test-Std data released, End of Challenge Phase 1 |
| Oct 5, 2020 | Entry submission deadline, End of Challenge Phase 2 |
| Oct 12, 2020 | [Final results announced](DSTC9_SIMMC_RESULTS.md) |## Track Description
### Tasks and Metrics
We present three sub-tasks primarily aimed at replicating human-assistant actions in order to enable rich and interactive shopping scenarios.
| Sub-Task #1 | [Multimodal Action Prediction](mm_action_prediction) |
|---------|---------------------------------------------------------------------------------------------------------------------------------------|
| Goal | To predict the correct Assistant API action(s) (classification) |
| Input | Current user utterance, Dialog context, Multimodal context |
| Output | Structural API (action & arguments) |
| Metrics | Action Accuracy, Attribute Accuracy, Action Perplexity || Sub-Task #2 | [Multimodal Dialog Response Generation & Retrieval](mm_response_generation) |
|---------|---------------------------------------------------------------------------------------------------------------------------------------|
| Goal | To generate Assistant responses or retrieve from a candidate pool |
| Input | Current user utterance, Dialog context, Multimodal context, (Ground-truth API Calls) |
| Output | Assistant response utterance |
| Metrics | Generation: BLEU-4, Retrieval: MRR, R@1, R@5, R@10, Mean Rank || Sub-Task #3 | [Multimodal Dialog State Tracking (MM-DST)](mm_dst) |
|---------|---------------------------------------------------------------------------------------------------------------------------------------|
| Goal | To track user belief states across multiple turns |
| Input | Current user utterance, Dialogue context, Multimodal context |
| Output | Belief state for current user utterance |
| Metrics | Slot F1, Intent F1 |Please check the [task input](./TASK_INPUTS.md) file for a full description of inputs
for each subtask.### Evaluation
For the DSTC9 SIMMC Track, we will do a two phase evaluation as follows.
**Challenge Period 1**:
Participants will evaluate the model performance on the provided `devtest` set.
At the end of Challenge Period 1 (Sept 28), we ask participants to submit their model prediction results and a link to their code repository.**Challenge Period 2**:
A `test-std` set will be released on Sept 28 for the participants who submitted the results for the Challenge Period 1.
We ask participants to submit their model predictions on the `test-std` set by Oct 5.
We will announce the final results and the winners on Oct 12.## Challenge Instructions
### (1) Challenge Registration
* Fill out [this form](https://forms.gle/jdT79eBeySHVoa1QA) to register at DSTC9. Check “**Track 4: Visually Grounded Dialog Track**” along with other tracks you are participating in.
### (2) Download Datasets and Code
* Irrespective of participation in the challenge, we'd like to encourge those interested in this dataset to complete this [optional survey](https://oculus.qualtrics.com/jfe/form/SV_1AlazoSV7iwepZH). This will also help us communicate any future updates on the codebase, the datasets, and the challenge track.
* Git clone our repository to download the datasets and the code. You may use the provided baselines as a starting point to develop your models.
```
$ git lfs install
$ git clone https://github.com/facebookresearch/simmc.git
```### (3) Reporting Results for Challenge Phase 1
* Submit your model prediction results on the `devtest` set, following the [submission instructions](./SUBMISSION_INSTRUCTIONS.md).
* We will release the `test-std` set (with ground-truth labels hidden) on Sept 28.### (4) Reporting Results for Challenge Phase 2
* Submit your model prediction results on the `test-std` set, following the [submission instructions](./SUBMISSION_INSTRUCTIONS.md).
* We will evaluate the participants’ model predictions using the same evaluation script for Phase 1, and announce the results.## Contact
### Questions related to SIMMC Track, Data, and Baselines
Please contact [email protected], or leave comments in the Github repository.### DSTC Mailing List
If you want to get the latest updates about DSTC9, join the [DSTC mailing list](https://groups.google.com/a/dstc.community/forum/#!forum/list/join).## Citations
If you want to publish experimental results with our datasets or use the baseline models, please cite the following articles:
```
@article{moon2020situated,
title={Situated and Interactive Multimodal Conversations},
author={Moon, Seungwhan and Kottur, Satwik and Crook, Paul A and De, Ankita and Poddar, Shivani and Levin, Theodore and Whitney, David and Difranco, Daniel and Beirami, Ahmad and Cho, Eunjoon and Subba, Rajen and Geramifard, Alborz},
journal={arXiv preprint arXiv:2006.01460},
year={2020}
}@article{crook2019simmc,
title={SIMMC: Situated Interactive Multi-Modal Conversational Data Collection And Evaluation Platform},
author={Crook, Paul A and Poddar, Shivani and De, Ankita and Shafi, Semir and Whitney, David and Geramifard, Alborz and Subba, Rajen},
journal={arXiv preprint arXiv:1911.02690},
year={2019}
}
```
**NOTE**: The [paper][simmc_arxiv] above describes in detail the datasets, the NLU/NLG/Coref annotations, and some of the baselines we provide in this challenge. The paper reports the results from an earlier version of the dataset and with different train-dev-test splits, hence the baseline performances on the challenge resources will be slightly different.## License
SIMMC is released under [CC-BY-NC-SA-4.0](https://creativecommons.org/licenses/by-nc-sa/4.0/legalcode), see [LICENSE](LICENSE) for details.
[dstc9]:https://sites.google.com/dstc.community/dstc9/home
[simmc_arxiv]:https://arxiv.org/abs/2006.01460
[screenshot_link_0]:./data/simmc_furniture/furniture_screenshots_part_0.zip
[screenshot_link_1]:./data/simmc_furniture/furniture_screenshots_part_1.zip
[screenshot_link_2]:./data/simmc_furniture/furniture_screenshots_part_2.zip
[span_fix_pr]:https://github.com/facebookresearch/simmc/pull/54
[screenshot_pr]:https://github.com/facebookresearch/simmc/pull/60