Ecosyste.ms: Awesome
An open API service indexing awesome lists of open source software.
https://github.com/opendilab/awesome-multi-modal-reinforcement-learning
A curated list of Multi-Modal Reinforcement Learning resources (continually updated)
https://github.com/opendilab/awesome-multi-modal-reinforcement-learning
List: awesome-multi-modal-reinforcement-learning
Last synced: about 1 month ago
JSON representation
A curated list of Multi-Modal Reinforcement Learning resources (continually updated)
- Host: GitHub
- URL: https://github.com/opendilab/awesome-multi-modal-reinforcement-learning
- Owner: opendilab
- License: apache-2.0
- Created: 2022-09-01T05:47:26.000Z (over 2 years ago)
- Default Branch: main
- Last Pushed: 2024-05-10T06:05:22.000Z (8 months ago)
- Last Synced: 2024-05-21T18:04:57.741Z (7 months ago)
- Homepage:
- Size: 32.2 KB
- Stars: 324
- Watchers: 8
- Forks: 10
- Open Issues: 0
-
Metadata Files:
- Readme: README.md
- Contributing: CONTRIBUTING.md
- License: LICENSE
Awesome Lists containing this project
- ultimate-awesome - awesome-multi-modal-reinforcement-learning - A curated list of Multi-Modal Reinforcement Learning resources (continually updated). (Other Lists / Monkey C Lists)
README
# Awesome Multi-Modal Reinforcement Learning
[![Awesome](https://cdn.rawgit.com/sindresorhus/awesome/d7305f38d29fed78fa85652e3a63e154dd8e8829/media/badge.svg)](https://github.com/sindresorhus/awesome)
![visitor badge](https://visitor-badge.lithub.cc/badge?page_id=opendilab.awesome-multi-modal-reinforcement-learning&left_text=Visitors)
[![docs](https://img.shields.io/badge/docs-latest-blue)](https://github.com/opendilab/awesome-multi-modal-reinforcement-learning)
![GitHub stars](https://img.shields.io/github/stars/opendilab/awesome-multi-modal-reinforcement-learning?color=yellow)
![GitHub forks](https://img.shields.io/github/forks/opendilab/awesome-multi-modal-reinforcement-learning?color=9cf)
[![GitHub license](https://img.shields.io/github/license/opendilab/awesome-multi-modal-reinforcement-learning)](https://github.com/opendilab/awesome-multi-modal-reinforcement-learning/blob/main/LICENSE)This is a collection of research papers for **Multi-Modal reinforcement learning (MMRL)**.
And the repository will be continuously updated to track the frontier of MMRL.
Some papers may not be relevant to RL, but we include them anyway as they may be useful for the research of MMRL.Welcome to follow and star!
## Introduction
Multi-Modal RL agents focus on learning from video (images), language (text), or both, as humans do. We believe that it is important for intelligent agents to learn directly from images or text, since such data can be easily obtained from the Internet.
![飞书20220922-161353](https://user-images.githubusercontent.com/4834562/191696555-2ff17e41-77f4-4d04-ba2a-ea9bc8d99d96.png)
## Table of Contents
- [Awesome Multi-Modal Reinforcement Learning](#awesome-multi-modal-reinforcement-learning)
- [Introduction](#introduction)
- [Table of Contents](#table-of-contents)
- [Papers](#papers)
- [ICLR 2024](#iclr-2024)
- [ICLR 2023](#iclr-2023)
- [ICLR 2022](#iclr-2022)
- [ICLR 2021](#iclr-2021)
- [ICLR 2019](#iclr-2019)
- [NeurIPS 2023](#neurips-2023)
- [NeurIPS 2022](#neurips-2022)
- [NeurIPS 2021](#neurips-2021)
- [NeurIPS 2018](#neurips-2018)
- [ICML 2024](#icml-2024)
- [ICML 2022](#icml-2022)
- [ICML 2019](#icml-2019)
- [ICML 2017](#icml-2017)
- [CVPR 2024](#cvpr-2024)
- [CVPR 2022](#cvpr-2022)
- [CoRL 2022](#corl-2022)
- [Other](#other)
- [ArXiv](#arxiv)
- [Contributing](#contributing)
- [License](#license)## Papers
```
format:
- [title](paper link) [links]
- authors.
- key words.
- experiment environment.
```### ICLR 2024
- [DrM: Mastering Visual Reinforcement Learning through Dormant Ratio Minimization](https://openreview.net/pdf?id=MSe8YFbhUE)
- Guowei Xu, Ruijie Zheng, Yongyuan Liang, Xiyao Wang, Zhecheng Yuan, Tianying Ji, Yu Luo, Xiaoyu Liu, Jiaxin Yuan, Pu Hua, Shuzhen Li, Yanjie Ze, Hal Daumé III, Furong Huang, Huazhe Xu
- Keyword: Visual RL; Dormant Ratio
- ExpEnv: [DeepMind Control Suite](https://github.com/deepmind/dm_control),[Meta-world](https://github.com/rlworkgroup/metaworld),[Adroit](https://github.com/Farama-Foundation/D4RL)- [Revisiting Data Augmentation in Deep Reinforcement Learning](https://openreview.net/pdf?id=EGQBpkIEuu)
- Jianshu Hu, Yunpeng Jiang, Paul Weng
- Keyword: Reinforcement Learning, Data Augmentation
- ExpEnv: [DeepMind Control Suite](https://github.com/deepmind/dm_control)- [Revisiting Plasticity in Visual Reinforcement Learning: Data, Modules and Training Stages](https://openreview.net/forum?id=0aR1s9YxoL)
- Guozheng Ma, Lu Li, Sen Zhang, Zixuan Liu, Zhen Wang, Yixin Chen, Li Shen, Xueqian Wang, Dacheng Tao
- Keyword: Plasticity, Visual Reinforcement Learning, Deep Reinforcement Learning, Sample Efficiency
- ExpEnv: [DeepMind Control Suite](https://github.com/deepmind/dm_control),[Atari](https://github.com/openai/gym)- [Entity-Centric Reinforcement Learning for Object Manipulation from Pixels](https://openreview.net/forum?id=uDxeSZ1wdI)
- Dan Haramati, Tal Daniel, Aviv Tamar
- Keyword: deep reinforcement learning, visual reinforcement learning, object-centric, robotic object manipulation, compositional generalization
- ExpEnv: [IsaacGym](https://github.com/NVIDIA-Omniverse/IsaacGymEnvs)### ICLR 2023
- [PaLI: A Jointly-Scaled Multilingual Language-Image Model](https://arxiv.org/abs/2209.06794)(**notable top 5%**)
- Xi Chen, Xiao Wang, Soravit Changpinyo, AJ Piergiovanni, Piotr Padlewski, Daniel Salz, Sebastian Goodman, Adam Grycner, Basil Mustafa, Lucas Beyer, Alexander Kolesnikov, Joan Puigcerver, Nan Ding, Keran Rong, Hassan Akbari, Gaurav Mishra, Linting Xue, Ashish Thapliyal, James Bradbury, Weicheng Kuo, Mojtaba Seyedhosseini, Chao Jia, Burcu Karagol Ayan, Carlos Riquelme, Andreas Steiner, Anelia Angelova, Xiaohua Zhai, Neil Houlsby, Radu Soricut
- Keyword: amazing zero-shot, language component and visual component
- ExpEnv: None- [VIMA: General Robot Manipulation with Multimodal Prompts](https://arxiv.org/abs/2210.03094)
- Yunfan Jiang, Agrim Gupta, Zichen Zhang, Guanzhi Wang, Yongqiang Dou, Yanjun Chen, Li Fei-Fei, Anima Anandkumar, Yuke Zhu, Linxi Fan. *NeurIPS Workshop 2022*
- Key Words: multimodal prompts, transformer-based generalist agent model, large-scale benchmark
- ExpEnv: [VIMA-Bench](https://github.com/vimalabs/VimaBench), [VIMA-Data](https://huggingface.co/datasets/VIMA/VIMA-Data)- [MIND ’S EYE: GROUNDED LANGUAGE MODEL REASONING THROUGH SIMULATION](https://arxiv.org/abs/2210.05359)
- Ruibo Liu, Jason Wei, Shixiang Shane Gu, Te-Yen Wu, Soroush Vosoughi, Claire Cui, Denny Zhou, Andrew M. Dai
- Keyword: language2physical-world, reasoning ability
- ExpEnv: [MuJoCo](https://mujoco.org/)### ICLR 2022
- [How Much Can CLIP Benefit Vision-and-Language Tasks?](https://openreview.net/forum?id=zf_Ll3HZWgy)
- Sheng Shen, Liunian Harold Li, Hao Tan, etc. *ICLR 2022*
- Key Words: Vision-and-Language, CLIP
- ExpEnv: None### ICLR 2021
- [Grounding Language to Entities and Dynamics for Generalization in Reinforcement Learning](https://arxiv.org/abs/2101.07393)
- Austin W. Hanjie, Victor Zhong, Karthik Narasimhan. *ICML 2021*
- Key Words: Multi-modal Attention
- ExpEnv: [Messenger](https://github.com/ahjwang/messenger-emma)- [Mastering Atari with Discrete World Models](https://arxiv.org/abs/2010.02193)
- Danijar Hafner, Timothy Lillicrap, Mohammad Norouzi, etc.
- Key Words: World models
- ExpEnv: [Atari](https://github.com/openai/gym)- [Decoupling Representation Learning from Reinforcement Learning](https://arxiv.org/abs/2009.08319)
- Adam Stooke,Kimin Lee,Pieter Abbeel, etc.
- Key Words: representation learning, unsupervised learning
- ExpEnv: [DeepMind Control](https://github.com/deepmind/dm_control), [Atari](https://github.com/openai/gym), [DMLab](https://github.com/deepmind/lab)### ICLR 2019
- [Learning Actionable Representations with Goal-Conditioned Policies](https://arxiv.org/abs/1811.07819)
- Dibya Ghosh, Abhishek Gupta, Sergey Levine.
- Key Words: Actionable Representations Learning
- ExpEnv: 2D navigation(2D Wall, 2D Rooms, Wheeled, Wheeled Rooms, Ant, Pushing)### NeurIPS 2023
- [Inverse Dynamics Pretraining Learns Good Representations for Multitask Imitation](https://openreview.net/pdf?id=kjMGHTo8Cs)
- David Brandfonbrener, Ofir Nachum, Joan Bruna
- Key Words: representation learning, imitation learning
- ExpEnv: [Sawyer Door Open](https://github.com/suraj-nair-1/metaworld), [MetaWorld](https://github.com/suraj-nair-1/metaworld), [Franka Kitchen, Adroit](https://github.com/aravindr93/mjrl)- [Frequency-Enhanced Data Augmentation for Vision-and-Language Navigation](https://proceedings.neurips.cc/paper_files/paper/2023/hash/0d9e08f247ca7fbbfd5e50b7ff9cf357-Abstract-Conference.html)
- Keji He, Chenyang Si, Zhihe Lu, Yan Huang, Liang Wang, Xinchao Wang
- Key Words: Vision-and-Language Navigation, High-Frequency, Data Augmentation
- ExpEnv: [Matterport3d](https://niessner.github.io/Matterport/)- [Language Is Not All You Need: Aligning Perception with Language Models](https://arxiv.org/pdf/2302.14045)
- Shaohan Huang, Li Dong, Wenhui Wang, Yaru Hao, etc.
- Key Words: Multimodal Perception, World Modeling
- ExpEnv: [IQ50](https://aka.ms/kosmos-iq50)- [MotionGPT: Human Motion as a Foreign Language](https://proceedings.neurips.cc/paper_files/paper/2023/file/3fbf0c1ea0716c03dea93bb6be78dd6f-Paper-Conference.pdf)
- Biao Jiang, Xin Chen, Wen Liu, Jingyi Yu, Gang Yu, Tao Chen
- Key Words: Human motion, text-driven motion generation
- ExpEnv: [HumanML3D](https://ericguo5513.github.io/text-to-motion),[KIT](https://motion-database.humanoids.kit.edu/)- [Large Language Models are Visual Reasoning Coordinators](https://proceedings.neurips.cc/paper_files/paper/2023/file/ddfe6bae7b869e819f842753009b94ad-Paper-Conference.pdf)
- Liangyu Chen, Bo Li, Sheng Shen, Jingkang Yang, Chunyuan Li, Kurt Keutzer, Trevor Darrell, Ziwei Liu
- Key Words: Visual Reasoning, Large Language Model
- ExpEnv: [A-OKVQA](), [OK-VQA](), [e-SNLI-VE](), [VSR]()### NeurIPS 2022
- [MineDojo: Building Open-Ended Embodied Agents with Internet-Scale Knowledge](https://arxiv.org/abs/2206.08853)
- Linxi Fan, Guanzhi Wang, Yunfan Jiang, etc.
- Key Words: multimodal dataset, MineClip
- ExpEnv: [Minecraft](https://minedojo.org/)- [Video PreTraining (VPT): Learning to Act by Watching Unlabeled Online Videos](https://arxiv.org/abs/2206.11795)
- Bowen Baker, Ilge Akkaya, Peter Zhokhov, etc.
- Key Words: Inverse Dynamics Model
- ExpEnv: [minerl](https://github.com/minerllabs/minerl)### NeurIPS 2021
- [SOAT: A Scene-and Object-Aware Transformer for Vision-and-Language Navigation](https://arxiv.org/pdf/2110.14143.pdf)
- Abhinav Moudgil, Arjun Majumdar,Harsh Agrawal, etc.
- Key Words: Vision-and-Language Navigation
- ExpEnv: [Room-to-Room](https://paperswithcode.com/dataset/room-to-room), [Room-Across-Room](https://github.com/google-research-datasets/RxR)- [Pretraining Representations for Data-Efficient Reinforcement Learning](https://papers.nips.cc/paper/2021/hash/69eba34671b3ef1ef38ee85caae6b2a1-Abstract.html)
- Max Schwarzer, Nitarshan Rajkumar, Michael Noukhovitch, etc.
- Key Words: latent dynamics modelling, unsupervised RL
- ExpEnv: [Atari](https://github.com/openai/gym)### NeurIPS 2018
- [Recurrent World Models Facilitate Policy Evolution](https://papers.nips.cc/paper/2018/hash/2de5d16682c3c35007e4e92982f1a2ba-Abstract.html)
- David Ha, Jürgen Schmidhuber.
- Key Words: World model, generative RNN, VAE
- ExpEnv: [VizDoom](https://github.com/mwydmuch/ViZDoom), [CarRacing](https://github.com/openai/gym)### ICML 2024
- [Investigating Pre-Training Objectives for Generalization in Vision-Based Reinforcement Learning](https://proceedings.mlr.press/v235/kim24u.html)
- Donghu Kim, Hojoon Lee, Kyungmin Lee, Dongyoon Hwang, Jaegul Choo
- Key Words: vision-based RL
- ExpEnv: [Atari](https://github.com/openai/gym)- [RL-VLM-F: Reinforcement Learning from Vision Language Foundation Model Feedback](https://proceedings.mlr.press/v235/wang24bn.html)
- Yufei Wang, Zhanyi Sun, Jesse Zhang, Zhou Xian, Erdem Biyik, David Held, Zackory Erickson
- Key Words: learning from VLM
- ExpEnv: [Gym](), [MetaWorld](https://github.com/suraj-nair-1/metaworld)- [Reward Shaping for Reinforcement Learning with An Assistant Reward Agent](https://proceedings.mlr.press/v235/ma24l.html)
- Haozhe Ma, Kuankuan Sima, Thanh Vinh Vo, Di Fu, Tze-Yun Leong
- Key Words: dual-agent reward shaping framework
- ExpEnv: [Mujoco](https://github.com/google-deepmind/mujoco)- [FuRL: Visual-Language Models as Fuzzy Rewards for Reinforcement Learning](https://proceedings.mlr.press/v235/fu24j.html)
- Yuwei Fu, Haichao Zhang, Di Wu, Wei Xu, Benoit Boulet
- Key Words: high-dimensional observations, representation learning for RL
- ExpEnv: [MetaWorld](https://github.com/suraj-nair-1/metaworld)- [Rich-Observation Reinforcement Learning with Continuous Latent Dynamics](https://proceedings.mlr.press/v235/song24i.html)
- Yuda Song, Lili Wu, Dylan J Foster, Akshay Krishnamurthy
- Key Words: VLM as reward function
- ExpEnv: [maze]()- [LLM-Empowered State Representation for Reinforcement Learning](https://proceedings.mlr.press/v235/wang24bh.html)
- Boyuan Wang, Yun Qu, Yuhang Jiang, Jianzhun Shao, Chang Liu, Wenming Yang, Xiangyang Ji
- Key Words: LLM-based state representation
- ExpEnv: [Mujoco](https://github.com/google-deepmind/mujoco)- [Code as Reward: Empowering Reinforcement Learning with VLMs](https://proceedings.mlr.press/v235/venuto24a.html)
- David Venuto, Mohammad Sami Nur Islam, Martin Klissarov, etc.
- Key Words: Vision-Language Models, reward functions
- ExpEnv: [MiniGrid](https://minigrid.farama.org/)### ICML 2022
- [Language Models as Zero-Shot Planners: Extracting Actionable Knowledge for Embodied Agents](https://arxiv.org/pdf/2201.07207.pdf)
- Wenlong Huang, Pieter Abbeel, Deepak Pathak, etc.
- Key Words: large language models, Embodied Agents
- ExpEnv: [VirtualHome](https://github.com/xavierpuigf/virtualhome)- [Reinforcement Learning with Action-Free Pre-Training from Videos](https://proceedings.mlr.press/v162/seo22a.html)
- Younggyo Seo, Kimin Lee, Stephen L James, etc.
- Key Words: action-free pretraining, videos
- ExpEnv: [Meta-world](https://github.com/rlworkgroup/metaworld), [DeepMind Control Suite](https://github.com/deepmind/dm_control)- [History Compression via Language Models in Reinforcement Learning](https://arxiv.org/abs/2205.12258)
- Fabian Paischer, Thomas Adler, Vihang Patil, etc.
- Key Words: Pretrained Language Transformer
- ExpEnv: [Minigrid](https://github.com/maximecb/gym-minigrid), [Procgen](https://github.com/openai/procgen)### ICML 2019
- [Learning Latent Dynamics for Planning from Pixels](https://arxiv.org/abs/1811.04551)
- Danijar Hafner, Timothy Lillicrap, Ian Fischer, etc.
- Key Words: latent dynamics model, pixel observations
- ExpEnv: [DeepMind Control Suite](https://github.com/deepmind/dm_control)### ICML 2017
- [Zero-Shot Task Generalization with Multi-Task Deep Reinforcement Learning](https://arxiv.org/abs/1706.05064)
- Junhyuk Oh, Satinder Singh, Honglak Lee, Pushmeet Kohli
- Key Words: unseen instruction, sequential instruction
- ExpEnv: [Minecraft](https://arxiv.org/abs/1605.09128)### CVPR 2024
- [DMR: Decomposed Multi-Modality Representations for Frames and Events Fusion in Visual Reinforcement Learning](https://openaccess.thecvf.com/content/CVPR2024/html/Xu_DMR_Decomposed_Multi-Modality_Representations_for_Frames_and_Events_Fusion_in_CVPR_2024_paper.html)
- Haoran Xu, Peixi Peng, Guang Tan, Yuan Li, Xinhai Xu, Yonghong Tian
- Key Words: Visual Reinforcement Learning, Multi-Modality Representation, Dynamic Vision Sensor
- ExpEnv: [Carla]()- [Vision-and-Language Navigation via Causal Learning](https://arxiv.org/abs/2404.10241)
- Liuyi Wang, Zongtao He, Ronghao Dang, Mengjiao Shen, Chengju Liu, Qijun Chen
- Key Words: vision-and-language navigation, cross-modal causal transformer
- ExpEnv: [R2R](https://bringmeaspoon.org/) [REVERIE](https://github.com/google-research-datasets/RxR) [RxR-English](https://github.com/google-research-datasets/RxR) [SOON]()### CVPR 2022
- [End-to-end Generative Pretraining for Multimodal Video Captioning](https://arxiv.org/abs/2201.08264)
- Paul Hongsuck Seo, Arsha Nagrani, Anurag Arnab, Cordelia Schmid
- Key Words: Multimodal video captioning, Pretraining using a future utterance, Multimodal Video Generative Pretraining
- ExpEnv: [HowTo100M](https://arxiv.org/abs/1906.03327)- [Image as a Foreign Language: BEiT Pretraining for All Vision and Vision-Language Tasks](https://arxiv.org/abs/2208.10442)
- Wenhui Wang, Hangbo Bao, Li Dong, Johan Bjorck, Zhiliang Peng, Qiang Liu, Kriti Aggarwal, Owais Khan Mohammed, Saksham Singhal, Subhojit Som, Furu Wei
- Key Words: backbone architecture, pretraining task, model scaling up
- ExpEnv: [ADE20K](https://groups.csail.mit.edu/vision/datasets/ADE20K/), [COCO](https://cocodataset.org/), [NLVR2](https://paperswithcode.com/dataset/nlvr), [Flickr30K](https://paperswithcode.com/dataset/flickr30k)- [Think Global, Act Local: Dual-scale Graph Transformer for Vision-and-Language Navigation](https://arxiv.org/abs/2202.11742)
- Shizhe Chen, Pierre-Louis Guhur, Makarand Tapaswi, Cordelia Schmid, Ivan Laptev
- Keyword: dual-scale graph transformer, dual-scale graph transformer, affordance detection
- ExpEnv: None- [Masked Visual Pre-training for Motor Control](https://arxiv.org/abs/2203.06173)
- Tete Xiao, Ilija Radosavovic, Trevor Darrell, etc. *ArXiv 2022*
- Key Words: self-supervised learning, motor control
- ExpEnv: [Isaac Gym](https://developer.nvidia.com/isaac-gym)### CoRL 2022
- [LM-Nav: Robotic Navigation with Large Pre-Trained Models of Language, Vision, and Action](https://arxiv.org/abs/2207.04429)
- Dhruv Shah, Blazej Osinski, Brian Ichter, Sergey Levine
- Key Words: robotic navigation, goal-conditioned, unannotated large dataset, CLIP, ViNG, GPT-3
- ExpEnv: None- [Real-World Robot Learning with Masked Visual Pre-training](https://arxiv.org/abs/2210.03109)
- Ilija Radosavovic, Tete Xiao, Stephen James, Pieter Abbeel, Jitendra Malik, Trevor Darrell
- Key Words: real-world robotic tasks,
- ExpEnv: None- [R3M: A Universal Visual Representation for Robot Manipulation](https://arxiv.org/abs/2203.12601)
- Suraj Nair, Aravind Rajeswaran, Vikash Kumar, etc.
- Key Words: Ego4D human video dataset, pre-train visual representation
- ExpEnv: [MetaWorld](https://github.com/suraj-nair-1/metaworld), [Franka Kitchen, Adroit](https://github.com/aravindr93/mjrl)### Other
- [RL-EMO: A Reinforcement Learning Framework for Multimodal Emotion Recognition](https://arxiv.org/abs/2005.07648) *ICASSP 2024*
- Chengwen Zhang, Yuhao Zhang, Bo Cheng
- Keyword: Multimodal Emotion Recognition, Reinforcement Learning, Graph Convolution Network
- ExpEnv: None- [Language Conditioned Imitation Learning over Unstructured Data](https://arxiv.org/abs/2005.07648) *RSS 2021*
- Corey Lynch, Pierre Sermanet
- Keyword: open-world environments
- ExpEnv: None- [Learning Generalizable Robotic Reward Functions from “In-The-Wild” Human Videos](https://arxiv.org/abs/2103.16817) *RSS 2021*
- Annie S. Chen, Suraj Nair, Chelsea Finn.
- Key Words: Reward Functions, “In-The-Wild” Human Videos
- ExpEnv: None- [Offline Reinforcement Learning from Images with Latent Space Models](https://proceedings.mlr.press/v144/rafailov21a.html) *L4DC 2021*
- Rafael Rafailov, Tianhe Yu, Aravind Rajeswaran, etc.
- Key Words: Latent Space Models
- ExpEnv: [DeepMind Control](https://github.com/deepmind/dm_control), [Adroit Pen](https://github.com/Farama-Foundation/D4RL), [Sawyer Door Open](https://github.com/suraj-nair-1/metaworld), [Robel D’Claw Screw](https://github.com/google-research/robel)- [Is Cross-Attention Preferable to Self-Attention for Multi-Modal Emotion Recognition?](https://arxiv.org/abs/2202.09263) *ICASSP 2022*
- Vandana Rajan, Alessio Brutti, Andrea Cavallaro.
- Key Words: Multi-Modal Emotion Recognition, Cross-Attention
- ExpEnv: None### ArXiv
- [Spatialvlm: Endowing vision-language models with spatial reasoning capabilities](https://arxiv.org/pdf/2401.12168)
- Boyuan Chen, Zhuo Xu, Sean Kirmani, Brian Ichter, Danny Driess, Pete Florence, Dorsa Sadigh, Leonidas Guibas, Fei Xia
- Key Words: Visual Question Answering, 3D Spatial Reasoning
- ExpEnv: spatial VQA dataset- [M2CURL: Sample-Efficient Multimodal Reinforcement Learning via Self-Supervised Representation Learning for Robotic Manipulation ](https://browse.arxiv.org/abs/2401.17032)
- Fotios Lygerakis, Vedant Dave, Elmar Rueckert
- Key Words: Robotic Manipulation, Self-supervised representation
- ExpEnv: Gym- [On Time-Indexing as Inductive Bias in Deep RL for Sequential Manipulation Tasks](https://arxiv.org/abs/2401.01993)
- M. Nomaan Qureshi, Ben Eisner, David Held
- Key Words: Multimodality of policy output, Action head switching
- ExpEnv: MetaWorld- [Parameterized Decision-making with Multi-modal Perception for Autonomous Driving](https://arxiv.org/abs/2312.11935)
- Yuyang Xia, Shuncheng Liu, Quanlin Yu, Liwei Deng, You Zhang, Han Su, Kai Zheng
- Key Words: Autonomous driving, GNN in RL
- ExpEnv: CARLA- [A Contextualized Real-Time Multimodal Emotion Recognition for Conversational Agents using Graph Convolutional Networks in Reinforcement Learning](https://arxiv.org/abs/2309.15683)
- Fathima Abdul Rahman, Guang Lu
- Key Words: Emotion Recognition, GNN in RL
- ExpEnv: IEMOCAP- [Reinforced UI Instruction Grounding: Towards a Generic UI Task Automation API](https://arxiv.org/abs/2310.04716)
- Zhizheng Zhang, Wenxuan Xie, Xiaoyi Zhang, Yan Lu
- Key Words: LLM, generic UI task automation API
- ExpEnv: RicoSCA, MoTIF- [Driving with LLMs: Fusing Object-Level Vector Modality for Explainable Autonomous Driving](https://arxiv.org/abs/2310.01957)
- Long Chen, Oleg Sinavski, Jan Hünermann, Alice Karnsund, Andrew James Willmott, Danny Birch, Daniel Maund, Jamie Shotton
- Key Words: LLM in Autonomous Driving, object-level multimodal LLM
- ExpEnv: RicoSCA, MoTIF- [Nonprehensile Planar Manipulation through Reinforcement Learning with Multimodal Categorical Exploration ](https://arxiv.org/abs/2308.02459)
- Juan Del Aguila Ferrandis, João Moura, Sethu Vijayakumar
- Key Words: multimodal exploration approach
- ExpEnv: KUKA iiwa robot arm- [End-to-End Streaming Video Temporal Action Segmentation with Reinforce Learning](https://arxiv.org/abs/2309.15683)
- Wujun Wen, Jinrong Zhang, Shenglan Liu, Yunheng Li, Qifeng Li, Lin Feng
- Key Words: Temporal Action Segmentation, RL in Video Analysis
- ExpEnv: EGTEA- [Do as I can, not as I get:Topology-aware multi-hop reasoningon multi-modal knowledge graphs](https://arxiv.org/abs/2306.10345)
- Shangfei Zheng, Hongzhi Yin, Tong Chen, Quoc Viet Hung Nguyen, Wei Chen, Lei Zhao
- Key Words: Multi-hop reasoning, multi-modal knowledge graphs, inductive setting, adaptive reinforcement learning
- ExpEnv: None- [Multimodal Reinforcement Learning for Robots Collaborating with Humans](https://arxiv.org/abs/2303.07265)
- Afagh Mehri Shervedani, Siyu Li, Natawut Monaikul, Bahareh Abbasi, Barbara Di Eugenio, Milos Zefran
- Key Words: robust and deliberate decisions, end-to-end training, importance enhancement, similarity, improve IRL training process multimodal RL domains
- ExpEnv: None- [See, Plan, Predict: Language-guided Cognitive Planning with Video Prediction](https://arxiv.org/abs/2210.03825v1)
- Maria Attarian, Advaya Gupta, Ziyi Zhou, Wei Yu, Igor Gilitschenski, Animesh Garg
- Keyword: cognitive planning, language-guided video prediction
- ExpEnv: None- [Open-vocabulary Queryable Scene Representations for Real World Planning](https://arxiv.org/abs/2209.09874)
- Boyuan Chen, Fei Xia, Brian Ichter, Kanishka Rao, Keerthana Gopalakrishnan, Michael S. Ryoo, Austin Stone, Daniel Kappler
- Key Words: Target Detection, Real World, Robotic Tasks
- ExpEnv: [Say Can](https://say-can.github.io/)- [Do As I Can, Not As I Say: Grounding Language in Robotic Affordances](https://arxiv.org/abs/2204.01691)
- Michael Ahn, Anthony Brohan, Noah Brown, Yevgen Chebotar, Omar Cortes, Byron David, Chelsea Finn, Chuyuan Fu, Keerthana Gopalakrishnan, Karol Hausman, Alex Herzog, Daniel Ho, Jasmine Hsu, Julian Ibarz, Brian Ichter, Alex Irpan, Eric Jang, Rosario Jauregui Ruano, Kyle Jeffrey, Sally Jesmonth, Nikhil J Joshi, Ryan Julian, Dmitry Kalashnikov, Yuheng Kuang, Kuang-Huei Lee, Sergey Levine, Yao Lu, Linda Luu, Carolina Parada, Peter Pastor, Jornell Quiambao, Kanishka Rao, Jarek Rettinghouse, Diego Reyes, Pierre Sermanet, Nicolas Sievers, Clayton Tan, Alexander Toshev, Vincent Vanhoucke, Fei Xia, Ted Xiao, Peng Xu, Sichun Xu, Mengyuan Yan, Andy Zeng
- Key Words: real world, natural language
- ExpEnv: [Say Can](https://say-can.github.io/)## Contributing
Our purpose is to make this repo even better. If you are interested in contributing, please refer to [HERE](CONTRIBUTING.md) for instructions in contribution.
## License
Awesome Multi-Modal Reinforcement Learning is released under the Apache 2.0 license.