Ecosyste.ms: Awesome
An open API service indexing awesome lists of open source software.
awesome-InteractiveNLP-papers
Paper List for a new paradigm of NLP: Interactive NLP (https://arxiv.org/abs/2305.13246) :fire:
https://github.com/InteractiveNLP-Team/awesome-InteractiveNLP-papers
Last synced: 6 days ago
JSON representation
-
Related Projects
-
Paper
-
👨💼Human-LM Interaction
- Deep reinforcement learning from human preferences - Reinforcement%20Learning-orange) ![img](https://img.shields.io/badge/Feedback-green)
- Constitutional Al:learning from ai feedback - Constitutional%20AI-blue) ![img](https://img.shields.io/badge/-Reinforcement%20Learning-orange) ![img](https://img.shields.io/badge/Feedback-green)
- LaMP: When Large Language Models Meet Personalization - LaMP-blue) ![img](https://img.shields.io/badge/Personalization-green)
- Training language models to follow instructions with human feedback - InstructGPT-blue) ![img](https://img.shields.io/badge/-Reinforcement%20Learning-orange) ![img](https://img.shields.io/badge/Feedback-green)
- Improving alignment of dialogue agents via targeted human judgements - Sparrow-blue) ![img](https://img.shields.io/badge/-Reinforcement%20Learning-orange) ![img](https://img.shields.io/badge/Feedback-green)
- AI Chains: Transparent and Controllable Human-AI Interaction by Chaining Large Language Model Prompts - AI%20Chains-blue) ![img](https://img.shields.io/badge/-Prompt%20Chaining-orange) ![img](https://img.shields.io/badge/-Natural%20Language-lightgrey) ![img](https://img.shields.io/badge/Configuration-green)
- Interactive Text Generation - ITG-blue) ![img](https://img.shields.io/badge/-Imitation%20Learning-orange) ![img](https://img.shields.io/badge/-Edits-lightgrey) ![img](https://img.shields.io/badge/Simulation-green)
- PromptChainer: Chaining Large Language Model Prompts through Visual Programming - PromptChainer-blue) ![img](https://img.shields.io/badge/-Prompt%20Chaining-orange) ![img](https://img.shields.io/badge/-Natural%20Language-lightgrey) ![img](https://img.shields.io/badge/Configuration-green)
- RRHF: Rank Responses to Align Language Models with Human Feedback without tears - RRHF-blue) ![img](https://img.shields.io/badge/-Reinforcement%20Learning-orange) ![img](https://img.shields.io/badge/Feedback-green)
- RAFT: Reward rAnked FineTuning for Generative Foundation Model Alignment - RAFT-blue) ![img](https://img.shields.io/badge/Feedback-green)
- Improving Grounded Language Understanding in a Collaborative Environment by Interacting with Agents Through Help Feedback - Natural%20Language-lightgrey) ![img](https://img.shields.io/badge/Chat,%20Feedback-green)
- Is Reinforcement Learning (Not) for Natural Language Processing: Benchmarks, Baselines, and Building Blocks for Natural Language Policy Optimization - RL4LMs-blue) ![img](https://img.shields.io/badge/-Reinforcement%20Learning-orange) ![img](https://img.shields.io/badge/Feedback-green)
- Towards Teachable Reasoning Systems: Using a Dynamic Memory of User Feedback for Continual System Improvement - TeachMe-blue) ![img](https://img.shields.io/badge/-Continual%20Learning-orange) ![img](https://img.shields.io/badge/-Natural%20Language-lightgrey) ![img](https://img.shields.io/badge/Feedback-green)
- MemPrompt: Memory-assisted Prompt Editing with User Feedback - MemPrompt-blue) ![img](https://img.shields.io/badge/-Active%20Learning-orange) ![img](https://img.shields.io/badge/-Edits-lightgrey) ![img](https://img.shields.io/badge/Personalization,%20Feedback-green)
- Craft an Iron Sword: Dynamically Generating Interactive Game Characters by Prompting Large Language Models Tuned on Code - green)
- Languages are Rewards: Hindsight Finetuning using Human Feedback - Chain%20of%20Hindsight-blue) ![img](https://img.shields.io/badge/-Natural%20Language-lightgrey) ![img](https://img.shields.io/badge/Feedback-green)
- InternChat: Solving Vision-Centric Tasks by Interacting with Chatbots Beyond Language - InternChat-blue) ![img](https://img.shields.io/badge/-Instruction%20Tuning-orange) ![img](https://img.shields.io/badge/Chat-green)
- Improving Code Generation by Training with Natural Language Feedback - ILF-blue) ![img](https://img.shields.io/badge/-Imitation%20Learning-orange) ![img](https://img.shields.io/badge/-Natural%20Language-lightgrey) ![img](https://img.shields.io/badge/Feedback-green)
-
🤖Model/🛠Tool-LM Interaction
- MindCraft: Theory of Mind Modeling for Situated Dialogue in Collaborative Tasks - MindCraft-blue) ![img](https://img.shields.io/badge/-Prompt%20Chaining-orange) ![img](https://img.shields.io/badge/-Natural%20Language-lightgrey) ![img](https://img.shields.io/badge/Theory%20of%20Mind,%20Communication-green)
- Toolformer: Language Models Can Teach Themselves to Use Tools - Toolformer-blue) ![img](https://img.shields.io/badge/Tool--use-green)
- Small Models are valuable Plug-ins for large language models - SuperICL-blue) ![img](https://img.shields.io/badge/-Standard%20Prompting-orange)
- Tree of Thoughts: Deliberate Problem Solving with Large Language Models - ToT-blue) ![img](https://img.shields.io/badge/-Prompt%20Chaining-orange) ![img](https://img.shields.io/badge/Chain%20of%20Thought-green)
- RECURRENTGPT: Interactive Generation of (Arbitrarily) Long Text - RecurrentGPT-blue) ![img](https://img.shields.io/badge/-Natural%20Language-lightgrey)
- Internet-augmented language models through few-shot prompting for open-domain question answering - -use-green)
- Internet-augmented language models through few-shot prompting for open-domain question answering - -use-green)
- Recitation-Augmented Language Models - RECITE-blue) ![img](https://img.shields.io/badge/-Elicitive%20Prompting-orange) ![img](https://img.shields.io/badge/-Natural%20Language-lightgrey) ![img](https://img.shields.io/badge/Chain%20of%20Thought-green)
- Iteratively Prompt Pre-trained Language Models for Chain of Thought - Prompt%20Chaining-orange) ![img](https://img.shields.io/badge/-Natural%20Language-lightgrey) ![img](https://img.shields.io/badge/Chain%20of%20Thought-green)
- ViperGPT: Visual Inference via Python Execution for Reasoning - ViperGPT-blue) ![img](https://img.shields.io/badge/-Elicitive%20Prompting-orange) ![img](https://img.shields.io/badge/-Formal%20Language-lightgrey) ![img](https://img.shields.io/badge/Tool--use-green)
- OpenAGI: When LLM Meets Domain Experts - OpenAGI-blue) ![img](https://img.shields.io/badge/-Standard%20Prompting,%20Reinforcement%20Learning-orange) ![img](https://img.shields.io/badge/-Natural%20Language-lightgrey) ![img](https://img.shields.io/badge/Tool--use-green)
- ChatGPT Asks, BLIP-2 Answers: Automatic Questioning Towards Enriched Visual Descriptions - ChatCaptioner-blue) ![img](https://img.shields.io/badge/-Prompt%20Chaining-orange) ![img](https://img.shields.io/badge/-Natural%20Language-lightgrey) ![img](https://img.shields.io/badge/Tool--use-green)
- Generative Agents: Interactive Simulacra of Human Behavior - Generative%20Agents-blue) ![img](https://img.shields.io/badge/-Natural%20Language-lightgrey) ![img](https://img.shields.io/badge/Society-green)
- ReAct: Synergizing reasoning and acting in language models - ReAct-blue) ![img](https://img.shields.io/badge/-Prompt%20Chaining-orange) ![img](https://img.shields.io/badge/Chain%20of%20Thought,%20Tool--use-green)
- Visual ChatGPT: Talking, Drawing and Editing with Visual Foundation Models - Visual%20ChatGPT-blue) ![img](https://img.shields.io/badge/Chain%20of%20Thought,%20Tool--use-green)
- HuggingGPT: Solving AI Tasks with ChatGPT and its Friends in Hugging Face - HuggingGPT-blue) ![img](https://img.shields.io/badge/Tool--use-green)
- CAMEL: Communicative Agents for "Mind" Exploration of Large Scale Language Model Society - CAMEL-blue) ![img](https://img.shields.io/badge/-Prompt%20Chaining-orange) ![img](https://img.shields.io/badge/Society-green)
- Program of Thoughts Prompting: Disentangling Computation from Reasoning for Numerical Reasoning Tasks - PoT%20Prompting-blue) ![img](https://img.shields.io/badge/-Elicitive%20Prompting-orange) ![img](https://img.shields.io/badge/-Formal%20Language-lightgrey) ![img](https://img.shields.io/badge/Tool--use,%20Chain%20of%20Thought-green)
- ART: Automatic multi-step reasoning and tool-use for large language models - ART-blue) ![img](https://img.shields.io/badge/-Formal%20Language,%20Natural%20Language-lightgrey) ![img](https://img.shields.io/badge/Tool--use,%20Chain%20of%20Thought-green)
- LEAST-TO-MOST PROMPTINGENABLESCOMPLEXREASONING IN LARGE LANGUAGE MODELS - least--to--most%20prompting-blue) ![img](https://img.shields.io/badge/-Prompt%20Chaining-orange) ![img](https://img.shields.io/badge/Chain%20of%20Thought-green)
- Decomposed Prompting: A Modular Approach for Solving Complex Tasks - Decomposed%20Prompting-blue) ![img](https://img.shields.io/badge/-Prompt%20Chaining-orange) ![img](https://img.shields.io/badge/Chain%20of%20Thought-green)
- ViperGPT: Visual Inference via Python Execution for Reasoning - ViperGPT-blue) ![img](https://img.shields.io/badge/-Elicitive%20Prompting-orange) ![img](https://img.shields.io/badge/-Formal%20Language-lightgrey) ![img](https://img.shields.io/badge/Tool--use-green)
- See,Think,Confirm:Interactive Prompting Between Vision and Language Models for Knowledge-based Visual Reasoning - IPVR-blue) ![img](https://img.shields.io/badge/-Prompt%20Chaining-orange) ![img](https://img.shields.io/badge/Tool--use-green)
- Large Language Models Are Reasoning Teachers - Fine--tune--CoT-blue) ![img](https://img.shields.io/badge/-Semi--Supervised%20Fine--Tuning,%20Elicitive%20Prompting-orange) ![img](https://img.shields.io/badge/Chain%20of%20Thought-green)
- STaR:Self-Taught ReasonerBootstrapping Reasoning With Reasoning - STaR-blue) ![img](https://img.shields.io/badge/-Elicitive%20Prompting,%20Semi--Supervised%20Fine--Tuning-orange) ![img](https://img.shields.io/badge/Chain%20of%20Thought-green)
- Search-in-the-Chain: Towards Accurate, Credible and Traceable Large Language Models for Knowledge-intensive Tasks - SearChain-blue) ![img](https://img.shields.io/badge/-Prompt%20Chaining-orange) ![img](https://img.shields.io/badge/Tool--use,%20Chain%20of%20Thought-green)
- PAL: Program-aided Language Models - PAL-blue) ![img](https://img.shields.io/badge/-Elicitive%20Prompting-orange) ![img](https://img.shields.io/badge/-Formal%20Language-lightgrey) ![img](https://img.shields.io/badge/Chain%20of%20Thought,%20Tool--use-green)
- Recitation-Augmented Language Models - RECITE-blue) ![img](https://img.shields.io/badge/-Elicitive%20Prompting-orange) ![img](https://img.shields.io/badge/-Natural%20Language-lightgrey) ![img](https://img.shields.io/badge/Chain%20of%20Thought-green)
- Iteratively Prompt Pre-trained Language Models for Chain of Thought - Prompt%20Chaining-orange) ![img](https://img.shields.io/badge/-Natural%20Language-lightgrey) ![img](https://img.shields.io/badge/Chain%20of%20Thought-green)
- MEASURING AND NARROWING THE COMPOSITIONALITY GAP IN LANGUAGE MODELS - self--ask-blue) ![img](https://img.shields.io/badge/-Prompt%20Chaining-orange) ![img](https://img.shields.io/badge/Tool--use,%20Chain%20of%20Thought-green)
- Self-Refine: Iterative Refinement with Self-Feedback - self--refine-blue) ![img](https://img.shields.io/badge/-Prompt%20Chaining-orange) ![img](https://img.shields.io/badge/-Natural%20Language-lightgrey) ![img](https://img.shields.io/badge/Self--Interaction-green)
- LEVER: Learning to Verify Language-to-Code Generation with Execution - LEVER-blue) ![img](https://img.shields.io/badge/-Standard%20Prompting-orange) ![img](https://img.shields.io/badge/-Formal%20Language,%20Natural%20Language-lightgrey) ![img](https://img.shields.io/badge/Tool--use-green)
- Computational Language Acquisition with Theory of Mind - green)
- Few-shot Language Coordination by Modeling Theory of Mind - green)
- OpenAGI: When LLM Meets Domain Experts - OpenAGI-blue) ![img](https://img.shields.io/badge/-Standard%20Prompting,%20Reinforcement%20Learning-orange) ![img](https://img.shields.io/badge/-Natural%20Language-lightgrey) ![img](https://img.shields.io/badge/Tool--use-green)
- MM-REACT: Prompting ChatGPT for Multimodal Reasoning and Action - MM--ReAct-blue) ![img](https://img.shields.io/badge/-Prompt%20Chaining-orange) ![img](https://img.shields.io/badge/Tool--use-green)
- Prompt, Generate, then Cache: Cascade of Foundation Models makes Strong Few-shot Learners - CaFo-blue) ![img](https://img.shields.io/badge/-Prompt%20Chaining-orange) ![img](https://img.shields.io/badge/-Natural%20Language,%20Machine%20Language-lightgrey) ![img](https://img.shields.io/badge/Tool--use-green)
- Principle-Driven Self-Alignment of Language Models from Scratch with Minimal Human Supervision - SELF--ALIGN-blue) ![img](https://img.shields.io/badge/-Instruction%20Tuning,Semi--Supervised%20Fine--Tuning-orange) ![img](https://img.shields.io/badge/-Natural%20Language-lightgrey) ![img](https://img.shields.io/badge/Alignment-green)
- Think Before You Act: Unified Policy for Interleaving Language Reasoning with Actions - Elicitive%20Prompting-orange) ![img](https://img.shields.io/badge/-Natural%20Language,%20Machine%20Language-lightgrey) ![img](https://img.shields.io/badge/Chain%20of%20Thought-green)
- TALM: Tool Augmented Language Models - TALM-blue) ![img](https://img.shields.io/badge/-Instruction%20Tuning,Semi--Supervised%20Fine--Tuning-orange) ![img](https://img.shields.io/badge/-Formal%20Language,%20Natural%20Language-lightgrey) ![img](https://img.shields.io/badge/Tool--use-green)
- Successive Prompting for Decomposing Complex Questions - Successive%20Prompting-blue) ![img](https://img.shields.io/badge/-Prompt%20Chaining-orange) ![img](https://img.shields.io/badge/Chain%20of%20Thought-green)
- REFINER: Reasoning Feedback on Intermediate Representations - REFINER-blue) ![img](https://img.shields.io/badge/-Semi--Supervised%20Fine--Tuning,%20Elicitive%20Prompting-orange) ![img](https://img.shields.io/badge/-Natural%20Language-lightgrey) ![img](https://img.shields.io/badge/Chain%20of%20Thought,%20Reasoning-green)
- LeTI: Learning to Generate from Textual Interactions - LeTI-blue) ![img](https://img.shields.io/badge/-Prompt%20Chaining-orange) ![img](https://img.shields.io/badge/-Formal%20Language,%20Natural%20Language-lightgrey) ![img](https://img.shields.io/badge/Tool--use-green)
- InternGPT: Solving Vision-Centric Tasks by Interacting with Chatbots Beyond Language - InternGPT-blue) ![img](https://img.shields.io/badge/-Formal%20Language,%20Natural%20Language-lightgrey) ![img](https://img.shields.io/badge/Tool--use-green)
- Human-level play in the game of Diplomacy by combining language models with strategic reasoning - Cicero-blue) ![img](https://img.shields.io/badge/-Reinforcement%20Learning-orange) ![img](https://img.shields.io/badge/-Natural%20Language-lightgrey) ![img](https://img.shields.io/badge/Decision%20Making-green)
- Generating Sequences by Learning to Self-Correct - Self--Correction-blue) ![img](https://img.shields.io/badge/-Semi--Supervised%20Fine--Tuning-orange) ![img](https://img.shields.io/badge/-Natural%20Language,%20Edits-lightgrey) ![img](https://img.shields.io/badge/Chain%20of%20Thought-green)
- ChatGPT-steered Editing Instructor for Customization of Abstractive Summarization - Semi--Supervised%20Fine--Tuning,%20Prompt%20Chaining-orange) ![img](https://img.shields.io/badge/-Natural%20Language,%20Edits-lightgrey) ![img](https://img.shields.io/badge/Chain%20of%20Thought-green)
- ChatGPT Asks, BLIP-2 Answers: Automatic Questioning Towards Enriched Visual Descriptions - ChatCaptioner-blue) ![img](https://img.shields.io/badge/-Prompt%20Chaining-orange) ![img](https://img.shields.io/badge/-Natural%20Language-lightgrey) ![img](https://img.shields.io/badge/Tool--use-green)
- Baize: An Open-Source Chat Model with Parameter-Efficient Tuning on Self-Chat Data - Baize-blue) ![img](https://img.shields.io/badge/-Parameter--Efficient%20Fine--Tuning,%20Semi--Supervised%20Fine--Tuning-orange) ![img](https://img.shields.io/badge/-Natural%20Language-lightgrey) ![img](https://img.shields.io/badge/Chat-green)
- Small Models are valuable Plug-ins for large language models - SuperICL-blue) ![img](https://img.shields.io/badge/-Standard%20Prompting-orange)
- Tree of Thoughts: Deliberate Problem Solving with Large Language Models - ToT-blue) ![img](https://img.shields.io/badge/-Prompt%20Chaining-orange) ![img](https://img.shields.io/badge/Chain%20of%20Thought-green)
- See,Think,Confirm:Interactive Prompting Between Vision and Language Models for Knowledge-based Visual Reasoning - IPVR-blue) ![img](https://img.shields.io/badge/-Prompt%20Chaining-orange) ![img](https://img.shields.io/badge/Tool--use-green)
- STaR:Self-Taught ReasonerBootstrapping Reasoning With Reasoning - STaR-blue) ![img](https://img.shields.io/badge/-Elicitive%20Prompting,%20Semi--Supervised%20Fine--Tuning-orange) ![img](https://img.shields.io/badge/Chain%20of%20Thought-green)
- Search-in-the-Chain: Towards Accurate, Credible and Traceable Large Language Models for Knowledge-intensive Tasks - SearChain-blue) ![img](https://img.shields.io/badge/-Prompt%20Chaining-orange) ![img](https://img.shields.io/badge/Tool--use,%20Chain%20of%20Thought-green)
- RECURRENTGPT: Interactive Generation of (Arbitrarily) Long Text - RecurrentGPT-blue) ![img](https://img.shields.io/badge/-Natural%20Language-lightgrey)
- PAL: Program-aided Language Models - PAL-blue) ![img](https://img.shields.io/badge/-Elicitive%20Prompting-orange) ![img](https://img.shields.io/badge/-Formal%20Language-lightgrey) ![img](https://img.shields.io/badge/Chain%20of%20Thought,%20Tool--use-green)
- MEASURING AND NARROWING THE COMPOSITIONALITY GAP IN LANGUAGE MODELS - self--ask-blue) ![img](https://img.shields.io/badge/-Prompt%20Chaining-orange) ![img](https://img.shields.io/badge/Tool--use,%20Chain%20of%20Thought-green)
- Self-Refine: Iterative Refinement with Self-Feedback - self--refine-blue) ![img](https://img.shields.io/badge/-Prompt%20Chaining-orange) ![img](https://img.shields.io/badge/-Natural%20Language-lightgrey) ![img](https://img.shields.io/badge/Self--Interaction-green)
- LEVER: Learning to Verify Language-to-Code Generation with Execution - LEVER-blue) ![img](https://img.shields.io/badge/-Standard%20Prompting-orange) ![img](https://img.shields.io/badge/-Formal%20Language,%20Natural%20Language-lightgrey) ![img](https://img.shields.io/badge/Tool--use-green)
- Computational Language Acquisition with Theory of Mind - green)
- Few-shot Language Coordination by Modeling Theory of Mind - green)
- Prompt, Generate, then Cache: Cascade of Foundation Models makes Strong Few-shot Learners - CaFo-blue) ![img](https://img.shields.io/badge/-Prompt%20Chaining-orange) ![img](https://img.shields.io/badge/-Natural%20Language,%20Machine%20Language-lightgrey) ![img](https://img.shields.io/badge/Tool--use-green)
- Principle-Driven Self-Alignment of Language Models from Scratch with Minimal Human Supervision - SELF--ALIGN-blue) ![img](https://img.shields.io/badge/-Instruction%20Tuning,Semi--Supervised%20Fine--Tuning-orange) ![img](https://img.shields.io/badge/-Natural%20Language-lightgrey) ![img](https://img.shields.io/badge/Alignment-green)
- Think Before You Act: Unified Policy for Interleaving Language Reasoning with Actions - Elicitive%20Prompting-orange) ![img](https://img.shields.io/badge/-Natural%20Language,%20Machine%20Language-lightgrey) ![img](https://img.shields.io/badge/Chain%20of%20Thought-green)
- TALM: Tool Augmented Language Models - TALM-blue) ![img](https://img.shields.io/badge/-Instruction%20Tuning,Semi--Supervised%20Fine--Tuning-orange) ![img](https://img.shields.io/badge/-Formal%20Language,%20Natural%20Language-lightgrey) ![img](https://img.shields.io/badge/Tool--use-green)
- Successive Prompting for Decomposing Complex Questions - Successive%20Prompting-blue) ![img](https://img.shields.io/badge/-Prompt%20Chaining-orange) ![img](https://img.shields.io/badge/Chain%20of%20Thought-green)
- REFINER: Reasoning Feedback on Intermediate Representations - REFINER-blue) ![img](https://img.shields.io/badge/-Semi--Supervised%20Fine--Tuning,%20Elicitive%20Prompting-orange) ![img](https://img.shields.io/badge/-Natural%20Language-lightgrey) ![img](https://img.shields.io/badge/Chain%20of%20Thought,%20Reasoning-green)
- LeTI: Learning to Generate from Textual Interactions - LeTI-blue) ![img](https://img.shields.io/badge/-Prompt%20Chaining-orange) ![img](https://img.shields.io/badge/-Formal%20Language,%20Natural%20Language-lightgrey) ![img](https://img.shields.io/badge/Tool--use-green)
- InternGPT: Solving Vision-Centric Tasks by Interacting with Chatbots Beyond Language - InternGPT-blue) ![img](https://img.shields.io/badge/-Formal%20Language,%20Natural%20Language-lightgrey) ![img](https://img.shields.io/badge/Tool--use-green)
- ChatGPT-steered Editing Instructor for Customization of Abstractive Summarization - Semi--Supervised%20Fine--Tuning,%20Prompt%20Chaining-orange) ![img](https://img.shields.io/badge/-Natural%20Language,%20Edits-lightgrey) ![img](https://img.shields.io/badge/Chain%20of%20Thought-green)
- Chameleon: Plug-and-Play Compositional Reasoning with Large Language Models - Chameleon-blue) ![img](https://img.shields.io/badge/-Formal%20Language,%20Natural%20Language-lightgrey) ![img](https://img.shields.io/badge/Tool--use-green)
- Baize: An Open-Source Chat Model with Parameter-Efficient Tuning on Self-Chat Data - Baize-blue) ![img](https://img.shields.io/badge/-Parameter--Efficient%20Fine--Tuning,%20Semi--Supervised%20Fine--Tuning-orange) ![img](https://img.shields.io/badge/-Natural%20Language-lightgrey) ![img](https://img.shields.io/badge/Chat-green)
-
🌎Environment-LM Interaction
- PaLM-E: An Embodied Multimodal Language Model - PaLM--E-blue)![img](https://img.shields.io/badge/Foundation%20Model-green)
- Mind's Eye: Grounded Language Model Reasoning through Simulation - Elicitive%20Prompting-orange) ![img](https://img.shields.io/badge/-Formal%20Language-lightgrey)
- Improving Multimodal Interactive Agents with Reinforcement Learning from Human Feedback - Reinforcement%20Learning-orange)
- BLIP-2: Bootstrapping Language-Image Pre-training with Frozen Image Encoders and Large Language Models - BLIP--2-blue) ![img](https://img.shields.io/badge/-Semi--Supervised%20Fine--Tuning-orange) ![img](https://img.shields.io/badge/-Machine%20Language-lightgrey) ![img](https://img.shields.io/badge/Modality%20Grounding-green)
- Do As I Can, Not As I Say: Grounding Language in Robotic Affordances - SayCan-blue) ![img](https://img.shields.io/badge/-Reinforcement%20Learning-orange) ![img](https://img.shields.io/badge/Affordance%20Grounding-green)
- RT-1: Robotics Transformer for Real-World Control at Scale - RT--1-blue)![img](https://img.shields.io/badge/Foundation%20Model-green)
- Grounded decoding: Guiding text generation with grounded models for robot control - blue) ![img](https://img.shields.io/badge/-Natural%20Language-lightgrey) ![img](https://img.shields.io/badge/Affordance%20Grounding-green)
- Inner monologue: Embodied reasoning through planning with language models. - blue)![img](https://img.shields.io/badge/Prompt%20Chaining-orange) ![img](https://img.shields.io/badge/-Natural%20Language-lightgrey)
- Code as Policies: Language Model Programs for Embodied Control - blue)![img](https://img.shields.io/badge/-Elicitive%20Prompting-orange) ![img](https://img.shields.io/badge/-Formal%20Language-lightgrey)
- Open-vocabulary Queryable Scene Representations for Real World Planning - NLMap-blue) ![img](https://img.shields.io/badge/-Active%20Learning-orange) ![img](https://img.shields.io/badge/-Natural%20Language-lightgrey) ![img](https://img.shields.io/badge/Affordance%20Grounding-green)
- Plan4MC: Skill Reinforcement Learning and Planning for Open-World Minecraft Tasks - Plan4MC-blue) ![img](https://img.shields.io/badge/-Reinforcement%20Learning-orange) ![img](https://img.shields.io/badge/-Formal%20Language-lightgrey) ![img](https://img.shields.io/badge/Affordance%20Grounding-green)
- MineDojo: Building Open-Ended Embodied Agents with Internet-Scale Knowledge - MineDojo-blue) ![img](https://img.shields.io/badge/-Active%20Learning,%20Reinforcement%20Learning-orange) ![img](https://img.shields.io/badge/-Natural%20Language-lightgrey)
- LM-Nav: Robotic Navigation with Large Pre-Trained Models of Language, Vision, and Action - -Nav-blue)![img](https://img.shields.io/badge/-Prompt%20Chaining-orange)
- WebShop: Towards Scalable Real-World Web Interaction with Grounded Language Agents - blue)![img](https://img.shields.io/badge/-Reinforcement%20Learning,%20Prompt%20Chaining-orange) ![img](https://img.shields.io/badge/-Natural%20Language,%20Formal%20Language-lightgrey) ![img](https://img.shields.io/badge/Affordance%20Grounding,%20Tool--use-green)
- VIMA: General Robot Manipulation with Multimodal Prompts - VIMA-blue) ![img](https://img.shields.io/badge/-Message%20Fusion-orange) ![img](https://img.shields.io/badge/Foundation%20Model-green)
- PIGLeT: Language Grounding Through Neuro-Symbolic Interaction in a 3D World - PIGLeT-blue) ![img](https://img.shields.io/badge/-Standard%20Prompting-orange) ![img](https://img.shields.io/badge/-Natural%20Language-lightgrey)
- Describe, Explain, Plan and Select: Interactive Planning with Large Language Models Enables Open-World Multi-Task Agents - Prompt%20Chaining,%20Reinforcement%20Learning-orange) ![img](https://img.shields.io/badge/-Formal%20Language,%20Natural%20Language-lightgrey)
- ProgPrompt: Generating Situated Robot Task Plans using Large Language Models - ProgPrompt-blue) ![img](https://img.shields.io/badge/-Elicitive%20Prompting,%20Prompt%20Chaining-orange) ![img](https://img.shields.io/badge/-Formal%20Language-lightgrey) ![img](https://img.shields.io/badge/Affordance%20Grounding-green)
- Plan,Eliminate,and Track-Language Models are Good Teachers for Embodied Agents. - Standard%20Prompting,%20Prompt%20Chaining-orange) ![img](https://img.shields.io/badge/-Natural%20Language-lightgrey) ![img](https://img.shields.io/badge/Affordance%20Grounding-green)
- Collaborating with language models for embodied reasoning - -Actor--Reporter-blue)![img](https://img.shields.io/badge/-Reinforcement%20Learning,%20Prompt%20Chaining-orange) ![img](https://img.shields.io/badge/-Natural%20Language-lightgrey)
- Open-World Object Manipulation using Pre-trained Vision-Language Models - MOO-blue)
- Guiding Pretraining in Reinforcement Learning with Large Language Models - ELLM-blue) ![img](https://img.shields.io/badge/-Reinforcement%20Learning-orange) ![img](https://img.shields.io/badge/-Natural%20Language-lightgrey)
- Grounding Large Language Models in Interactive Environments with Online Reinforcement Learning - GLAM-blue)![img](https://img.shields.io/badge/-Reinforcement%20Learning-orange) ![img](https://img.shields.io/badge/-Natural%20Language-lightgrey) ![img](https://img.shields.io/badge/Affordance%20Grounding-green)
- Don't Generate, Discriminate: A Proposal for Grounding Language Models to Real-World Environments - Prompt%20Chaining-orange)
- Interactive Language: Talking to Robots in Real Time - blue)![img](https://img.shields.io/badge/-Imitation%20Learning-orange) ![img](https://img.shields.io/badge/-Natural%20Language-lightgrey)
- Socratic Models: Composing Zero-Shot Multimodal Reasoning with Language - blue)![img](https://img.shields.io/badge/-Prompt%20Chaining-orange) ![img](https://img.shields.io/badge/-Natural%20Language%20,%20Shared%20Memory-lightgrey) ![img](https://img.shields.io/badge/Modality%20Grounding-green)
-
👍Evaluation
- Evaluating Human-Language Model Interaction - HALIE-blue) ![img](https://img.shields.io/badge/Human--LM%20Interaction-green)
- Behavior-1k: A benchmark for embodied ai with 1,000 everyday activities and realistic simulation - Behavior--1k-blue) ![img](https://img.shields.io/badge/Embodied%20AI-green)
- An AI Dungeon Master’s Guide: Learning to Converse and Guide with Intents and Theory-of-Mind in Dungeons and Dragons - -of--Mind-green)
- Evaluating Verifiability in Generative Search Engines
- Alexa Arena: A User-Centric Interactive Platform for Embodied AI - Alexa%20Arena-blue) ![img](https://img.shields.io/badge/Embodied%20AI-green)
- ReCEval: Evaluating Reasoning Chains via Correctness and Informativeness - ReCEval-blue) ![img](https://img.shields.io/badge/-Prompt%20Chaining-orange) ![img](https://img.shields.io/badge/CoT-green)
- Multi-Party Chat: Conversational Agents in Group Settings with Humans and Models - Multi--Party%20Chat-blue) ![img](https://img.shields.io/badge/Communicative%20Agents-green)
- ORBIT: A Unified Simulation Framework for Interactive Robot Learning Environments - ORBIT-blue) ![img](https://img.shields.io/badge/Embodied%20AI-green)
- Learn to Explain: Multimodal Reasoning via Thought Chains for Science Question Answering - ScienceQA-blue) ![img](https://img.shields.io/badge/CoT-green)
- Alexa Arena: A User-Centric Interactive Platform for Embodied AI - Alexa%20Arena-blue) ![img](https://img.shields.io/badge/Embodied%20AI-green)
- CB2: Collaborative Natural Language Interaction Research Platform - CB2-blue) ![img](https://img.shields.io/badge/-Natural%20Language-lightgrey) ![img](https://img.shields.io/badge/Collaborative%20Behavior-green)
- Evaluating Human-Language Model Interaction - HALIE-blue) ![img](https://img.shields.io/badge/Human--LM%20Interaction-green)
- ReCEval: Evaluating Reasoning Chains via Correctness and Informativeness - ReCEval-blue) ![img](https://img.shields.io/badge/-Prompt%20Chaining-orange) ![img](https://img.shields.io/badge/CoT-green)
- Evaluating Verifiability in Generative Search Engines
- Multi-Party Chat: Conversational Agents in Group Settings with Humans and Models - Multi--Party%20Chat-blue) ![img](https://img.shields.io/badge/Communicative%20Agents-green)
- ORBIT: A Unified Simulation Framework for Interactive Robot Learning Environments - ORBIT-blue) ![img](https://img.shields.io/badge/Embodied%20AI-green)
- Learn to Explain: Multimodal Reasoning via Thought Chains for Science Question Answering - ScienceQA-blue) ![img](https://img.shields.io/badge/CoT-green)
- CB2: Collaborative Natural Language Interaction Research Platform - CB2-blue) ![img](https://img.shields.io/badge/-Natural%20Language-lightgrey) ![img](https://img.shields.io/badge/Collaborative%20Behavior-green)
- An AI Dungeon Master’s Guide: Learning to Converse and Guide with Intents and Theory-of-Mind in Dungeons and Dragons - -of--Mind-green)
-
📚KB-LM Interaction
- ERNIE: Enhanced Representation through Knowledge Integration - ERNIE-blue) ![img](https://img.shields.io/badge/-Corpus%20Knowledge-green)
- Recitation-Augmented Language Models - RECITE-blue) ![img](https://img.shields.io/badge/-Elicitive%20Prompting-orange)![img](https://img.shields.io/badge/-Corpus%20Knowledge-green)
- Don't Generate, Discriminate: A Proposal for Grounding Language Models to Real-World Environments - Pangu-blue) ![img](https://img.shields.io/badge/-Message%20Fusion-orange) ![img](https://img.shields.io/badge/-Internet%20Knowledge-green)
- Check Your Facts and Try Again: Improving Large Language Models with External Knowledge and Automated Feedback - LLM--Augmenter-blue) ![img](https://img.shields.io/badge/-Prompt%20Chaining-orange) ![img](https://img.shields.io/badge/-Internet%20Knowledge-green)
- WebGPT: Browser-assisted question-answering with human feedback - WebGPT-blue) ![img](https://img.shields.io/badge/-Reinforcement%20Learning-orange) ![img](https://img.shields.io/badge/-Internet%20Knowledge-green)
- K-BERT: Enabling Language Representation with Knowledge Graph - K--BERT-blue) ![img](https://img.shields.io/badge/-Message%20Fusion-orange) ![img](https://img.shields.io/badge/-Corpus%20Knowledge-green)
- Atlas: Few-shot learning with retrieval augmented language models - Atlas-blue) ![img](https://img.shields.io/badge/-Corpus%20Knowledge-green)
- MineDojo: Building Open-Ended Embodied Agents with Internet-Scale Knowledge - MineDojo-blue) ![img](https://img.shields.io/badge/-Reinforcement%20Learning-orange) ![img](https://img.shields.io/badge/-Internet%20Knowledge-green)
- WebGPT: Browser-assisted question-answering with human feedback - WebGPT-blue) ![img](https://img.shields.io/badge/-Reinforcement%20Learning-orange) ![img](https://img.shields.io/badge/-Internet%20Knowledge-green)
- Improving language models by retrieving from trillions of tokens - Retro-blue) ![img](https://img.shields.io/badge/-Message%20Fusion-orange) ![img](https://img.shields.io/badge/-Corpus%20Knowledge-green)
- KELM: Knowledge Enhanced Pre-Trained Language Representations with Message Passing on Hierarchical Relational Graphs - KELM-blue) ![img](https://img.shields.io/badge/-Message%20Fusion-orange) ![img](https://img.shields.io/badge/-Corpus%20Knowledge-green)
- K-Adapter: Infusing Knowledge into Pre-Trained Models with Adapters - K--Adapter-blue) ![img](https://img.shields.io/badge/-Parameter%20Efficient%20Fine--Tuning-orange) ![img](https://img.shields.io/badge/-Formal%20Language-lightgrey) ![img](https://img.shields.io/badge/-Corpus%20Knowledge-green)
- Binding Language Models in Symbolic Languages - Binder-blue) ![img](https://img.shields.io/badge/-Elicitive%20Prompting-orange) ![img](https://img.shields.io/badge/-Formal%20Language-lightgrey) ![img](https://img.shields.io/badge/-Corpus%20Knowledge-green)
- Teaching language models to support answers with verified quotes - GopherCite-blue) ![img](https://img.shields.io/badge/-Reinforcement%20Learning-orange) ![img](https://img.shields.io/badge/-Internet%20Knowledge-green)
- Retrieval-Augmented Generation for Knowledge-Intensive NLP Tasks - RAG-blue) ![img](https://img.shields.io/badge/-Corpus%20Knowledge-green)
- DocPrompting: Generating Code by Retrieving the Docs - DocPrompting-blue) ![img](https://img.shields.io/badge/-Standard%20Prompting-orange) ![img](https://img.shields.io/badge/-Formal%20Language-lightgrey) ![img](https://img.shields.io/badge/-Internet%20Knowledge-green)
- Mind's Eye: Grounded Language Model Reasoning through Simulation - Mind%E2%80%99s%20Eye-blue) ![img](https://img.shields.io/badge/-Elicitive%20Prompting-orange) ![img](https://img.shields.io/badge/-Formal%20Language-lightgrey)
- Don't Generate, Discriminate: A Proposal for Grounding Language Models to Real-World Environments - Pangu-blue) ![img](https://img.shields.io/badge/-Message%20Fusion-orange) ![img](https://img.shields.io/badge/-Internet%20Knowledge-green)
- Check Your Facts and Try Again: Improving Large Language Models with External Knowledge and Automated Feedback - LLM--Augmenter-blue) ![img](https://img.shields.io/badge/-Prompt%20Chaining-orange) ![img](https://img.shields.io/badge/-Internet%20Knowledge-green)
- REALM: retrieval-augmented language model pre-training - REALM-blue) ![img](https://img.shields.io/badge/-Message%20Fusion-orange) ![img](https://img.shields.io/badge/-Corpus%20Knowledge-green)
- Binding Language Models in Symbolic Languages - Binder-blue) ![img](https://img.shields.io/badge/-Elicitive%20Prompting-orange) ![img](https://img.shields.io/badge/-Formal%20Language-lightgrey) ![img](https://img.shields.io/badge/-Corpus%20Knowledge-green)
- Teaching language models to support answers with verified quotes - GopherCite-blue) ![img](https://img.shields.io/badge/-Reinforcement%20Learning-orange) ![img](https://img.shields.io/badge/-Internet%20Knowledge-green)
-
🔭Surveys and Position Papers
- Interactive Natural Language Processing - green)
- Tool Learning with Foundation Models - Tool--use-green)
- Augmented Language Models: a Survey - Reasoning-green) ![img](https://img.shields.io/badge/-Tool--use-green)
- Foundation Models for Decision Making: Problems, Methods, and Opportunities - Tool--use-green) ![img](https://img.shields.io/badge/-Decision%20Making-green)
-
-
Keywords Convention
-
Contribution
-
Contributors
- ![img - Team/awesome-InteractiveNLP-papers/graphs/contributors)
-
-
Star History
-
Contributing to this paper list
- ![Star History Chart - history.com/#InteractiveNLP-Team/awesome-InteractiveNLP-papers&Date)
- ![Star History Chart - history.com/#InteractiveNLP-Team/awesome-InteractiveNLP-papers&Date)
-
Programming Languages
Sub Categories
Keywords
visual-instruction-tuning
1
visual-in-context-learning
1
visual-chain-of-thought
1
multimodal-large-language-models
1
multimodal-instruction-tuning
1
multimodal-in-context-learning
1
multimodal-chain-of-thought
1
multi-modality
1
large-vision-language-models
1
large-vision-language-model
1
large-language-models
1
instruction-tuning
1
instruction-following
1
in-context-learning
1
chain-of-thought
1
llm
1
gpt-4
1
gpt
1
ai
1
agent
1