1
PI-QT-Opt: Predictive Information Improves Multi-Task Robotic Reinforcement Learning at Scale
2
Inner Monologue: Embodied Reasoning through Planning with Language Models
3
PaLM: Scaling Language Modeling with Pathways
4
R3M: A Universal Visual Representation for Robot Manipulation
5
Training language models to follow instructions with human feedback
6
A data-driven approach for learning to control computers
7
BC-Z: Zero-Shot Task Generalization with Robotic Imitation Learning
8
Pre-Trained Language Models for Interactive Decision-Making
9
Chain of Thought Prompting Elicits Reasoning in Large Language Models
10
Can Wikipedia Help Offline Reinforcement Learning?
11
LaMDA: Language Models for Dialog Applications
12
Language Models as Zero-Shot Planners: Extracting Actionable Knowledge for Embodied Agents
13
Scaling Language Models: Methods, Analysis & Insights from Training Gopher
14
Value Function Spaces: Skill-Centric State Abstractions for Long-Horizon Reasoning
15
Skill Induction and Planning with Latent Language
16
CLIPort: What and Where Pathways for Robotic Manipulation
17
Example-Driven Model-Based Reinforcement Learning for Solving Long-Horizon Visuomotor Tasks
18
Finetuned Language Models Are Zero-Shot Learners
19
Learning Language-Conditioned Robot Behavior from Offline Data and Crowd-Sourced Annotation
20
On the Opportunities and Risks of Foundation Models
21
Embodied BERT: A Transformer Model for Embodied, Language-guided Visual Task Completion
22
BEHAVIOR: Benchmark for Everyday Household Activities in Virtual, Interactive, and Ecological Environments
23
Language Grounding with 3D Objects
24
A Persistent Spatial Semantic Representation for High-level Natural Language Instruction Execution
25
MERLOT: Multimodal Neural Script Knowledge Models
26
PIGLeT: Language Grounding Through Neuro-Symbolic Interaction in a 3D World
27
ReLMoGen: Integrating Motion Generation in Reinforcement Learning for Mobile Manipulation
28
Episodic Transformer for Vision-and-Language Navigation
29
Understanding by Understanding Not: Modeling Negation in Language Models
30
Open-vocabulary Object Detection via Vision and Language Knowledge Distillation
31
MT-Opt: Continuous Multi-Task Robotic Reinforcement Learning at Scale
32
Actionable Models: Unsupervised Offline Reinforcement Learning of Robotic Skills
33
On the Dangers of Stochastic Parrots: Can Language Models Be Too Big? 🦜
34
Learning Transferable Visual Models From Natural Language Supervision
35
RetinaGAN: An Object-aware Approach to Sim-to-Real Transfer
36
Language-Conditioned Imitation Learning for Robot Manipulation Tasks
37
Broadly-Exploring, Local-Policy Trees for Long-Horizon Task Planning
38
PixL2R: Guiding Reinforcement Learning Using Natural Language by Mapping Pixels to Rewards
39
Climbing towards NLU: On Meaning, Form, and Understanding in the Age of Data
40
Language Models are Few-Shot Learners
41
Human Instruction-Following with Deep Reinforcement Learning via Transfer-Learning from Text
42
Grounding Language in Play
43
Robots That Use Language
44
Thinking While Moving: Deep Reinforcement Learning with Concurrent Control
45
Jointly Improving Parsing and Perception for Natural Language Commands through Human-Robot Dialog
46
ALFRED: A Benchmark for Interpreting Grounded Instructions for Everyday Tasks
47
Online Replanning in Belief Space for Partially Observable Task and Motion Problems
48
HRL4IN: Hierarchical Reinforcement Learning for Interactive Navigation with Mobile Manipulators
49
Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer
50
Self-Educated Language Agent with Hindsight Experience Replay for Instruction Following
51
Hierarchical Foresight: Self-Supervised Learning of Long-Horizon Tasks via Visual Subgoal Generation
52
Regression Planning Networks
53
VisualBERT: A Simple and Performant Baseline for Vision and Language
54
ViLBERT: Pretraining Task-Agnostic Visiolinguistic Representations for Vision-and-Language Tasks
55
Language as an Abstraction for Hierarchical Deep Reinforcement Learning
56
Search on the Replay Buffer: Bridging Planning and Reinforcement Learning
57
A Survey of Reinforcement Learning Informed by Natural Language
58
VideoBERT: A Joint Model for Video and Language Representation Learning
60
Neural Task Graphs: Generalizing to Unseen Tasks From a Single Video Demonstration
61
Differentiable Physics and Stable Modes for Tool-Use and Manipulation Planning
62
Universal Sentence Encoder
63
Semi-parametric Topological Memory for Navigation
64
Guiding Exploratory Behaviors for Multi-Modal Grounding of Linguistic Descriptions
65
Neural Task Programming: Learning to Generalize Across Hierarchical Tasks
66
Grounded Language Learning in a Simulated 3D World
67
Zero-Shot Task Generalization with Multi-Task Deep Reinforcement Learning
68
Attention is All you Need
69
Visual Semantic Planning Using Deep Successor Representations
70
Mapping Instructions and Visual Observations to Actions with Reinforcement Learning
71
Modular Multitask Reinforcement Learning with Policy Sketches
72
Learning Language Games through Interaction
73
Prioritized Experience Replay
74
Logic-Geometric Programming: An Optimization-Based Approach to Combined Task and Motion Planning
75
Listen, Attend, and Walk: Neural Mapping of Navigational Instructions to Action Sequences
76
Combined task and motion planning through an extensible planner-independent interface layer
77
Tell me Dave: Context-sensitive grounding of natural language to manipulation instructions
78
Asking for Help Using Inverse Semantics
79
A Joint Model of Language and Perception for Grounded Attribute Learning
80
RoboFrameNet: Verb-centric semantics for actions in robot middleware
81
Understanding Natural Language Commands for Robotic Navigation and Mobile Manipulation
82
Hierarchical Planning in the Now
83
Toward understanding natural language directions
84
The theory of affordances
85
Walk the Talk: Connecting Language, Knowledge, and Action in Route Instructions
87
SHOP: Simple Hierarchical Ordered Planner
88
Grounding language in perception
89
The Ecological Approach to Visual Perception
90
STRIPS: A New Approach to the Application of Theorem Proving to Problem Solving
91
Inventing Relational State and Action Abstractions for Effective and Efficient Bilevel Planning
92
Grounding Language to Autonomously-Acquired Skills via Goal Generation
93
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
94
TF-Agents: A library for reinforcement learning in tensorflow
95
Polosukhin. Attention is all you need. Advances in neural information processing systems
96
A Structure for Plans and Behavior
97
Understanding natural language
98
Robot: 1. find a water bottle, 2. pick up the water bottle, 3. go to the counter, 4. put down the water bottle, 5. done
99
Move the grapefruit drink from the table to the close counter
100
How would you put the grapes in the bowl and then move the cheese to the table?
101
Explanation: The user has asked for snacks, I will bring jalapeno chips and an apple
102
Robot: 1. put down the banana, 2. done
103
Explanation: The user has asked me to move the grapefruit drink to the counter