3260 papers • 126 benchmarks • 313 datasets
This task has no description! Would you like to contribute one?
(Image credit: Papersgraph)
These leaderboards are used to track progress in transfer-reinforcement-learning-15
No benchmarks available.
Use these libraries to find transfer-reinforcement-learning-15 models and implementations
No subtasks available.
This work defines a novel method of multitask and transfer learning that enables an autonomous agent to learn how to behave in multiple tasks simultaneously, and then generalize its knowledge to new domains, and uses Atari games as a testing environment to demonstrate these methods.
This work explores a new challenge in transfer RL, where only a set of source policies collected under diverse unknown dynamics is available for learning a target task efficiently, and proposes MULTI-source POLicy AggRegation (MULTIPOLAR), which learns to aggregate the actions provided by the source policies adaptively to maximize the target task performance.
Shapechanger, a library for transfer reinforcement learning specifically designed for robotic tasks, considers three types of knowledge transfer---from simulation to simulation, from simulation to real, and from real to real---and a wide range of tasks with continuous states and actions.
This work finds that the representations learned are not only effective for goal-directed visual imitation via gradient-based trajectory optimization, but can also provide a metric for specifying goals using images.
This work proposes a reinforcement learning framework based on a self-critic policy gradient approach which achieves good generalization and state-of-the-art results on a variety of datasets and provides a generic solution that works well on unseen data.
An upgraded, real world application oriented version of gym-gazebo, the Robot Operating System (ROS) and Gazebo based Reinforcement Learning (RL) toolkit, which complies with OpenAI Gym is presented.
This work uses the kinematic structure directly as the hardware encoding and shows great zero-shot transfer to completely novel robots not seen during training and demonstrates that fine-tuning the policy network is significantly more sample-efficient than training a model from scratch.
The results indicate that this approach can be used to solve robotic manipulation problems that would otherwise be infeasible without expert demonstrations.
This work proposes a two-stage RL agent that first learns a latent unified state representation (LUSR) which is consistent across multiple domains in the first stage, and then does RL training in one source domain based on LUSR in the second stage to achieve state-of-the-art domain adaptation performance.
Adding a benchmark result helps the community track progress.