https://gym.openai.com/envs/CarRacing-v0/
3260 papers • 126 benchmarks • 313 datasets
https://gym.openai.com/envs/CarRacing-v0/
(Image credit: Papersgraph)
These leaderboards are used to track progress in continuous-control-2
No benchmarks available.
Use these libraries to find continuous-control-2 models and implementations
No datasets available.
No subtasks available.
This manual describes the competition software for the Simulated Car Racing Championship and provides an overview of the architecture, the instructions to install the software and to run the simple drivers provided in the package.
This paper demonstrates the surprising finding that models with the same precise parts can be instead efficiently trained end-to-end through a genetic algorithm (GA), reaching a comparable performance to the original world model by solving a challenging car racing task.
This work presents a safety-critical model predictive control strategy utilizing discrete-time control barrier functions (CBFs), which guarantees system safety and accomplishes optimal performance via model predictivecontrol.
An extension of the DAgger, called SafeDAgger, is proposed that is query-efficient and more suitable for end-to-end autonomous driving and observes a significant speed up in convergence, which is conjecture to be due to the effect of automated curriculum learning.
This work addresses the issue of computationally inefficiency with a model that does not need to generate a high-dimensional image at each time-step and can be used to improve exploration and is adaptable to many diverse environments.
The proposed framework for autonomous driving using deep reinforcement learning incorporates Recurrent Neural Networks for information integration, enabling the car to handle partially observable scenarios and integrates the recent work on attention models to focus on relevant information, thereby reducing the computational complexity for deployment on embedded hardware.
This work approaches an alternative Interactive Machine Learning strategy for training DNN policies based on human corrective feedback, with a method called Deep COACH (D-COACH), which takes advantage of the knowledge and insights of human teachers as well as the power of DNNs, but also has no need of a reward function.
This work proposes a search method for neural network architectures that can already perform a task without any explicit weight training, and demonstrates that this method can find minimal neural network architecture that can perform several reinforcement learning tasks without weight training.
An algorithm that safely and interactively learns a model of the user's reward function and actively synthesizes hypothetical behaviors from scratch by maximizing tractable proxies for the value of information, without interacting with the environment.
Adding a benchmark result helps the community track progress.