Learn-to-Race is an OpenAI gym compliant, multimodal control environment where agents learn how to race. Unlike many simplistic learning environments, ours is built around Arrival’s high-fidelity racing simulator featuring full software-in-the-loop (SIL), and even hardware-in-the-loop (HIL), simulation capabilities. This simulator has played a key role in bringing autonomous racing technology to real life in the Roborace series, the world’s first extreme competition of teams developing self-driving AI.
Please visit our official docs for a comprehensive guide on getting started with the environment. Happy racing!
While learning-based agents continue to demonstrate superhuman performance in many areas, we believe that they still lack in terms of generalization abilities and often require too many interactions. In summary, agents will have the ability to learn on training racetracks, but will be evaluated on their performance on an unseen evaluation track. However, the evaluation track is not truly unseen. Much like a Formula-1 driver, we will let agents interact with the new track for 60 minutes during a pre-evaluation stage before true evaluation.
We provide multiple baseline agents to demonstrate how to use Learn-to-Race including both classical and learning-based controllers. The first is a RandomActionAgent
to show basic functionality. We also include a Soft Actor-Critic agent, tabula rasa, trained for 1000 epsiodes. On the Las Vegas track, it is able to consistently complete laps in under 2 minutes each using only visual features from the virtual camera as input.
One of the key features of this environment is the ability to create arbitrary configurations of vehicle sensors. This provides users a rich sandbox for multimodal, learning based approaches. The following sensors are supported and can be placed, if applicable, at any location relative to the vehicle:
- RGB cameras
- Depth cameras
- Ground truth segmentation cameras
- Fisheye cameras
- Ray trace LiDARs
- Depth 2D LiDARs
- Radars
Additionally, these sensors are parameterized and can be customized further; for example, cameras have modifiable image size, field-of-view, and exposure. We provide a sample configuration below which has front, birdseye, and side facing cameras both in RGB mode and with ground truth segmentation.
Left Facing | Front Facing | Right Facing | Birdseye |
---|---|---|---|
Please visit our documentation for more information about sensor customization.
Python: We use Learn-to-Race with Python 3.8+.
Graphics Hardware: An Nvidia graphics card & associated drives is required. An Nvidia 970 GTX graphics card is minimally sufficient to simply run the simulator, but a better card is recommended.
Docker: Commonly, the racing simulator runs in a Docker container.
Container GPU Access: If running the simulator in a container, the container needs access to the GPU, so nvidia-container-runtime is also required.
Due to the container GPU access requirement, this installation assumes a Linux operating system. If you do not have a Linux OS, we recommend running Learn-to-Race on a public cloud instance that has a sufficient GPU.
- Request access to the Racing simulator: https://www.aicrowd.com/challenges/learn-to-race-autonomous-racing-virtual-challenge
We recommmend running the simulator as a Python subprocess which simply requires that you specify the path of the simulator in the env_kwargs.controller_kwargs.sim_path
of your configuration file. Alternatively, you can run the simulator as a Docker container by setting env_kwargs.controller_kwargs.start_container
to True. If you prefer the latter, you can load the docker image as follows:
$ docker load < arrival-sim-image.tar.gz
- Download the source code from this repository and install the package requirements. We recommend using a virtual environment:
$ conda create -n l2r python=3.6
$ conda activate # activate the environment
(l2r) $ pip3 install git+https://github.com/learn-to-race/l2r.git@aicrowd-environment
Please cite this work if you use L2R as a part of your research.
Main:
@inproceedings{herman2021learn,
title={Learn-to-race: A multimodal control environment for autonomous racing},
author={Herman, James and Francis, Jonathan and Ganju, Siddha and Chen, Bingqing and Koul, Anirudh and Gupta, Abhinav and Skabelkin, Alexey and Zhukov, Ivan and Kumskoy, Max and Nyberg, Eric},
booktitle={Proceedings of the IEEE/CVF International Conference on Computer Vision},
pages={9793--9802},
year={2021}
}
L2R Task 2 Benchmark:
@article{francis2022learn,
title={Learn-to-race challenge 2022: Benchmarking safe learning and cross-domain generalisation in autonomous racing},
author={Francis, Jonathan and Chen, Bingqing and Ganju, Siddha and Kathpal, Sidharth and Poonganam, Jyotish and Shivani, Ayush and Genc, Sahika and Zhukov, Ivan and Kumskoy, Max and Koul, Anirudh and others},
journal={arXiv preprint arXiv:2205.02953},
year={2022}
}