Category: Post categories
Robert Moni’s experience after winning AI-DO 5
An interview with Robert Moni
Robert is a Ph. D. student at the Budapest University of Technology and Economics.
His work focuses on deep learning and he has (co)authored papers on reinforcement learning (RL), imitation learning (IL), and sim-to-real learning using for autonomous vehicles using Duckietown.
Robert and his team won the LFV_multi
hardware challenge of the 2020 AI Driving Olympics.
Today, Robert shares some of his thoughts with us!
What brought you to work on AVs?
I started my journey in the world of AV’s in 2016 when I was hired at the automotive supplier company “Continental” in Romania. In 2018 I moved to Budapest, Hungary, to join Continental’s Deep Learning Competence Center where we develop novel perception methods for AVs.
In 2019, with the support of the company, I started my Ph.D. at Budapest University of Technology and Economics on the topic “Deep Reinforcement Learning in Complex environments”.
At this time, I crossed paths with the Duckietown environment. Continental bought 12 Duckiebots and supplementary materials to build our own Duckietown environment in a lab at the university.
Tell us about you and your team
At the beginning of my Ph. D. and with the arrival of the Duckietown materials we established the “PIA” (Professional Intelligence for Automotive) project with the aim to provide education and mentorship for undergrad and master students in the field on Machine Learning and AV.
In each semester since 2019 February I managed a team of 4-6 people developing their own solutions for AI-DO challenges. I wrote a short blogpost presenting my team and our solutions submitted to AI-DO 5.
"With the arrival of the Duckietown material we established the PIA project with the aim to provide education and mentorship for undergrad and master students in the field on Machine Learning and autonomous vehicles (AV)."
Robert Moni
What approach did you choose for AI-DO, and why?
I started to tackle the AI-DO challenges applying deep reinforcement learning (DRL) for driver policy learning and state representation learning (SRL) for sim2real transfer.
The reason for my chosen approach is my Ph. D. topic, and I plan to develop and test my hypotheses in the Duckietown environment.
What are the hardest challenges that you faced in the competition?
In the beginning, there was a simple agent training task that caused some headaches: finding a working DRL method, composing a good reward function, preprocessing the observations to reduce the search space, and fine-tuning all the parameters. All these were challenges, but well-known ones in the field.
One unexpected challenge was the continuous updates of the gym-duckietown environment. While we are thrilled that the environment gets improved by the Duckietown team, we faced occasional breakdowns in our methods when applying them to the newest releases, which caused some frustration.
The biggest headache was caused by the different setups in the training and evaluation environments: in the evaluation environment, the images are dimmed while during training they are clear. Furthermore, the real world is full of nuisances – for example lags introduced by WiFi communication, which causes different outcomes in the real environment. This challenge can be mitigated to some degree with the algorithms running directly on the Duckiebot’s hardware, and by using a more powerful onboard computer, e.g., the Jetson Nano 2GB development board.
Are you satisfied with the final outcome?
I am satisfied with the achievements of my team, which kept the resolve throughout the technical challenges faced.
I’m sure we would’ve done even better in the real-world challenge if we had seen our submission running earlier in the Autolab, so we could have adjusted our algorithms. We are going to work to bring one to our University in the next future.
What are you going to change next time?
I believe the AI-DO competition as well as the Duckietown platform would improve through more powerful hardware. I hope to see Duckiebots (DB19) upgraded to support the new Jetson Nano hardware!
(Since the date of the interview, Duckiebots model DB21 supports Jetson Nano boards)
Learn more about Duckietown
The Duckietown platform offers robotics and AI learning experiences.
Duckietown is modular, customizable and state-of-the-art. It is designed to teach, learn, and do research: from exploring the fundamentals of computer science and automation to pushing the boundaries of knowledge.
Tell us your story
Are you an instructor, learner, researcher or professional with a Duckietown story to tell? Reach out to us!
AI Driving Olympics 5th edition: results
AI-DO 5: Urban league winners
This year’s challenges were lane following (LF
), lane following with pedestrians (LFP
) and lane following with other vehicles, multibody (LFV_multi
).
Let’s find out the results in each category:
LF
- Andras Beres 🇭🇺
- Zoltan Lorincz 🇭🇺
- András Kalapos 🇭🇺
LFP
- Bea Baselines 🐤
- Melisande Teng 🇨🇦
- Raphael Jean 🇨🇦
LFV_multi
- Robert Moni 🇭🇺
- Márton Tim 🇭🇺
- Anastasiya Nikolskay 🇷🇺
Congratulations to the Hungarian Team from the Budapest University of Technology and Economics for collecting the highest rankings in the urban league!
Here’s how the winners in each category performed both in the qualification (simulation) and in the finals running on real hardware:
Andras Beres - Lane following (LF) winner
Melisande Teng - Lane following with pedestrians (LFP) winner
Robert Moni - Lane following with other vehicles, multibody (LFV_multi) winner
AI-DO 5: Advanced Perception league winners
Great participation and results in the Advanced Perception league! Check out this year’s winners in the video below:
AI-DO 5 sponsors
Many thanks to our amazing sponsors, without which none of this would have been possible!
Stay tuned for next year AI Driving Olympics. Visit the AI-DO page for more information on the competition and to browse this year’s introductory webinars, or check out the Duckietown massive open online course (MOOC) and prepare for next year’s competition!
AI-DO 5 competition leaderboard update
AI-DO 5 pre-finals update
With the fifth edition of the AI Driving Olympics finals day approaching, 1326 solutions submitted from 94 competitors in three challenges, it is time to glance over at the leaderboards!
Leaderboards updates
LF
), lane following with pedestrians (LFP
) and lane following with other vehicles, multibody (LFV_multi
). Learn more about the challenges here.
Each submission can be sent to multiple challenges.
Let’s look at some of the most promising or interesting submissions. The Montréal menace
Raphael Jean at Mila / University of Montréal is a new entrant for this year.
An interesting submission: submission #12962
The submissions from the cold
Team JetBrains from Saint Petersburg was a winner of previous editions of AI-DO. They have been dominating the leaderboards also this year.
Interesting submissions: submission #12905
All of JetBrains submissions: JBRRussia1.
BME Conti
PhD student Robert Moni (BME-Conti) from Hungary.
Interesting submissions: submission #12999
All submissions: timur-BMEconti
Deadline for submissions
AI-DO 5 competition Update
AI-DO 5 Update
AI-DO 5 is in full swing and we want to bring you some updates: better graphics, more maps, faster and more reliable backend and an improved GUI to submit to challenges!
Challenges visualization
We updated the visualization. Now the evaluation produces videos with your name and evaluation number (as below).
Challenges updates
We fixed some of the bugs in the simulator regarding the visualization (“phantom robots” popping in and out).
We updated the maps in the challenges to have more variety in the road network; we put more grass and trees to make the maps more joyful!
Faster and more reliable backend
The server was getting slow given the number of submissions, and sometime the service was unavailable. We have revamped the server code and added some backend capacity to be more fault-tolerant. It is now much faster!
Thanks so much to the participants that helped us debug this problem!
More evaluators
We brought online many more CPU and GPU evaluators. We now encourage you to submit more often as we have a lot more capacity.
Submit to testing challenges
We also remind you that the challenges on the front page are the validation challenges, in which everybody can see the output. However what counts for winning are the testing challenges!
To do that you can use dts challenges submit with the –challenges option
Or, you can use a new way using the website that we just implemented, described below.
Submitting to other challenges
Step 1: Go to your user page, by clicking “login” and then going to “My Submissions”.
Step 2: In this page you will find your submissions grouped by “component”.
Click the component icon as in the figure.
Step 3: The page will contain some buttons that allow you to submit to other challenges that you didn’t submit to yet.
Imitation Learning Approach for AI Driving Olympics Trained on Real-world and Simulation Data Simultaneously
- Imitation Learning Approach for AI Driving Olympics Trained on Real-world and Simulation Data Simultaneously
- Mikita Sazanovich, Konstantin Chaika, Kirill Krinkin, Aleksei Shpilman
- Workshop on AI for Autonomous Driving (AIAD), the 37th International Conference on Machine Learning, Vienna, Austria, 2020
- ArXiv version download: arXiv:2007.03514
- Find code here
Imitation Learning Approach for AI Driving Olympics Trained on Real-world and Simulation Data Simultaneously
The AIDO challenge is divided into two global stages: simulation and real-world. A single algorithm needs to perform well in both. It was quickly identified that one of the major problems is the simulation to real-world transfer.
Many algorithms trained in the simulated environment performed very poorly in the real world, and many classic control algorithms that are known to perform well in a real-world environment, once tuned to that environment, do not perform well in the simulation. Some approaches suggest randomizing the domain for the simulation to real-world transfer.
We propose a novel method of training a neural network model that can perform well in diverse environments, such as simulations and real-world environment.
Dataset Generation
To that end, we have trained our model through imitation learning on a dataset compiled from four different sources:
- Real-world Duckietown dataset from logs.duckietown.com (REAL-DT).
- Simulation dataset on a simple loop map (SIM-LP).
- Simulation dataset on an intersection map (SIM-IS).
- Real-world dataset collected by us in our environment with car driven by PD controller (REAL-IH).
We aimed to collect data with as many possible situations such as twists in the road, driving in circles clockwise/counterclockwise, and so on. We have also tried to diversify external factors such as scene lighting, items in the room that can get into the camera’s field of view, roadside objects, etc. If we keep these conditions constant, our model may overfit to them and perform poorly in a different environment. For this reason, we changed the lighting and environment after each duckiebot run. The lane detection was calibrated for every lighting condition since different lighting changes the color scheme of the image input.
We made the following change to the standard PD algorithm: since most Duckietown turns and intersections are standard-shaped, we hard-coded the robot’s motion in these situations, but we did not exclude imperfect trajectories. For example, the ones that would go slightly out of bounds of the lane. Imperfections in the robot’s actions increase the robustness of the model.
Neural network architecture and training
Original images are 640×480 RGB. As a preprocessing step, we remove the top third of the image, since it mostly contains the sky, resize the image to 64×32 pixels and convert it into the YUV colorspace.
We have used 5 convolutional layers with a small number of filters, followed by 2 fully-connected layers. The small size of the network is not only due to it being less prone to overfitting, but we also need a model that can run on a single CPU on RaspberryPi.
We have also incorporated Independent-Component (IC) layers. These layers aim to make the activations of each layer more independent by combining two popular techniques, BatchNorm and Dropout. For convolutional layers, we substitute Dropout with Spatial Dropout which has been shown to work better with them. The model outputs two values for voltages of the left and the right wheel drives. We use the mean square error (MSE) as our training loss.
Results
For the training evaluation, we compute the mean square error (MSE) of the left and the right wheels outputs on the validation set of each data source.
The first table shows the results for the models trained on all data sources (HYBRID), on real-world data sources only (REAL) and on simulation data sources only (SIM). As we can see, while training on a single dataset sometimes achieves lower error on the same dataset than our hybrid approach. We can also see that our method performs on par with the best single methods. In terms of the average error it outperforms the closest one tenfold. This demonstrates definitively the high dependence of MSE on the training method, and highlights the differences between the data sources.
The next table shows simulation closed-loop performance for all our approaches using the Duckietown simulator. All methods drove for 15 seconds without major infractions, and the SIM model that was trained specifically on the simulation data only drove just 1.8 tiles more than our hybrid approach.
The third table shows the closed-loop performance in the real-world environment. Comparing the number of tiles, we see that our hybrid approach drove about 3.5 tiles more than the following in the rankings model trained on real-world data only.
Conclusion
Our method follows the imitation learning approach and consists of a convolutional neural network which is trained on a dataset compiled from data from different sources, such as simulation model and real-world Duckietown vehicle driven by a PD controller, tuned to various conditions, such as different map configuration and lighting.
We believe that our approach of emphasizing neurons independence and monitoring generalization performance can offer more robustness to control models that have to perform in diverse environments. We also believe that the described approach of imitation learning on data obtained from several algorithms that are fitted to specific environments may yield a single algorithm that will perform well in general.
The “Self-Driving cars with Duckietown” Massive Open Online Course on edX
"Self-Driving Cars with Duckietown" hands-on MOOC on edX
We are launching a massive open online course (MOOC): “Self-Driving Cars with Duckietown” on edX, and it is free to attend!
This course is made possible thanks to the support of the Swiss Federal Institute of Technology in Zurich (ETHZ), in collaboration with the University of Montreal, the Duckietown Foundation, and the Toyota Technological Institute at Chicago.
This course combines remote and hands-on learning with real-world robots. It is offered on edX, the trusted platform for learning, and it is now open for enrollment.
Learning autonomy
Participants will engage in software and hardware hands-on learning experiences, with focus on overcoming the challenges of deploying autonomous robots in the real world.
This course will explore the theory and implementation of model- and data-driven approaches for making a model self-driving car drive autonomously in an urban environment.
Pedestrian detection
MOOC Factsheet
- Name: Self-driving cars with Duckietown
- Start: March 2021
- Platform: edX
- Cost: free to attend
Prerequisites
- Basic Linux, Python, Git
- Elements of linear algebra, probability, calculus
- Elements of kinematics, dynamics
- Computer with native Ubuntu installation
What you will learn
- Computer Vision
- Robot operations
- Object Detection
- Localization, planning and control
- Reinforcement Learning
Why Self-driving cars with Duckietown?
Teaching autonomy requires a fundamentally different approach when compared to other computer science and engineering disciplines, because it is multi-disciplinary. Mastering it requires expertise in domains ranging from fundamental mathematics to practical machine-learning skills.
Robot Perception
Robots operate in the real world, and theory and practice often do not play well together. There are many hardware platforms and software tools, each with its own strengths and weaknesses. It is not always clear what tools are worth investing time in mastering, and how these skills will generalize to different platforms.
Duckiebot Detection
Learning through challenges
Progressing through behaviors of increasing complexity, participants uncover concepts and tools that address the limitations of previous approaches. This allows to get Duckiebots to actually do things, while gradually re-iterating concepts through various technical frameworks. Simulation and real-world experiments will be performed using a Python, ROS, and Docker based software stack.
Robot Planning
(Hidden) This line and everything under this line are hidden
This course combines remote and hands-on learning with real-world robots.
It is offered on edX, the trusted platform for learning, and it is now open for enrollment.
Learning activities will support the use of NVIDIA Jetson Nano powered Duckiebots.
Join the AI Driving Olympics, 5th edition, starting now!
Compete in the 5th AI Driving Olympics (AI-DO)
The 5th edition of the Artificial Intelligence Driving Olympics (AI-DO 5) has officially started!
The AI-DO serves to benchmark the state of the art of artificial intelligence in autonomous driving by providing standardized simulation and hardware environments for tasks related to multi-sensory perception and embodied AI.
Duckietown hosts AI-DO competitions biannually, with finals events held at machine learning and robotics conferences such as the International Conference on Robotics and Automation (ICRA) and the Neural Information Processing Systems (NeurIPS).
The AI-DO 5 will be in conjunction with NeurIPS 2020 and have two leagues: Urban Driving and Advanced Perception.
Urban driving league challenges
This year’s Urban League includes a traditional AI-DO challenge (LF) and introduces two new ones (LFP, LFVM).
The most traditional of AI-DO challenges: have a Duckiebot navigate a road loop without intersection, pedestrians (duckies) or other vehicles. The objective is traveling the longest path in a given time while staying in the lane.
The LFP challenge is new to AI-DO. It builds upon LF by introducing static obstacles (duckies) on the road. The objectives are the same as for lane following, but do not hit the duckies!
In this traditional AI-DO challenge, contestants seek to travel the longest path in a city without intersections nor pedestrians, but with other vehicles on the road. Except this year there’s a twist. In this year’s novel multi-body variant, all vehicles on the road are controlled by the submission.
Getting started: the webinars
We offer a short webinar series to guide contestants through the steps for participating: from running our baselines in simulation as well as deploying them on hardware. All webinars are 9 am EST and free!
Introduction
Learn about the Duckietown project and the Artificial Intelligence Driving Olympics.
- Nov. 9, 2020
ROS baseline
How to run and build upon the “traditional” Robotic Operation System (ROS) baseline.
- Nov. 11, 2020
Local development
On the workflow for developing and deploying to Duckiebots, for hardware-based testing.
- Nov. 13, 2020
RL baseline
Learn how to use the Pytorch template for reinforcement learning approaches.
- Nov. 16, 2020
IL baseline
Introduction to the Tensorflow template, use of logs and simulator for imitation learning.
- Nov. 18, 2020
Advanced sensing league challenges
Previous AI-DO editions featured: detection, tracking and prediction challenges around the nuScenes dataset.
For the 5th iteration of AI-DO we have a brand new lidar segmentation challenge.
The challenge is based on the recently released lidar segmentation annotations for nuScenes and features an astonishing 1,400,000,000 lidar points annotated with one of 32 labels.
We hope that this new benchmark will help to push the boundaries in lidar segmentation. Please see https://www.nuscenes.org/lidar-segmentation for more details.
Furthermore, due to popular demand, we will organize the 3rd iteration of the nuScenes 3d detection challenge. Please see https://www.nuscenes.org/object-detection for more details.
AI-DO 5 Finals event
The AI-DO finals will be streamed LIVE during 2020 edition of the Neural Information Processing Systems (NeurIPS 2020) conference in December.
Learn more about the AI-DO here.
Thank you to our generous sponsors!
The Duckietown Foundation is grateful to its sponsors for supporting this fifth edition of the AI Driving Olympics!
Integrated Benchmarking and Design for Reproducible and Accessible Evaluation of Robotic Agents
- Integrated Benchmarking and Design for Reproducible and Accessible Evaluation of Robotic Agents
- Jacopo Tani, Andrea F. Daniele, Gianmarco Bernasconi, Amaury Camus, Aleksandar Petrov, Anthony Courchesne, Bhairav Mehta, Rohit Suri, Tomasz Zaluska, Matthew R. Walter, Emilio Frazzoli, Liam Paull, Andrea Censi
- 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) October 25-29, 2020, Las Vegas, NV, USA (Virtual)
- ArXiv version download: arXiv:2009.04362v1
- Find the code here
Integrated Benchmarking and Design for Reproducible and Accessible Evaluation of Robotic Agents
Why is this important?
As robotics matures and increases in complexity, it is more necessary than ever that robot autonomy research be reproducible.
Compared to other sciences, there are specific challenges to benchmarking autonomy, such as the complexity of the software stacks, the variability of the hardware and the reliance on data-driven techniques, amongst others.
We describe a new concept for reproducible robotics research that integrates development and benchmarking, so that reproducibility is obtained by design from the beginning of the research/development processes.
We first provide the overall conceptual objectives to achieve this goal and then a concrete instance that we have built: the DUCKIENet.
The Duckietown Automated Laboratories (Autolabs)
One of the central components of this setup is the Duckietown Autolab (DTA), a remotely accessible standardized setup that is itself also relatively low-cost and reproducible.
DTAs include an off-the-shelf camera-based localization system. The accessibility of the hardware testing environment through enables experimental benchmarking that can be performed on a network of DTAs in different geographical locations.
The DUCKIENet
When evaluating agents, careful definition of interfaces allows users to choose among local versus remote evaluation using simulation, logs, or remote automated hardware setups. The Decentralized Urban Collaborative Benchmarking Environment Network (DUCKIENet) is an instantiation of this design based on the Duckietown platform that provides an accessible and reproducible framework focused on autonomous vehicle fleets operating in model urban environments.
The DUCKIENet enables users to develop and test a wide variety of different algorithms using available resources (simulator, logs, cloud evaluations, etc.), and then deploy their algorithms locally in simulation, locally on a robot, in a cloud-based simulation, or on a real robot in a remote lab. In each case, the submitter receives feedback and scores based on well-defined metrics.
Validation
We validate the system by analyzing the repeatability of experiments conducted using the infrastructure and show that there is low variance across different robot hardware and across different remote labs. We built DTAs at the Swiss Federal Institute of Technology in Zurich (ETHZ) and at the Toyota Technological Institute at Chicago (TTIC).
Conclusions
Our contention is that there is a need for stronger efforts towards reproducible research for robotics, and that to achieve this we need to consider the evaluation in equal terms as the algorithms themselves. In this fashion, we can obtain reproducibility by design through the research and development processes. Achieving this on a large-scale will contribute to a more systemic evaluation of robotics research and, in turn, increase the progress of development.
If you found this interesting, you might want to:
IROS2020: Watch The Workshop on Benchmarking Progress in Autonomous Driving
What a start for IROS 2020 with the "Benchmarking Progress in Autonomous Driving" workshop!
The 2020 edition of the International Conference on Intelligent Robots and Systems (IROS) started great with the workshop on “Benchmarking Progress in Autonomous Driving”.
The workshop was held virtually on October 25th, 2020, using an engaging and concise format of a sequence of four 1.5-hour moderated round-table discussions (including an introduction) centered around 4 themes.
The discussions on the methods by which progress in autonomous driving is evaluated, benchmarked, and verified were exciting. Many thanks to all the panelists and the organizers!
Here are the videos of the various sessions.
Opening remarks
Theme 1: Assessing progress for the field of autonomous vehicles (AVs)
Moderator: Andrea Censi
Invited Panelists:
- Emilio Frazzoli (ETH Zürich / Motional)
- Alex Kendall (Wayve)
- Jane Lappin (Chair, Transportation Research Board Standing Committee on Vehicle-Highway Automation)
- Atul Acharya (Director, AV Strategy at AAA Northern California)
Theme 2: How to evaluate AV risk from the perspective of real world deployment (public acceptance, insurance, liability, …)?
Moderator: Jacopo Tani
Invited Panelists:
- Bryant Walker Smith (USC Faculty of Law)
- Luigi Di Lillo (Swiss Re Insurance)
- John Leonard (MIT)
Theme 3: Best practices for AV benchmarking
Moderator: Liam Paull
Invited Panelists:
- Fabio Bonsignorio (Heron Robots)
- Michael Milford (QUT)
- Oscar Beijbom (Motional)
- Marcelo H. Ang Jr (National University of Singapore)
Theme 4: Do we need new paradigms for AV development?
Moderator: Matt Walter
Invited Panelists:
- Raquel Urtasun (U of Toronto / Uber ATG)
- Edwin Olson (May Mobility)
- Ram Vasudevan (University of Michigan)
- Sertac Karaman (MIT / Optimus Ride)
Closing remarks
You can find additional information about the workshop here.