ICRA 2023 BARN Challenge


Wondering how the state-of-the-art navigation systems work on autonomous navigation in highly-constrained spaces? Check out our detailed report on the results and findings of The BARN Challenge at ICRA 2023 here!


Congrats KUL+FM, INVENTEC, and University of Almeria for winning the 1st, 2nd, and 3rd place of The BARN Challenge at ICRA 2023! See you next year!

Physical Example

About

Designing autonomous robot navigation systems has been a topic of interest to the robotics community for decades. Indeed, many existing navigation systems allow robots to move from one point to another in a collision-free manner, which may create the impression that navigation is a solved problem. However, autonomous mobile robots still struggle in many scenarios, e.g., colliding with or getting stuck in novel and tightly constrained spaces. These troublesome scenarios have many real-world implications, such as poor navigation performance in adversarial search and rescue environments, in naturally cluttered daily households, and in congested social spaces such as classrooms, offices, and cafeterias. Meeting these real-world challenges requires systems that can both successfully and efficiently navigate the environment with confidence, posing fundamental challenges to current autonomous systems, from perception to control. Therefore, the Benchmark Autonomous Robot Navigation (BARN) Challenge aims at creating a benchmark for state-of-the-art navigation systems and pushing the boundaries of their performance in these challenging and highly constrained environments.

At ICRA 2022, we successfully launched the first BARN Challenge in Philadelphia. Five teams competed in the simulation competition and the top three teams advanced into the physical finals. We congratulate UT AMRL, UVA AMR, and Temple TRAIL for winning the 1st, 2nd, and 3rd place of The BARN Challenge at ICRA 2022. We published the results and findings about how the state-of-the-art navigation systems work in highly-constrained spaces in IEEE Robotics and Automation Magazine.



The Challenge

The BARN Challenge will take place primarily on the simulated BARN dataset and also physically at the conference venue in London.

The BARN dataset comprises 300 pre-generated navigation environments, ranging from easy open spaces to difficult highly constrained ones, and an environment generator that can generate novel BARN environments. The task is to navigate a Clearpath Jackal robot from a predefined start to a goal location as quickly as possible without any collision. The Jackal robot will be standardized with a 2D LiDAR, a motor controller with a max speed of 2m/s, and appropriate computational resources. Participants will need to develop navigation systems which consume the standardized LiDAR input, run all computation onboard using the provided resources, and output motion commands to drive the motors. Participants are welcome to use any approaches to tackle the navigation problem, such as using classical sampling-based or optimization-based planners, end-to-end learning, or hybrid approaches. The following infrastructure will be provided by the competition organizers:



Competition Rules

During the competition, another 50 new BARN evaluation environments will be generated, which will not be accessible to the public. Each participating team is required to submit the developed navigation system as a (collection of) launchable ROS node(s). The final performance will be evaluated based on a standardized metric that considers navigation success rate (collision or not reaching the goal count as failure), actual traversal time, and environment difficulty (measured by optimal traversal time). Specially, the score $s$ for navigating each environment $i$ will be computed as

$$s_i = 1^{\textrm{success}} \times \frac{\textrm{OT}}{\textrm{clip}(\textrm{AT}_i, 4\textrm{OT}_i, 8\textrm{OT}_i)}$$

where the indicator $1^{\textrm{success}}$ is set to $1$ if the robot reaches the navigation goal without any collisions, and set to $0$ otherwise. AT denotes the actual traversal time, while OT denotes the optimal traversal time, as an indicator of the environment difficulty and measured by the shortest traversal time assuming the robot always travels at its maximum speed ($2\textrm{m/s}$):

$$\textrm{OT}_i = \frac{\textrm{Path Length}_i}{\textrm{Maximum Speed}}$$

The Path Length is provided by the BARN dataset based on Dijkstra's search. The clip function clips the AT within 4OT and 8OT, in order to assure navigating extremely quickly or slowly in easy or difficult environments won't disproportionally scale the score. The overall score of each team is the score averaged over all 50 test BARN environments, with 10 trials in each environment. Higher scores indicate better navigation performance.

During the in-person conference in London, physical obstacle courses will be set up at the conference venue using cardboard boxes (see example below). A physical Clearpath Jackal robot will be provided by the competition sponsor, Clearpath Robotics, with the same standard sensor and actuator suites as in the simulation. However, a sim-to-real gap may still exist due to factors such as friction coefficient between physical wheels and venue floor, electric motor impedance and resistance, real-world sensor noise distribution, etc. The top three teams will be invited to compete in the physical competition. In case a team is not able to attend the conference in person, the organizers will run their submitted navigation stack on their behalf on the physical robot at the conference. The team who achieves the highest collision-free navigation success rate and shortest traversal time wins the competition.


Physical Example

Due to the overhead of the simulator in the simulation phase, we do not specifically constrain computation. We will use a computer with Intel Xeon Gold 6342 CPU @ 2.80 GHz to evaluate all submissions. However, during the physical competition, robot onboard computation will be limited to an Intel i3 CPU (i3-9100TE or similar) with 16GB of DDR4 RAM. GPUs are not available considering the small footprint of the LiDAR perception data.



Leaderboard

Real World

2023

Ranking Team Success / Total Trials Code Link
1 KUL+FM 9/9 GitHub
2 INVENTEC 6/9 GitHub
3 University of Almeria 5/9 GitHub

2022

Ranking Team Success / Total Trials Code Link
1 UT AMRL 8/9 GitHub
2 UVA AMR 4/9 GitHub
3 Temple TRAIL 2/9 GitHub
Real-world challenge in 2022 includes three different obstacle courses. Each team has 30 minutes to finish five timed trials. Top three trials are counted. The team that finishes the most successful trials (reaching the goal without any collision) wins. In the case of a tie, the team with the fasted average traversal time wins.

Simulation

Ranking Team Score Comment
1 KUL+FM 0.2490 KU Leuven & Flanders Make
2 INVENTEC 0.2445 Inventec Corporation
3 University of Almeria 0.2439 University of Almeria & Teladoc
4 UT AMRL 0.2424 The University of Texas at Austin
5 Temple TRAIL22 0.2415 Temple University
6 LfLHb 0.2334 Details
7 UT AMRL22 0.2310 The University of Texas at Austin
8 Temple TRAIL 0.2290 Temple University
9 UVA AMR 0.2237 University of Virginia
10 RIL 0.2203 Indian Institute of Science, Bengaluru
11 UVA AMR22 0.2200 University of Virginia
12 E-Bandb 0.2053 Details
13 e2eb 0.2042 Details
14 Staxel 0.2019 Staxel
15 APPLR-DWAb 0.1979 Details
16 Yiyuiii22 0.1969 Nanjing University
17 The MECO Barners 0.1829 KU Leuven
18 NavBot22 0.1733 Indian Institute of Science, Bangalore
19 Fast DWAb 0.1709 Details
20 Default DWAb 0.1627 Details
21 Lumnonicity NA AICoE, Jio
b denotes baseline.
22 denotes 2022 participants.
Last updated: May 22 2023. If you do not see your submission, it is still under evaluation.


Participation

We have standardized the entire navigation pipeline using the BARN dataset in a standardized Singularity container. You should only modify the navigation system, and leave other parts intact. You can develop your navigation system in your local environment or in a container. You will need to upload your code (e.g., to GitHub) and submit a Singularityfile.def file which will access your code. We will build your Singularity container using the Singularityfile.def file you submitted, run the built container for evaluation, and then publish your score on our website. More detailed instructions can be found here.
Please use this Submission Form to submit your navigation stack to be evaluated by our standardized evaluation pipeline.

Citing BARN

If you find BARN useful for your research, please cite the following paper:

@inproceedings{perille2020benchmarking,
title = {Benchmarking Metric Ground Navigation},
author = {Perille, Daniel and Truong, Abigail and Xiao, Xuesu and Stone, Peter},
booktitle = {2020 IEEE International Symposium on Safety, Security and Rescue Robotics (SSRR)},
year = {2020},
organization = {IEEE}
}


Schedule

Date Event
January 1 2023 Online submission open
April 20 2023 Soft Submission Deadline1
May 1 2023 1st Round Physical Competition Invitations
May 20 2023 Hard Submission Deadline2
May 23 2023 2nd Round Physical Competition Invitations
May 29 2023 noon-6pm Team Set-up based on Availability
May 30 2023 9p-2p Team Set-up
May 30 2023 2p-5p Physical Competition Obstacle Course 1
May 31 2023 9a-noon Physical Competition Obstacle Course 2
May 31 2023 1p-4p Physical Competition Obstacle Course 3
May 31 2023 4p-5p Award Ceremony and Open Discussions
1Deadline to guarantee an invitation to the final physical competition given good navigation performance in simulation.
2Deadline to participate in the simulation competition. However, final physical competition eligibility will depend on the available capacity and travel arrangement made beforehand.


Organizers

Xuesu Xiao
Xuesu Xiao
George Mason University
Everyday Robots
Zifan Xu
Zifan Xu
UT Austin
Garrett Warnell
Garrett Warnell
US Army Research Lab
UT Austin
Peter Stone
Peter Stone
UT Austin
Sony AI
Clearpath Robotics
Clearpath Robotics
Competition Sponsor


Contact

For questions, please contact:

Dr. Xuesu Xiao
Assistant Professor
Department of Computer Science
George Mason University
4400 University Drive, Fairfax, VA 22030, USA
xiao@gmu.edu