ShelfMCL - Semantic particle filter localization with low-cost sensors
2023-2024 - Paper in progress
Team members:
Shivendra Agrawal (lead)
Ashutosh Naik (Graduate researcher)
Dusty Woods (Lab Manager)
Jake Brawer (Postdoc)
Bradley Hayes (Advisor)
Overview
- Problem: Estimating where an autonomous system is in a dynamic and cluttered environment such as a grocery store is challenging and unsolved. Such localization ability is essential to providing navigation guidance.
![Our system's capabilities across tasks](/assets/img/paper/shelfmcl/husky.gif)
- Solution: We present a novel Semantic Monte Carlo Localization algorithm that requires just an RGB-D camera and a visual odometry camera. Our system can be 1. mounted on a cart or 2. worn as a wearable.
Challenges
Environmental Complexity:
- The uniform and symmetric nature of aisles makes depth or LiDAR observations essentially featureless and unreliable.
![Our system's capabilities across tasks](/assets/img/paper/shelfmcl/rays.gif)
- Constantly changing product stock and poses exacerbates the localization challenge.
![First Image](/assets/img/paper/shelfmcl/dusty_stock.gif)
![Second Image](/assets/img/paper/shelfmcl/lidarline.gif)
Our Solution: A Novel Semantic Particle Filter
Minimal Sensor Requirements
Our system uses low-cost sensors such as an RGB-D camera and a visual odometry camera, without requiring LiDAR or wheel odometry.
![Sensors Used in the System](/assets/img/paper/shelfmcl/sensors.png)
Modularity
- Mountable on Carts/Strollers: Can add autonomous capabilities to existing equipment.
- Wearable: Can support assistive technology for navigation.
![Mountable Rig on a Cart](/assets/img/paper/shelfmcl/maria_cart.gif)
![Wearable Rig for Assistive Technology](/assets/img/paper/shelfmcl/dusty_wearable.gif)
Algorithm
-
Semantic Mapping
We trained a custom classifier to classify products into a fixed number of classes.
![Semantic Mapping - Classification](/assets/img/paper/shelfmcl/classification_rgb.gif)
![Semantic Mapping - Classification](/assets/img/paper/shelfmcl/classification_d.gif)
-
Pose Correction
Real-world pose estimates obtained through inverse camera projection are refined using ray casting on the semantic map.
![Product Clusters After Pose Correction](/assets/img/paper/shelfmcl/raycast.gif)
-
Semantic Localization
Semantic information is fused with the depth observation to in a Monte Carlo Localization framework.
Demo
![Ground Truth Trajectory Plot](/assets/img/paper/shelfmcl/optitrack_trajectory_plot.png)
![Localization Trajectory Plot](/assets/img/paper/shelfmcl/particle_filter_trajectory_plot.png)
(Left) Ground truth pose for the above & (Right) Localization pose estimates from our system.
Preminary Results
(Detailed results are coming soon)
- Without semantic information, the system converges incorrectly.
- Using semantic information, our system performs robust global localization and maintains it.