Skip to content

Commit ce57289

Browse files
committed
added ICRA26 papers
1 parent 3f16a08 commit ce57289

1 file changed

Lines changed: 22 additions & 8 deletions

File tree

_bibliography/references.bib

Lines changed: 22 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -1,3 +1,25 @@
1+
@inproceedings{kim2026raven,
2+
title = {RAVEN: Resilient Aerial Navigation via Open-Set Semantic Memory and Behavior Adaptation},
3+
author = {Kim, Seungchan and Alama, Omar and Kurdydyk, Dmytro and Keetha, Nikhil and Wang, Wenshan and Bisk, Yonatan and Scherer, Sebastian},
4+
year = {2026},
5+
booktitle = {2026 IEEE International Conference on Robotics and Automation (ICRA)},
6+
url = {https://arxiv.org/abs/2509.23563},
7+
video = {https://youtu.be/slLuZv3-zIs?si=zoYQKNCBQVVUJ6Ez},
8+
abstract = {Aerial outdoor semantic navigation requires robots to explore large, unstructured environments to locate target objects. Recent advances in semantic navigation have demonstrated open-set object-goal navigation in indoor settings, but these methods remain limited by constrained spatial ranges and structured layouts, making them unsuitable for long-range outdoor search. While outdoor semantic navigation approaches exist, they either rely on reactive policies based on current observations, which tend to produce short-sighted behaviors, or precompute scene graphs offline for navigation, limiting adaptability to online deployment. We present RAVEN, a 3D memory-based, behavior tree framework for aerial semantic navigation in unstructured outdoor environments. It (1) uses a spatially consistent semantic voxel-ray map as persistent memory, enabling long-horizon planning and avoiding purely reactive behaviors, (2) combines short-range voxel search and long-range ray search to scale to large environments, (3) leverages a large vision-language model to suggest auxiliary cues, mitigating sparsity of outdoor targets. These components are coordinated by a behavior tree, which adaptively switches behaviors for robust operation. We evaluate RAVEN in 10 photorealistic outdoor simulation environments over 100 semantic tasks, encompassing single-object search, multi-class, multi-instance navigation and sequential task changes. Results show RAVEN outperforms baselines by 85.25% in simulation and demonstrate its real-world applicability through deployment on an aerial robot in outdoor field tests.}
9+
}
10+
@inproceedings{gupta2026umi,
11+
title = {UMI-on-Air: Embodiment-Aware Guidance for Embodiment-Agnostic Visuomotor Policies},
12+
author = {Gupta, Harsh and Guo, Xiaofeng and Ha, Huy and Pan, Chuer and Cao, Muqing and Lee, Dongjae and Scherer, Sebastian and Song, Shuran and Shi, Guanya},
13+
year = {2026},
14+
booktitle = {2026 IEEE International Conference on Robotics and Automation (ICRA)},
15+
url = {https://arxiv.org/abs/2510.02614}
16+
}
17+
@inproceedings{maheshwari2026anythermal,
18+
title = {AnyThermal: Towards Learning Universal Representations for Thermal Perception},
19+
author = {Maheshwari, Parv and Karhade, Jay and Chawla, Yogesh and Adu, Isaiah and Heisen, Florian and Porco, Andrew and Jong, Andrew and Liu, Yifei and Pitla, Santosh and Scherer, Sebastian and Wang, Wenshan},
20+
year = {2026},
21+
booktitle = {2026 IEEE International Conference on Robotics and Automation (ICRA)}
22+
}
123
@inproceedings{keetha2026mapanything,
224
title = {{MapAnything}: Universal Feed-Forward Metric 3D Reconstruction},
325
author = {Keetha, Nikhil and M{\"u}ller, Norman and Sch{\"o}nberger, Johannes and Porzi, Lorenzo and Zhang, Yuchen and Fischer, Tobias and Knapitsch, Arno and Zauss, Duncan and Weber, Ethan and Antunes, Nelson and others},
@@ -47,14 +69,6 @@ @article{chen2025cometokens
4769
journal = {arXiv preprint arXiv:2511.14751},
4870
abstract = {We propose Confidence-Guided Token Merging (Co-Me), an acceleration mechanism for visual geometric transformers without retraining or finetuning the base model. Co-Me distilled a light-weight confidence predictor to rank tokens by uncertainty and selectively merge low-confidence ones, effectively reducing computation while maintaining spatial coverage. Compared to similarity-based merging or pruning, the confidence signal in Co-Me reliably indicates regions emphasized by the transformer, enabling substantial acceleration without degrading performance. Co-Me applies seamlessly to various multi-view and streaming visual geometric transformers, achieving speedups that scale with sequence length. When applied to VGGT and MapAnything, Co-Me achieves up to 11.3x and 7.2x speedup, making visual geometric transformers practical for real-time 3D perception and reconstruction.}
4971
}
50-
@article{kim2025raven,
51-
title = {RAVEN: Resilient Aerial Navigation via Open-Set Semantic Memory and Behavior Adaptation},
52-
author = {Kim, Seungchan and Alama, Omar and Kurdydyk, Dmytro and Keetha, Nikhil and Wang, Wenshan and Bisk, Yonatan and Scherer, Sebastian},
53-
year = {2025},
54-
url = {https://arxiv.org/abs/2509.23563},
55-
journal = {arXiv preprint arXiv:2509.23563},
56-
abstract = {Aerial outdoor semantic navigation requires robots to explore large, unstructured environments to locate target objects. Recent advances in semantic navigation have demonstrated open-set object-goal navigation in indoor settings, but these methods remain limited by constrained spatial ranges and structured layouts, making them unsuitable for long-range outdoor search. While outdoor semantic navigation approaches exist, they either rely on reactive policies based on current observations, which tend to produce short-sighted behaviors, or precompute scene graphs offline for navigation, limiting adaptability to online deployment. We present RAVEN, a 3D memory-based, behavior tree framework for aerial semantic navigation in unstructured outdoor environments. It (1) uses a spatially consistent semantic voxel-ray map as persistent memory, enabling long-horizon planning and avoiding purely reactive behaviors, (2) combines short-range voxel search and long-range ray search to scale to large environments, (3) leverages a large vision-language model to suggest auxiliary cues, mitigating sparsity of outdoor targets. These components are coordinated by a behavior tree, which adaptively switches behaviors for robust operation. We evaluate RAVEN in 10 photorealistic outdoor simulation environments over 100 semantic tasks, encompassing single-object search, multi-class, multi-instance navigation and sequential task changes. Results show RAVEN outperforms baselines by 85.25% in simulation and demonstrate its real-world applicability through deployment on an aerial robot in outdoor field tests.}
57-
}
5872
@misc{yu2025unified,
5973
title = {Unified Spherical Frontend: Learning Rotation-Equivariant Representations of Spherical Images from Any Camera},
6074
shorttitle = {Unified Spherical Frontend},

0 commit comments

Comments
 (0)