Skip to content

Commit 25b089d

Browse files
committed
Update reference.bib
1 parent 097d170 commit 25b089d

1 file changed

Lines changed: 14 additions & 0 deletions

File tree

_bibliography/references.bib

Lines changed: 14 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -1,3 +1,17 @@
1+
@misc{yu2025unified,
2+
title = {Unified Spherical Frontend: Learning Rotation-Equivariant Representations of Spherical Images from Any Camera},
3+
shorttitle = {Unified Spherical Frontend},
4+
author = {Mukai Yu and Mosam Dabhi and Liuyue Xie and Sebastian Scherer and L{\'a}szl{\'o} A. Jeni},
5+
year = {2025},
6+
publisher = {arXiv},
7+
doi = {10.48550/arXiv.2511.18174},
8+
url = {https://arxiv.org/pdf/2511.18174},
9+
eprint = {2511.18174},
10+
primaryclass = {cs.CV},
11+
abstract = {Modern perception increasingly relies on fisheye, panoramic, and other wide field-of-view (FoV) cameras, yet most pipelines still apply planar CNNs designed for pinhole imagery on 2D grids, where image-space neighborhoods misrepresent physical adjacency and models are sensitive to global rotations. Frequency-domain spherical CNNs partially address this mismatch but require costly spherical harmonic transforms that constrain resolution and efficiency. We introduce the Unified Spherical Frontend (USF), a lens-agnostic framework that transforms images from any calibrated camera into a unit-sphere representation via ray-direction correspondences, and performs spherical resampling, convolution, and pooling directly in the spatial domain. USF is modular: projection, location sampling, interpolation, and resolution control are fully decoupled. Its distance-only spherical kernels offer configurable rotation-equivariance (mirroring translation-equivariance in planar CNNs) while avoiding harmonic transforms entirely. We compare standard planar backbones with their spherical counterparts across classification, detection, and segmentation tasks on synthetic (Spherical MNIST) and real-world datasets (PANDORA, Stanford 2D-3D-S), and stress-test robustness to extreme lens distortions, varying FoV, and arbitrary rotations. USF processes high-resolution spherical imagery efficiently and maintains less than 1\% performance drop under random test-time rotations, even without rotational augmentation, and even enables zero-shot generalization from one lens type to unseen wide-FoV lenses with minimal performance degradation.},
12+
archiveprefix = {arXiv},
13+
keywords = {Computer Science - Computer Vision and Pattern Recognition}
14+
}
115
@inproceedings{alama2025rayfronts,
216
title = {RayFronts: Open-Set Semantic Ray Frontiers for Online Scene Understanding and Exploration},
317
author = {Omar Alama and Avigyan Bhattacharya and Haoyang He and Seungchan Kim and Yuheng Qiu and Wenshan Wang and Cherie Ho and Nikhil Keetha and Sebastian Scherer},

0 commit comments

Comments
 (0)