Oct. 14, 2022, 1:16 a.m. | Yanwei Li, Yilun Chen, Xiaojuan Qi, Zeming Li, Jian Sun, Jiaya Jia

cs.CV updates on arXiv.org arxiv.org

In this work, we present a unified framework for multi-modality 3D object
detection, named UVTR. The proposed method aims to unify multi-modality
representations in the voxel space for accurate and robust single- or
cross-modality 3D detection. To this end, the modality-specific space is first
designed to represent different inputs in the voxel feature space. Different
from previous work, our approach preserves the voxel space without height
compression to alleviate semantic ambiguity and enable spatial connections. To
make full use of …

arxiv detection representation transformer voxel

AI Research Scientist

@ Vara | Berlin, Germany and Remote

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne

Lead Data Scientist, Commercial Analytics

@ Checkout.com | London, United Kingdom

Data Engineer I

@ Love's Travel Stops | Oklahoma City, OK, US, 73120