April 19, 2024, 4:44 a.m. | Xunsong Li, Pengzhan Sun, Yangcen Liu, Lixin Duan, Wen Li

cs.CV updates on arXiv.org arxiv.org

arXiv:2404.11903v1 Announce Type: new
Abstract: The interactions between human and objects are important for recognizing object-centric actions. Existing methods usually adopt a two-stage pipeline, where object proposals are first detected using a pretrained detector, and then are fed to an action recognition model for extracting video features and learning the object relations for action recognition. However, since the action prior is unknown in the object detection stage, important objects could be easily overlooked, leading to inferior action recognition performance. In …

abstract action recognition arxiv cs.cv detection features fed human interactions object objects pipeline proposals reasoning recognition stage type video

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne

Senior Machine Learning Engineer (MLOps)

@ Promaton | Remote, Europe

Senior Data Engineer

@ Quantexa | Sydney, New South Wales, Australia

Staff Analytics Engineer

@ Warner Bros. Discovery | NY New York 230 Park Avenue South