all AI news
Social-MAE: Social Masked Autoencoder for Multi-person Motion Representation Learning
April 9, 2024, 4:47 a.m. | Mahsa Ehsanpour, Ian Reid, Hamid Rezatofighi
cs.CV updates on arXiv.org arxiv.org
Abstract: For a complete comprehension of multi-person scenes, it is essential to go beyond basic tasks like detection and tracking. Higher-level tasks, such as understanding the interactions and social activities among individuals, are also crucial. Progress towards models that can fully understand scenes involving multiple people is hindered by a lack of sufficient annotated data for such high-level tasks. To address this challenge, we introduce Social-MAE, a simple yet effective transformer-based masked autoencoder framework for multi-person …
abstract arxiv autoencoder basic beyond cs.cv detection interactions masked autoencoder multiple people person progress representation representation learning social tasks tracking type understanding
More from arxiv.org / cs.CV updates on arXiv.org
Jobs in AI, ML, Big Data
Data Architect
@ University of Texas at Austin | Austin, TX
Data ETL Engineer
@ University of Texas at Austin | Austin, TX
Lead GNSS Data Scientist
@ Lurra Systems | Melbourne
Senior Machine Learning Engineer (MLOps)
@ Promaton | Remote, Europe
Data Scientist
@ Publicis Groupe | New York City, United States
Bigdata Cloud Developer - Spark - Assistant Manager
@ State Street | Hyderabad, India