all AI news
MESEN: Exploit Multimodal Data to Design Unimodal Human Activity Recognition with Few Labels
April 3, 2024, 4:41 a.m. | Lilin Xu, Chaojie Gu, Rui Tan, Shibo He, Jiming Chen
cs.LG updates on arXiv.org arxiv.org
Abstract: Human activity recognition (HAR) will be an essential function of various emerging applications. However, HAR typically encounters challenges related to modality limitations and label scarcity, leading to an application gap between current solutions and real-world requirements. In this work, we propose MESEN, a multimodal-empowered unimodal sensing framework, to utilize unlabeled multimodal data available during the HAR model design phase for unimodal HAR enhancement during the deployment phase. From a study on the impact of supervised …
abstract application applications arxiv challenges cs.lg current data design exploit function gap however human labels limitations multimodal multimodal data recognition requirements solutions type will work world
More from arxiv.org / cs.LG updates on arXiv.org
Jobs in AI, ML, Big Data
AI Research Scientist
@ Vara | Berlin, Germany and Remote
Data Architect
@ University of Texas at Austin | Austin, TX
Data ETL Engineer
@ University of Texas at Austin | Austin, TX
Lead GNSS Data Scientist
@ Lurra Systems | Melbourne
Senior Machine Learning Engineer (MLOps)
@ Promaton | Remote, Europe
AI Engineering Manager
@ M47 Labs | Barcelona, Catalunya [Cataluña], Spain