all AI news
Modeling Output-Level Task Relatedness in Multi-Task Learning with Feedback Mechanism
April 2, 2024, 7:42 p.m. | Xiangming Xi, Feng Gao, Jun Xu, Fangtai Guo, Tianlei Jin
cs.LG updates on arXiv.org arxiv.org
Abstract: Multi-task learning (MTL) is a paradigm that simultaneously learns multiple tasks by sharing information at different levels, enhancing the performance of each individual task. While previous research has primarily focused on feature-level or parameter-level task relatedness, and proposed various model architectures and learning algorithms to improve learning performance, we aim to explore output-level task relatedness. This approach introduces a posteriori information into the model, considering that different tasks may produce correlated outputs with mutual influences. …
abstract algorithms architectures arxiv cs.lg feature feedback information modeling multiple multi-task learning paradigm performance research tasks type
More from arxiv.org / cs.LG updates on arXiv.org
Jobs in AI, ML, Big Data
AI Research Scientist
@ Vara | Berlin, Germany and Remote
Data Architect
@ University of Texas at Austin | Austin, TX
Data ETL Engineer
@ University of Texas at Austin | Austin, TX
Lead GNSS Data Scientist
@ Lurra Systems | Melbourne
Senior Data Engineer (m/f/d)
@ Project A Ventures | Berlin, Germany
Principle Research Scientist
@ Analog Devices | US, MA, Boston