all AI news
Improving the performance of weak supervision searches using transfer and meta-learning
March 4, 2024, 5:43 a.m. | Hugues Beauchesne, Zong-En Chen, Cheng-Wei Chiang
cs.LG updates on arXiv.org arxiv.org
Abstract: Weak supervision searches have in principle the advantages of both being able to train on experimental data and being able to learn distinctive signal properties. However, the practical applicability of such searches is limited by the fact that successfully training a neural network via weak supervision can require a large amount of signal. In this work, we seek to create neural networks that can learn from less experimental signal by using transfer and meta-learning. The …
abstract advantages arxiv cs.lg data experimental hep-ph learn meta meta-learning network neural network performance practical signal supervision train training transfer type via
More from arxiv.org / cs.LG updates on arXiv.org
Jobs in AI, ML, Big Data
AI Research Scientist
@ Vara | Berlin, Germany and Remote
Data Architect
@ University of Texas at Austin | Austin, TX
Data ETL Engineer
@ University of Texas at Austin | Austin, TX
Lead GNSS Data Scientist
@ Lurra Systems | Melbourne
Data Science Analyst
@ Mayo Clinic | AZ, United States
Sr. Data Scientist (Network Engineering)
@ SpaceX | Redmond, WA