July 4, 2022, 1:11 a.m. | Mohamad Rida Rammal, Alessandro Achille, Aditya Golatkar, Suhas Diggavi, Stefano Soatto

cs.LG updates on arXiv.org arxiv.org

We derive information theoretic generalization bounds for supervised learning
algorithms based on a new measure of leave-one-out conditional mutual
information (loo-CMI). Contrary to other CMI bounds, which are black-box bounds
that do not exploit the structure of the problem and may be hard to evaluate in
practice, our loo-CMI bounds can be computed easily and can be interpreted in
connection to other notions such as classical leave-one-out cross-validation,
stability of the optimization algorithm, and the geometry of the
loss-landscape. It …

arxiv information leave-one-out lg

Software Engineer for AI Training Data (School Specific)

@ G2i Inc | Remote

Software Engineer for AI Training Data (Python)

@ G2i Inc | Remote

Software Engineer for AI Training Data (Tier 2)

@ G2i Inc | Remote

Data Engineer

@ Lemon.io | Remote: Europe, LATAM, Canada, UK, Asia, Oceania

Artificial Intelligence – Bioinformatic Expert

@ University of Texas Medical Branch | Galveston, TX

Lead Developer (AI)

@ Cere Network | San Francisco, US