June 7, 2022, 1:11 a.m. | Luqin Gan, Lili Zheng, Genevera I. Allen

stat.ML updates on arXiv.org arxiv.org

In order to trust machine learning for high-stakes problems, we need models
to be both reliable and interpretable. Recently, there has been a growing body
of work on interpretable machine learning which generates human understandable
insights into data, models, or predictions. At the same time, there has been
increased interest in quantifying the reliability and uncertainty of machine
learning predictions, often in the form of confidence intervals for predictions
using conformal inference. Yet, there has been relatively little attention
given …

arxiv confidence feature importance inference learning machine machine learning ml model-agnostic

AI Research Scientist

@ Vara | Berlin, Germany and Remote

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne

Senior Machine Learning Engineer (MLOps)

@ Promaton | Remote, Europe

AI Engineering Manager

@ M47 Labs | Barcelona, Catalunya [Cataluña], Spain