March 19, 2024, 4:44 a.m. | Angelos Chatzimparmpas, Rafael M. Martins, Alexandru C. Telea, Andreas Kerren

cs.LG updates on arXiv.org arxiv.org

arXiv:2304.00133v4 Announce Type: replace
Abstract: As the complexity of machine learning (ML) models increases and their application in different (and critical) domains grows, there is a strong demand for more interpretable and trustworthy ML. A direct, model-agnostic, way to interpret such models is to train surrogate models-such as rule sets and decision trees-that sufficiently approximate the original ones while being simpler and easier-to-explain. Yet, rule sets can become very lengthy, with many if-else statements, and decision tree depth grows rapidly …

abstract analysis application arxiv behavior behavior analysis complexity cs.hc cs.lg decision demand domains machine machine learning machine learning models model-agnostic train trustworthy type

AI Research Scientist

@ Vara | Berlin, Germany and Remote

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne

Senior Machine Learning Engineer (MLOps)

@ Promaton | Remote, Europe

AI Engineering Manager

@ M47 Labs | Barcelona, Catalunya [Cataluña], Spain