all AI news
Interpretability in Symbolic Regression: a benchmark of Explanatory Methods using the Feynman data set
April 10, 2024, 4:41 a.m. | Guilherme Seidyo Imai Aldeia (Federal University of ABC), Fabricio Olivetti de Franca (Federal University of ABC)
cs.LG updates on arXiv.org arxiv.org
Abstract: In some situations, the interpretability of the machine learning models plays a role as important as the model accuracy. Interpretability comes from the need to trust the prediction model, verify some of its properties, or even enforce them to improve fairness. Many model-agnostic explanatory methods exists to provide explanations for black-box models. In the regression task, the practitioner can use white-boxes or gray-boxes models to achieve more interpretable results, which is the case of symbolic …
abstract accuracy arxiv benchmark cs.ai cs.lg data data set fairness feynman interpretability machine machine learning machine learning models model accuracy prediction regression role set them trust type verify
More from arxiv.org / cs.LG updates on arXiv.org
Jobs in AI, ML, Big Data
Software Engineer for AI Training Data (School Specific)
@ G2i Inc | Remote
Software Engineer for AI Training Data (Python)
@ G2i Inc | Remote
Software Engineer for AI Training Data (Tier 2)
@ G2i Inc | Remote
Data Engineer
@ Lemon.io | Remote: Europe, LATAM, Canada, UK, Asia, Oceania
Artificial Intelligence – Bioinformatic Expert
@ University of Texas Medical Branch | Galveston, TX
Lead Developer (AI)
@ Cere Network | San Francisco, US