all AI news
The Model Openness Framework: Promoting Completeness and Openness for Reproducibility, Transparency and Usability in AI
March 21, 2024, 4:42 a.m. | Matt White (Yanglet), Ibrahim Haddad (Yanglet), Cailean Osborne (Yanglet), Xiao-Yang (Yanglet), Liu, Ahmed Abdelmonsef, Sachin Varghese
cs.LG updates on arXiv.org arxiv.org
Abstract: Generative AI (GAI) offers unprecedented possibilities but its commercialization has raised concerns about transparency, reproducibility, bias, and safety. Many "open-source" GAI models lack the necessary components for full understanding and reproduction, and some use restrictive licenses, a practice known as "openwashing." We propose the Model Openness Framework (MOF), a ranked classification system that rates machine learning models based on their completeness and openness, following principles of open science, open source, open data, and open access. …
abstract arxiv bias components concerns cs.ai cs.cy cs.lg cs.se framework gai generative practice reproducibility restrictive safety transparency type understanding usability
More from arxiv.org / cs.LG updates on arXiv.org
Jobs in AI, ML, Big Data
AI Research Scientist
@ Vara | Berlin, Germany and Remote
Data Architect
@ University of Texas at Austin | Austin, TX
Data ETL Engineer
@ University of Texas at Austin | Austin, TX
Lead GNSS Data Scientist
@ Lurra Systems | Melbourne
Data Science Analyst
@ Mayo Clinic | AZ, United States
Sr. Data Scientist (Network Engineering)
@ SpaceX | Redmond, WA