all AI news
Model Compression in Practice: Lessons Learned from Practitioners Creating On-device Machine Learning Experiences
April 5, 2024, 4:43 a.m. | Fred Hohman, Mary Beth Kery, Donghao Ren, Dominik Moritz
cs.LG updates on arXiv.org arxiv.org
Abstract: On-device machine learning (ML) promises to improve the privacy, responsiveness, and proliferation of new, intelligent user experiences by moving ML computation onto everyday personal devices. However, today's large ML models must be drastically compressed to run efficiently on-device, a hurtle that requires deep, yet currently niche expertise. To engage the broader human-centered ML community in on-device ML experiences, we present the results from an interview study with 30 experts at Apple that specialize in producing …
abstract arxiv compression computation cs.ai cs.hc cs.lg devices however intelligent lessons learned machine machine learning ml models moving practice privacy type
More from arxiv.org / cs.LG updates on arXiv.org
Jobs in AI, ML, Big Data
AI Research Scientist
@ Vara | Berlin, Germany and Remote
Data Architect
@ University of Texas at Austin | Austin, TX
Data ETL Engineer
@ University of Texas at Austin | Austin, TX
Lead GNSS Data Scientist
@ Lurra Systems | Melbourne
Senior Machine Learning Engineer (MLOps)
@ Promaton | Remote, Europe
Senior Software Engineer, Generative AI (C++)
@ SoundHound Inc. | Toronto, Canada