Feb. 13, 2024, 5:42 a.m. | Francisco Dur\'an Silverio Mart\'inez-Fern\'andez Matias Martinez Patricia Lago

cs.LG updates on arXiv.org arxiv.org

The growing use of large machine learning models highlights concerns about their increasing computational demands. While the energy consumption of their training phase has received attention, fewer works have considered the inference phase. For ML inference, the binding of ML models to the ML system for user access, known as ML serving, is a critical yet understudied step for achieving efficiency in ML applications.
We examine the literature in ML architectural design decisions and Green AI, with a special focus …

attention computational concerns consumption cs.lg cs.se decisions design energy green highlights inference machine machine learning machine learning models ml inference ml models training

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne

Senior Machine Learning Engineer (MLOps)

@ Promaton | Remote, Europe

Principal Data Engineering Manager

@ Microsoft | Redmond, Washington, United States

Machine Learning Engineer

@ Apple | San Diego, California, United States