all AI news
Power-law Scaling to Assist with Key Challenges in Artificial Intelligence. (arXiv:2211.08430v1 [cs.LG])
Nov. 17, 2022, 2:11 a.m. | Yuval Meir, Shira Sardi, Shiri Hodassman, Karin Kisos, Itamar Ben-Noam, Amir Goldental, Ido Kanter
cs.LG updates on arXiv.org arxiv.org
Power-law scaling, a central concept in critical phenomena, is found to be
useful in deep learning, where optimized test errors on handwritten digit
examples converge as a power-law to zero with database size. For rapid decision
making with one training epoch, each example is presented only once to the
trained network, the power-law exponent increased with the number of hidden
layers. For the largest dataset, the obtained test error was estimated to be in
the proximity of state-of-the-art algorithms for …
artificial artificial intelligence arxiv challenges intelligence law power power-law scaling
More from arxiv.org / cs.LG updates on arXiv.org
Jobs in AI, ML, Big Data
Data Architect
@ University of Texas at Austin | Austin, TX
Data ETL Engineer
@ University of Texas at Austin | Austin, TX
Lead GNSS Data Scientist
@ Lurra Systems | Melbourne
Senior Machine Learning Engineer (MLOps)
@ Promaton | Remote, Europe
#13721 - Data Engineer - AI Model Testing
@ Qualitest | Miami, Florida, United States
Elasticsearch Administrator
@ ManTech | 201BF - Customer Site, Chantilly, VA