all AI news
Sleep-Like Unsupervised Replay Improves Performance when Data are Limited or Unbalanced
Feb. 20, 2024, 5:43 a.m. | Anthony Bazhenov, Pahan Dewasurendra, Giri Krishnan, Jean Erik Delanois
cs.LG updates on arXiv.org arxiv.org
Abstract: The performance of artificial neural networks (ANNs) degrades when training data are limited or imbalanced. In contrast, the human brain can learn quickly from just a few examples. Here, we investigated the role of sleep in improving the performance of ANNs trained with limited data on the MNIST and Fashion MNIST datasets. Sleep was implemented as an unsupervised phase with local Hebbian type learning rules. We found a significant boost in accuracy after the sleep …
abstract anns artificial artificial neural networks arxiv brain contrast cs.lg cs.ne data examples human learn networks neural networks performance role sleep training training data type unsupervised
More from arxiv.org / cs.LG updates on arXiv.org
Jobs in AI, ML, Big Data
ML/AI Engineer / NLP Expert - Custom LLM Development (x/f/m)
@ HelloBetter | Remote
Doctoral Researcher (m/f/div) in Automated Processing of Bioimages
@ Leibniz Institute for Natural Product Research and Infection Biology (Leibniz-HKI) | Jena
Seeking Developers and Engineers for AI T-Shirt Generator Project
@ Chevon Hicks | Remote
Cloud Data Platform Engineer
@ First Central | Home Office (Remote)
Associate Director, Data Science
@ MSD | USA - New Jersey - Rahway
Data Scientist Sr.
@ MSD | CHL - Santiago - Santiago (Calle Mariano)