all AI news
Sleep-Like Unsupervised Replay Improves Performance when Data are Limited or Unbalanced
Feb. 20, 2024, 5:43 a.m. | Anthony Bazhenov, Pahan Dewasurendra, Giri Krishnan, Jean Erik Delanois
cs.LG updates on arXiv.org arxiv.org
Abstract: The performance of artificial neural networks (ANNs) degrades when training data are limited or imbalanced. In contrast, the human brain can learn quickly from just a few examples. Here, we investigated the role of sleep in improving the performance of ANNs trained with limited data on the MNIST and Fashion MNIST datasets. Sleep was implemented as an unsupervised phase with local Hebbian type learning rules. We found a significant boost in accuracy after the sleep …
abstract anns artificial artificial neural networks arxiv brain contrast cs.lg cs.ne data examples human learn networks neural networks performance role sleep training training data type unsupervised
More from arxiv.org / cs.LG updates on arXiv.org
Jobs in AI, ML, Big Data
Data Architect
@ University of Texas at Austin | Austin, TX
Data ETL Engineer
@ University of Texas at Austin | Austin, TX
Lead GNSS Data Scientist
@ Lurra Systems | Melbourne
Senior Machine Learning Engineer (MLOps)
@ Promaton | Remote, Europe
Software Engineer, Data Tools - Full Stack
@ DoorDash | Pune, India
Senior Data Analyst
@ Artsy | New York City