all AI news
A Post-Training Approach for Mitigating Overfitting in Quantum Convolutional Neural Networks
March 5, 2024, 2:45 p.m. | Aakash Ravindra Shinde, Charu Jain, Amir Kalev
cs.LG updates on arXiv.org arxiv.org
Abstract: Quantum convolutional neural network (QCNN), an early application for quantum computers in the NISQ era, has been consistently proven successful as a machine learning (ML) algorithm for several tasks with significant accuracy. Derived from its classical counterpart, QCNN is prone to overfitting. Overfitting is a typical shortcoming of ML models that are trained too closely to the availed training dataset and perform relatively poorly on unseen datasets for a similar problem. In this work we …
abstract accuracy algorithm application arxiv computers convolutional neural network convolutional neural networks cs.lg machine machine learning network networks neural network neural networks nisq overfitting quant-ph quantum quantum computers tasks training type
More from arxiv.org / cs.LG updates on arXiv.org
Jobs in AI, ML, Big Data
Data Architect
@ University of Texas at Austin | Austin, TX
Data ETL Engineer
@ University of Texas at Austin | Austin, TX
Lead GNSS Data Scientist
@ Lurra Systems | Melbourne
Senior Machine Learning Engineer (MLOps)
@ Promaton | Remote, Europe
AIML - Sr Machine Learning Engineer, Data and ML Innovation
@ Apple | Seattle, WA, United States
Senior Data Engineer
@ Palta | Palta Cyprus, Palta Warsaw, Palta remote