all AI news
Learning in Convolutional Neural Networks Accelerated by Transfer Entropy
April 5, 2024, 4:41 a.m. | Adrian Moldovan, Angel Ca\c{t}aron, R\u{a}zvan Andonie
cs.LG updates on arXiv.org arxiv.org
Abstract: Recently, there is a growing interest in applying Transfer Entropy (TE) in quantifying the effective connectivity between artificial neurons. In a feedforward network, the TE can be used to quantify the relationships between neuron output pairs located in different layers. Our focus is on how to include the TE in the learning mechanisms of a Convolutional Neural Network (CNN) architecture. We introduce a novel training mechanism for CNN architectures which integrates the TE feedback connections. …
abstract artificial arxiv connectivity convolutional neural networks cs.ai cs.it cs.lg entropy focus math.it network networks neural networks neuron neurons relationships transfer type
More from arxiv.org / cs.LG updates on arXiv.org
Jobs in AI, ML, Big Data
Software Engineer for AI Training Data (School Specific)
@ G2i Inc | Remote
Software Engineer for AI Training Data (Python)
@ G2i Inc | Remote
Software Engineer for AI Training Data (Tier 2)
@ G2i Inc | Remote
Data Engineer
@ Lemon.io | Remote: Europe, LATAM, Canada, UK, Asia, Oceania
Artificial Intelligence – Bioinformatic Expert
@ University of Texas Medical Branch | Galveston, TX
Lead Developer (AI)
@ Cere Network | San Francisco, US