March 12, 2024, 4:47 a.m. | Frincy Clement, Ji Yang, Irene Cheng

cs.CV updates on arXiv.org arxiv.org

arXiv:2403.05658v1 Announce Type: new
Abstract: Deep Neural Networks have often been called the black box because of the complex, deep architecture and non-transparency presented by the inner layers. There is a lack of trust to use Artificial Intelligence in critical and high-precision fields such as security, finance, health, and manufacturing industries. A lot of focused work has been done to provide interpretable models, intending to deliver meaningful insights into the thoughts and behavior of neural networks. In our research, we …

abstract architecture artificial artificial intelligence arxiv black box box classification cs.ai cs.cv cs.mm feature fields finance health image industries intelligence interpretable ai manufacturing networks neural networks precision security transparency trust type

Software Engineer for AI Training Data (School Specific)

@ G2i Inc | Remote

Software Engineer for AI Training Data (Python)

@ G2i Inc | Remote

Software Engineer for AI Training Data (Tier 2)

@ G2i Inc | Remote

Data Engineer

@ Lemon.io | Remote: Europe, LATAM, Canada, UK, Asia, Oceania

Artificial Intelligence – Bioinformatic Expert

@ University of Texas Medical Branch | Galveston, TX

Lead Developer (AI)

@ Cere Network | San Francisco, US