all AI news
Distributional Black-Box Model Inversion Attack with Multi-Agent Reinforcement Learning
April 23, 2024, 4:42 a.m. | Huan Bao, Kaimin Wei, Yongdong Wu, Jin Qian, Robert H. Deng
cs.LG updates on arXiv.org arxiv.org
Abstract: A Model Inversion (MI) attack based on Generative Adversarial Networks (GAN) aims to recover the private training data from complex deep learning models by searching codes in the latent space. However, they merely search a deterministic latent space such that the found latent code is usually suboptimal. In addition, the existing distributional MI schemes assume that an attacker can access the structures and parameters of the target model, which is not always viable in practice. …
abstract adversarial agent arxiv box code cs.cr cs.lg data deep learning found gan generative generative adversarial networks however multi-agent networks reinforcement reinforcement learning search searching space training training data type
More from arxiv.org / cs.LG updates on arXiv.org
Jobs in AI, ML, Big Data
Data Engineer
@ Lemon.io | Remote: Europe, LATAM, Canada, UK, Asia, Oceania
Artificial Intelligence – Bioinformatic Expert
@ University of Texas Medical Branch | Galveston, TX
Lead Developer (AI)
@ Cere Network | San Francisco, US
Research Engineer
@ Allora Labs | Remote
Ecosystem Manager
@ Allora Labs | Remote
Founding AI Engineer, Agents
@ Occam AI | New York