all AI news
Distributional Black-Box Model Inversion Attack with Multi-Agent Reinforcement Learning
April 23, 2024, 4:42 a.m. | Huan Bao, Kaimin Wei, Yongdong Wu, Jin Qian, Robert H. Deng
cs.LG updates on arXiv.org arxiv.org
Abstract: A Model Inversion (MI) attack based on Generative Adversarial Networks (GAN) aims to recover the private training data from complex deep learning models by searching codes in the latent space. However, they merely search a deterministic latent space such that the found latent code is usually suboptimal. In addition, the existing distributional MI schemes assume that an attacker can access the structures and parameters of the target model, which is not always viable in practice. …
abstract adversarial agent arxiv box code cs.cr cs.lg data deep learning found gan generative generative adversarial networks however multi-agent networks reinforcement reinforcement learning search searching space training training data type
More from arxiv.org / cs.LG updates on arXiv.org
Jobs in AI, ML, Big Data
AI Engineer Intern, Agents
@ Occam AI | US
AI Research Scientist
@ Vara | Berlin, Germany and Remote
Data Architect
@ University of Texas at Austin | Austin, TX
Data ETL Engineer
@ University of Texas at Austin | Austin, TX
Lead GNSS Data Scientist
@ Lurra Systems | Melbourne
Data Engineer - Takealot Group (Takealot.com | Superbalist.com | Mr D Food)
@ takealot.com | Cape Town