Feb. 22, 2024, 5:42 a.m. | William N. Caballero, Jose Manuel Camacho, Tahir Ekin, Roi Naveiro

cs.LG updates on arXiv.org arxiv.org

arXiv:2402.13287v1 Announce Type: cross
Abstract: Time-series models typically assume untainted and legitimate streams of data. However, a self-interested adversary may have incentive to corrupt this data, thereby altering a decision maker's inference. Within the broader field of adversarial machine learning, this research provides a novel, probabilistic perspective toward the manipulation of hidden Markov model inferences via corrupted data. In particular, we provision a suite of corruption problems for filtering, smoothing, and decoding inferences leveraging an adversarial risk analysis approach. Multiple …

abstract adversarial adversarial machine learning arxiv cs.ai cs.cr cs.lg data decision hidden inference inferences machine machine learning maker manipulation markov novel perspective research series type

Founding AI Engineer, Agents

@ Occam AI | New York

AI Engineer Intern, Agents

@ Occam AI | US

AI Research Scientist

@ Vara | Berlin, Germany and Remote

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne