all AI news
On Output Activation Functions for Adversarial Losses: A Theoretical Analysis via Variational Divergence Minimization and An Empirical Study on MNIST Classification. (arXiv:1901.08753v3 [cs.LG] UPDATED)
Nov. 8, 2022, 2:13 a.m. | Hao-Wen Dong, Yi-Hsuan Yang
stat.ML updates on arXiv.org arxiv.org
Recent years have seen adversarial losses been applied to many fields. Their
applications extend beyond the originally proposed generative modeling to
conditional generative and discriminative settings. While prior work has
proposed various output activation functions and regularization approaches,
some open questions still remain unanswered. In this paper, we aim to study the
following two research questions: 1) What types of output activation functions
form a well-behaved adversarial loss? 2) How different combinations of output
activation functions and regularization approaches perform …
More from arxiv.org / stat.ML updates on arXiv.org
Mixture of partially linear experts
an hour ago |
arxiv.org
Adaptive deep learning for nonlinear time series models
1 day, 1 hour ago |
arxiv.org
A Full Adagrad algorithm with O(Nd) operations
1 day, 1 hour ago |
arxiv.org
Minimax Regret Learning for Data with Heterogeneous Subgroups
1 day, 1 hour ago |
arxiv.org
Jobs in AI, ML, Big Data
Founding AI Engineer, Agents
@ Occam AI | New York
AI Engineer Intern, Agents
@ Occam AI | US
AI Research Scientist
@ Vara | Berlin, Germany and Remote
Data Architect
@ University of Texas at Austin | Austin, TX
Data ETL Engineer
@ University of Texas at Austin | Austin, TX
Sr. BI Analyst
@ AkzoNobel | Pune, IN