April 24, 2023, 12:44 a.m. | Lie He, Shiva Prasad Kasiviswanathan

stat.ML updates on arXiv.org arxiv.org

In this paper, we study the conditional stochastic optimization (CSO) problem
which covers a variety of applications including portfolio selection,
reinforcement learning, robust learning, causal inference, etc. The
sample-averaged gradient of the CSO objective is biased due to its nested
structure and therefore requires a high sample complexity to reach convergence.
We introduce a general stochastic extrapolation technique that effectively
reduces the bias. We show that for nonconvex smooth objectives, combining this
extrapolation with variance reduction techniques can achieve a …

algorithms applications arxiv bias causal inference complexity convergence cso etc general gradient inference optimization paper portfolio reinforcement reinforcement learning stochastic study variance

Lead Developer (AI)

@ Cere Network | San Francisco, US

Research Engineer

@ Allora Labs | Remote

Ecosystem Manager

@ Allora Labs | Remote

Founding AI Engineer, Agents

@ Occam AI | New York

AI Engineer Intern, Agents

@ Occam AI | US

AI Research Scientist

@ Vara | Berlin, Germany and Remote