May 23, 2022, 1:12 a.m. | Tao Yang, Shenglong Zhou, Yuwang Wang, Yan Lu, Nanning Zheng

cs.CV updates on arXiv.org arxiv.org

Deep neural networks often suffer the data distribution shift between
training and testing, and the batch statistics are observed to reflect the
shift. In this paper, targeting of alleviating distribution shift in test time,
we revisit the batch normalization (BN) in the training process and reveals two
key insights benefiting test-time optimization: $(i)$ preserving the same
gradient backpropagation form as training, and $(ii)$ using dataset-level
statistics for robust optimization and inference. Based on the two insights, we
propose a novel …

arxiv normalization test time

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne

Senior Machine Learning Engineer (MLOps)

@ Promaton | Remote, Europe

Data Engineer

@ Chubb | Simsbury, CT, United States

Research Analyst , NA Light Vehicle Powertrain Forecasting

@ S&P Global | US - MI - VIRTUAL

Sr. Data Scientist - ML Ops Job

@ Yash Technologies | Indore, IN

Alternance-Data Management

@ Keolis | Courbevoie, FR, 92400