May 16, 2022, 9:24 p.m. | /u/KarlKani44

Machine Learning www.reddit.com

I recently studied ways to improve the training time of big neural networks, especially ResNets. On my way, I could not help but notice the big claims you can find in the paper [Deep Networks with Stochastic Depth](https://arxiv.org/pdf/1603.09382.pdf)

To summarize informally, their contribution consists of a new hyperparameter for ResBlocks, which is used to skip the inner part of the residual connection with the given probability (they use 0.5 in their experiments). Quoting from the Paper:

> Let b ∈ …

machinelearning stochastic

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne

Senior Machine Learning Engineer (MLOps)

@ Promaton | Remote, Europe

Data Management Associate

@ EcoVadis | Ebène, Mauritius

Senior Data Engineer

@ Telstra | Telstra ICC Bengaluru