Feb. 24, 2024, 6:19 p.m. | /u/mgostIH

Machine Learning www.reddit.com

I was thinking on how to visualize the layernorm for a vector, and figured out that it's just applying two projections when the learned parameters are null (if not, then they just add a rescaling and translation).

You project onto the hyperplane where the mean (or sum) of components is null (e.g. in 3D x + y + z = 0) by removing the mean and then project onto the sphere of radius sqrt(D) by dividing by the standard deviation. …

components hyperplane machinelearning mean null parameters project thinking translation vector

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne

Senior Machine Learning Engineer (MLOps)

@ Promaton | Remote, Europe

Principal Data Engineering Manager

@ Microsoft | Redmond, Washington, United States

Machine Learning Engineer

@ Apple | San Diego, California, United States