all AI news
GELU : Gaussian Error Linear Unit Code (Python, TF, Torch)
Oct. 17, 2022, 8:02 p.m. | Poulinakis Kon
Towards AI - Medium pub.towardsai.net
GELU : Gaussian Error Linear Unit Code (Python, TF, Torch)
Code tutorial for GELU, Gaussian Error Linear Unit activation function. Includes bare python, Tensorflow and Pytorch code.
Photo by Markus Winkler on UnsplashGELU Activation Function
Gaussian Error Linear Unit, GELU, is the most-used activation function in state-of-the-art models including BERT, GPT, Vision Transformers, etc..
If you want to understand the intuition and math behind GELU I suggest you check my previous article covering the GELU paper (GELU, the ReLU …
activation-functions code deep learning error linear neural networks python
More from pub.towardsai.net / Towards AI - Medium
Best Resources to Learn & Understand Evaluating LLMs
2 days, 15 hours ago |
pub.towardsai.net
Deploying Your Models (Cheap and Dirty Way) Using Binder
2 days, 17 hours ago |
pub.towardsai.net
Learn AI Together — Towards AI Community Newsletter #22
3 days, 14 hours ago |
pub.towardsai.net
Jobs in AI, ML, Big Data
Founding AI Engineer, Agents
@ Occam AI | New York
AI Engineer Intern, Agents
@ Occam AI | US
AI Research Scientist
@ Vara | Berlin, Germany and Remote
Data Architect
@ University of Texas at Austin | Austin, TX
Data ETL Engineer
@ University of Texas at Austin | Austin, TX
Codec Avatars Research Engineer
@ Meta | Pittsburgh, PA