Dec. 4, 2023, 3:12 a.m. | /u/suspicious_beam

Machine Learning www.reddit.com

I have been learning NLP for a little while, and always found it hard to find full implementations of Transformers with **explicit forward and backprop**. This project was my attempt at building that, while learning how to better understand optimization in Transformers.

It is as well-documented as I could make it, and can be trained or fine-tuned easily by editing the **config.py** file and running the **run.py** script. Get the code with:

git clone https://github.com/eduardoleao052/Transformer-from-scratch.git

I managed to generate some …

autograd building educational found machinelearning nlp optimization project transformer transformers

Artificial Intelligence – Bioinformatic Expert

@ University of Texas Medical Branch | Galveston, TX

Lead Developer (AI)

@ Cere Network | San Francisco, US

Research Engineer

@ Allora Labs | Remote

Ecosystem Manager

@ Allora Labs | Remote

Founding AI Engineer, Agents

@ Occam AI | New York

AI Engineer Intern, Agents

@ Occam AI | US