Feb. 26, 2024, 1 p.m. | /u/mono1110

Deep Learning www.reddit.com

Quick context. I am trying to solve a very easy problem with self attention.

Let me provide the approaches I tried.

1. Training data was one hot encoded, then positional encoding was added. After that self attention followed by two feedforward layers.

2. Training data was label encoded, passed through an embedding layer, positional encoding was added. After that it was passed through self attention followed by two feedforward layers.

Now the results.
In the first approach, the training accuracy …

attention context data deeplearning easy encoding faster hot positional encoding solve training training data

AI Engineer Intern, Agents

@ Occam AI | US

AI Research Scientist

@ Vara | Berlin, Germany and Remote

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne

Data Engineer - Takealot Group (Takealot.com | Superbalist.com | Mr D Food)

@ takealot.com | Cape Town