July 12, 2023, 5:21 p.m. | /u/eternalmathstudent

Computer Vision www.reddit.com

I just started reading about Transformers model. I have barely scratched the surface of this concept. For starters, I have the following 2 questions

1. How positional encoding are incorporated in the transformer model? I see that immediately after the word embedding, they have positional encoding. But I'm not getting in which part of the entire network it is being used?

2. For a given sentence, the weight matrices of the query, key and value, all of these 3 have …

computervision concept embedding encoding network part positional encoding questions reading transformer transformer model transformers word

Founding AI Engineer, Agents

@ Occam AI | New York

AI Engineer Intern, Agents

@ Occam AI | US

AI Research Scientist

@ Vara | Berlin, Germany and Remote

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne