July 13, 2023, 1:49 a.m. | /u/Western-Image7125

Machine Learning www.reddit.com

I understand that encoder only models (like BERT etc) are mainly for learning representations of words taking context on both sides. What I’m confused about is why you would need decoder only vs encoder-decoder models. GPT and Bloom are decoder only while I think T5 is enc-dec, I’m not sure why you would use one vs the other. Intuitively enc-dec model has more parameters and should be better at tasks where you have both complex text input and output, like …

bert bloom context decoder encoder encoder-decoder etc gpt machinelearning think words

Senior Machine Learning Engineer

@ GPTZero | Toronto, Canada

ML/AI Engineer / NLP Expert - Custom LLM Development (x/f/m)

@ HelloBetter | Remote

Doctoral Researcher (m/f/div) in Automated Processing of Bioimages

@ Leibniz Institute for Natural Product Research and Infection Biology (Leibniz-HKI) | Jena

Seeking Developers and Engineers for AI T-Shirt Generator Project

@ Chevon Hicks | Remote

Senior Applied Data Scientist

@ dunnhumby | London

Principal Data Architect - Azure & Big Data

@ MGM Resorts International | Home Office - US, NV