June 29, 2022, 1:58 a.m. | /u/FlyingQuokka

Natural Language Processing www.reddit.com

For context: my research is in applied ML, but not NLP. I'm running experiments to test if, in my field, word/document embeddings from transformers are better than word2vec. As of now, I'm using BERT, but it seems [that was a poor choice](https://www.reddit.com/r/MachineLearning/comments/vm2sti/n_inverse_scaling_prize_250k_in_prizes_for/ie2jy90/). Which brings me to this sub's expertise--is there a class of transformers (T5, RoBERTa, etc.) that are better for embeddings that are used in downstream tasks?

languagetechnology word embeddings

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne

Senior Machine Learning Engineer (MLOps)

@ Promaton | Remote, Europe

Senior Business Intelligence Developer / Analyst

@ Transamerica | Work From Home, USA

Data Analyst (All Levels)

@ Noblis | Bethesda, MD, United States