Sept. 30, 2022, 9:07 p.m. | /u/vikigenius

Machine Learning www.reddit.com

I have a Seq2Seq dataset where a given target sentence can occur for multiple source sentences.

Eg:

S1 -> T1
S2 -> T1
S3 -> T1

I tried to use `BartForConditionalGeneration` and `T5` models as `Seq2Seq` models.

But I notice that after a single epoch the model starts collapsing to produce the same sentence as output regardless of input. And this most likely happens because of the repetition. If I simply eleiminate all repetition in my dataset by picking only …

machinelearning overfitting seq2seq transformers

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne

Senior Machine Learning Engineer (MLOps)

@ Promaton | Remote, Europe

Lead Data Engineer

@ JPMorgan Chase & Co. | Jersey City, NJ, United States

Senior Machine Learning Engineer

@ TELUS | Vancouver, BC, CA

CT Technologist - Ambulatory Imaging - PRN

@ Duke University | Morriville, NC, US, 27560

BH Data Analyst

@ City of Philadelphia | Philadelphia, PA, United States