all AI news
[R] Preventing mode collapse and overfitting in Seq2Seq transformers
Sept. 30, 2022, 9:07 p.m. | /u/vikigenius
Machine Learning www.reddit.com
Eg:
S1 -> T1
S2 -> T1
S3 -> T1
I tried to use `BartForConditionalGeneration` and `T5` models as `Seq2Seq` models.
But I notice that after a single epoch the model starts collapsing to produce the same sentence as output regardless of input. And this most likely happens because of the repetition. If I simply eleiminate all repetition in my dataset by picking only …
More from www.reddit.com / Machine Learning
Jobs in AI, ML, Big Data
Lead GNSS Data Scientist
@ Lurra Systems | Melbourne
Senior Machine Learning Engineer (MLOps)
@ Promaton | Remote, Europe
Lead Data Engineer
@ JPMorgan Chase & Co. | Jersey City, NJ, United States
Senior Machine Learning Engineer
@ TELUS | Vancouver, BC, CA
CT Technologist - Ambulatory Imaging - PRN
@ Duke University | Morriville, NC, US, 27560
BH Data Analyst
@ City of Philadelphia | Philadelphia, PA, United States