Oct. 11, 2023, 2 p.m. | /u/Successful-Western27

Machine Learning www.reddit.com

Transformers are great at NLP and computer vision tasks, but I was surprised to learn they still lag behind simple linear models at time series forecasting.

The issue is how most Transformer architectures treat each timestamp as a token and fuse all the variable data from that moment. This makes two big problems:

* **Variables recorded at slightly different times get blurred together,** losing important timing info
* **Each token can only see a single moment,** no long-term dependencies

So …

architectures computer computer vision data forecasting issue learn linear machinelearning nlp series simple tasks time series time series forecasting token transformer transformers tsinghua university university vision

Senior Machine Learning Engineer

@ GPTZero | Toronto, Canada

ML/AI Engineer / NLP Expert - Custom LLM Development (x/f/m)

@ HelloBetter | Remote

Doctoral Researcher (m/f/div) in Automated Processing of Bioimages

@ Leibniz Institute for Natural Product Research and Infection Biology (Leibniz-HKI) | Jena

Seeking Developers and Engineers for AI T-Shirt Generator Project

@ Chevon Hicks | Remote

Technical Program Manager, Expert AI Trainer Acquisition & Engagement

@ OpenAI | San Francisco, CA

Director, Data Engineering

@ PatientPoint | Cincinnati, Ohio, United States