July 31, 2022, 3:55 a.m. | /u/Megixist

Machine Learning www.reddit.com

As per the transformer paper, sinusoidal embeddings help inference on longer sequences than the ones it was trained on. This isn't specific to transformers and this property has been extensively used for time series modeling in the past. From what I can see, this is due to the oscillatory property of sinusoidal waves which can be combined in specific manners to embed temporal information. This makes a lot of sense but has there been any method to embed temporal information …

encoding machinelearning temporal

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne

Senior Machine Learning Engineer (MLOps)

@ Promaton | Remote, Europe

Principal Engineer, Deep Learning

@ Outrider | Remote

Data Analyst (Bangkok based, relocation provided)

@ Agoda | Bangkok (Central World Office)

Data Scientist II

@ MoEngage | Bengaluru

Machine Learning Engineer

@ Sika AG | Welwyn Garden City, United Kingdom