July 4, 2022, 6:34 p.m. | Rachit Jain

Towards Data Science - Medium towardsdatascience.com

go_backwards() — Unravelling its ‘hidden’ secrets

Understanding its hidden nuances & exploring its leaky nature!

Representation of an LSTM cell | Image by Christopher Olah

Introduction

Long Short Term Memory (LSTM) are superior versions of Recurrent Neural Networks (RNN) and are capable of storing ‘context’, as the name suggests, over relatively long sequences. This allows them to be a perfect utility for NLP tasks such as document classification, speech recognition, Named Entity Recognition (NER), etc.

In many applications, such as …

bi-lstm elmo lstm next-word-prediction tensorflow transfer learning

Founding AI Engineer, Agents

@ Occam AI | New York

AI Engineer Intern, Agents

@ Occam AI | US

AI Research Scientist

@ Vara | Berlin, Germany and Remote

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Sr. BI Analyst

@ AkzoNobel | Pune, IN