Dec. 15, 2023, 2:25 a.m. | /u/FallMindless3563

Machine Learning www.reddit.com

They state in the paper:

We highlight the most important connection: the classical gating mechanism of RNNs is an instance of our selection mechanism for SSMs.

Is it mainly the discretization step and different set of parameters in A,B, and C that are different?

Otherwise it feels like the same mental model to me. Encode information into a hidden space, use a gating or "selection" mechanism to figure out what to remember and forget, then unroll it over time to …

concepts gru highlight instance lstm machinelearning mamba paper parameters set state

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne

Senior Machine Learning Engineer (MLOps)

@ Promaton | Remote, Europe

Intern Large Language Models Planning (f/m/x)

@ BMW Group | Munich, DE

Data Engineer Analytics

@ Meta | Menlo Park, CA | Remote, US