Jan. 18, 2022, 12:26 p.m. | /u/mlvpj

Machine Learning www.reddit.com

How do you visualize attention especially when there are many tokens (256,512) with multiple layers and multiple heads? Most visualizations and frameworks I’ve tried fail when there are more than 100 tokens.

This is quite important since most nlp and vision models now use multi head attention.

submitted by /u/mlvpj
[link] [comments]

attention machinelearning

Data Engineer

@ Bosch Group | San Luis Potosí, Mexico

DATA Engineer (H/F)

@ Renault Group | FR REN RSAS - Le Plessis-Robinson (Siège)

Advisor, Data engineering

@ Desjardins | 1, Complexe Desjardins, Montréal

Data Engineer Intern

@ Getinge | Wayne, NJ, US

Software Engineer III- Java / Python / Pyspark / ETL

@ JPMorgan Chase & Co. | Jersey City, NJ, United States

Lead Data Engineer (Azure/AWS)

@ Telstra | Telstra ICC Bengaluru