all AI news
Self attention with better complexity?
April 14, 2023, 4:24 p.m. | /u/abstract000
Deep Learning www.reddit.com
Those lase weeks I looked at papers trying to reduce self attention complexity. The first was LongFormer. As I love the idea in the paper, I think the implementation is currently impossible as it would need sparse tensors. We tried those at work and have no speedup if the tensor is not VERY sparse. If you have a good way to deal …
attention complexity deal deeplearning good implementation layoutlm love paper reduce tensor think transformers work
More from www.reddit.com / Deep Learning
How does gradient descent work in random forest
1 day, 19 hours ago |
www.reddit.com
Prerequisites for jumping into transformers?
1 day, 21 hours ago |
www.reddit.com
[Reading] Deeplearning by goodfellow
2 days, 3 hours ago |
www.reddit.com
Linearizing Large Language Models
2 days, 19 hours ago |
www.reddit.com
Converting Soft tokens to Hard tokens in Llama2
2 days, 21 hours ago |
www.reddit.com
Detection of free parking spaces
3 days, 4 hours ago |
www.reddit.com
Jobs in AI, ML, Big Data
Software Engineer for AI Training Data (School Specific)
@ G2i Inc | Remote
Software Engineer for AI Training Data (Python)
@ G2i Inc | Remote
Software Engineer for AI Training Data (Tier 2)
@ G2i Inc | Remote
Data Engineer
@ Lemon.io | Remote: Europe, LATAM, Canada, UK, Asia, Oceania
Artificial Intelligence – Bioinformatic Expert
@ University of Texas Medical Branch | Galveston, TX
Lead Developer (AI)
@ Cere Network | San Francisco, US