all AI news
Topic: transformer architecture
Transformers, Contextualism, and Polysemy
2 days, 20 hours ago |
arxiv.org
Revealing Trends in Datasets from the 2022 ACL and EMNLP Conferences
2 days, 20 hours ago |
arxiv.org
TransformerFAM: Feedback attention is working memory
2 days, 20 hours ago |
arxiv.org
The Illusion of State in State-Space Models
2 days, 20 hours ago |
arxiv.org
Adaptive Query Prompting for Multi-Domain Landmark Detection
2 weeks, 2 days ago |
arxiv.org
The FIRST Production-grade Mamba-based LLM!!!
2 weeks, 4 days ago |
www.youtube.com
Jamba: The LLM with Mamba Mentality
2 weeks, 6 days ago |
gradientflow.com
SEA: Sparse Linear Attention with Estimated Attention Mask
3 weeks, 2 days ago |
arxiv.org
Kernel-Elastic Autoencoder for Molecular Design
3 weeks, 2 days ago |
arxiv.org
CFAT: Unleashing TriangularWindows for Image Super-resolution
3 weeks, 2 days ago |
arxiv.org
Stronger Graph Transformer with Regularized Attention Scores
4 weeks, 1 day ago |
arxiv.org
Looped Transformers are Better at Learning Learning Algorithms
4 weeks, 2 days ago |
arxiv.org
Magnushammer: A Transformer-Based Approach to Premise Selection
4 weeks, 2 days ago |
arxiv.org
Merging Text Transformer Models from Different Initializations
1 month, 1 week ago |
arxiv.org
Transformers, Contextualism, and Polysemy
2 days, 20 hours ago |
arxiv.org
TransformerFAM: Feedback attention is working memory
2 days, 20 hours ago |
arxiv.org
Items published with this topic over the last 90 days.
Latest
Transformers, Contextualism, and Polysemy
2 days, 20 hours ago |
arxiv.org
Revealing Trends in Datasets from the 2022 ACL and EMNLP Conferences
2 days, 20 hours ago |
arxiv.org
TransformerFAM: Feedback attention is working memory
2 days, 20 hours ago |
arxiv.org
The Illusion of State in State-Space Models
2 days, 20 hours ago |
arxiv.org
Adaptive Query Prompting for Multi-Domain Landmark Detection
2 weeks, 2 days ago |
arxiv.org
The FIRST Production-grade Mamba-based LLM!!!
2 weeks, 4 days ago |
www.youtube.com
Jamba: The LLM with Mamba Mentality
2 weeks, 6 days ago |
gradientflow.com
SEA: Sparse Linear Attention with Estimated Attention Mask
3 weeks, 2 days ago |
arxiv.org
Kernel-Elastic Autoencoder for Molecular Design
3 weeks, 2 days ago |
arxiv.org
CFAT: Unleashing TriangularWindows for Image Super-resolution
3 weeks, 2 days ago |
arxiv.org
Stronger Graph Transformer with Regularized Attention Scores
4 weeks, 1 day ago |
arxiv.org
Looped Transformers are Better at Learning Learning Algorithms
4 weeks, 2 days ago |
arxiv.org
Magnushammer: A Transformer-Based Approach to Premise Selection
4 weeks, 2 days ago |
arxiv.org
Merging Text Transformer Models from Different Initializations
1 month, 1 week ago |
arxiv.org
Topic trend (last 90 days)
Top (last 7 days)
Transformers, Contextualism, and Polysemy
2 days, 20 hours ago |
arxiv.org
TransformerFAM: Feedback attention is working memory
2 days, 20 hours ago |
arxiv.org
Jobs in AI, ML, Big Data
Data Scientist (m/f/x/d)
@ Symanto Research GmbH & Co. KG | Spain, Germany
Enterprise Data Quality, Senior Analyst
@ Toyota North America | Plano
Data Analyst & Audit Management Software (AMS) Coordinator
@ World Vision | Philippines - Home Working
Product Manager Power BI Platform Tech I&E Operational Insights
@ ING | HBP (Amsterdam - Haarlerbergpark)
Sr. Director, Software Engineering, Clinical Data Strategy
@ Moderna | USA-Washington-Seattle-1099 Stewart Street
Data Engineer (Data as a Service)
@ Xplor | Atlanta, GA, United States