all AI news
Exploring Extreme Parameter Compression for Pre-trained Language Models. (arXiv:2205.10036v1 [cs.CL])
May 23, 2022, 1:12 a.m. | Yuxin Ren, Benyou Wang, Lifeng Shang, Xin Jiang, Qun Liu
cs.CL updates on arXiv.org arxiv.org
Recent work explored the potential of large-scale Transformer-based
pre-trained models, especially Pre-trained Language Models (PLMs) in natural
language processing. This raises many concerns from various perspectives, e.g.,
financial costs and carbon emissions. Compressing PLMs like BERT with
negligible performance loss for faster inference and cheaper deployment has
attracted much attention. In this work, we aim to explore larger compression
ratios for PLMs, among which tensor decomposition is a potential but
under-investigated one. Two decomposition and reconstruction protocols are
further proposed …
More from arxiv.org / cs.CL updates on arXiv.org
Jobs in AI, ML, Big Data
Senior Machine Learning Engineer (MLOps)
@ Promaton | Remote, Europe
Data Analyst (CPS-GfK)
@ GfK | Bucharest
Consultant Data Analytics IT Digital Impulse - H/F
@ Talan | Paris, France
Data Analyst
@ Experian | Mumbai, India
Data Scientist
@ Novo Nordisk | Princeton, NJ, US
Data Architect IV
@ Millennium Corporation | United States