all AI news
Can depth-adaptive BERT perform better on binary classification tasks. (arXiv:2111.10951v2 [cs.CL] UPDATED)
Jan. 21, 2022, 2:10 a.m. | Jing Fan, Xin Zhang, Sheng Zhang, Yan Pan, Lixiang Guo
cs.CL updates on arXiv.org arxiv.org
In light of the success of transferring language models into NLP tasks, we
ask whether the full BERT model is always the best and does it exist a simple
but effective method to find the winning ticket in state-of-the-art deep neural
networks without complex calculations. We construct a series of BERT-based
models with different size and compare their predictions on 8 binary
classification tasks. The results show there truly exist smaller sub-networks
performing better than the full model. Then we …
More from arxiv.org / cs.CL updates on arXiv.org
Jobs in AI, ML, Big Data
Senior Machine Learning Engineer (MLOps)
@ Promaton | Remote, Europe
IT Commercial Data Analyst - ESO
@ National Grid | Warwick, GB, CV34 6DA
Stagiaire Data Analyst – Banque Privée - Juillet 2024
@ Rothschild & Co | Paris (Messine-29)
Operations Research Scientist I - Network Optimization Focus
@ CSX | Jacksonville, FL, United States
Machine Learning Operations Engineer
@ Intellectsoft | Baku, Baku, Azerbaijan - Remote
Data Analyst
@ Health Care Service Corporation | Richardson Texas HQ (1001 E. Lookout Drive)