all AI news
Beyond Prompting: Making Pre-trained Language Models Better Zero-shot Learners by Clustering Representations. (arXiv:2210.16637v2 [cs.CL] UPDATED)
Nov. 24, 2022, 7:18 a.m. | Yu Fei, Ping Nie, Zhao Meng, Roger Wattenhofer, Mrinmaya Sachan
cs.CL updates on arXiv.org arxiv.org
Recent work has demonstrated that pre-trained language models (PLMs) are
zero-shot learners. However, most existing zero-shot methods involve heavy
human engineering or complicated self-training pipelines, hindering their
application to new situations. In this work, we show that zero-shot text
classification can be improved simply by clustering texts in the embedding
spaces of PLMs. Specifically, we fit the unlabeled texts with a Bayesian
Gaussian Mixture Model after initializing cluster positions and shapes using
class names. Despite its simplicity, this approach achieves …
More from arxiv.org / cs.CL updates on arXiv.org
Jobs in AI, ML, Big Data
Lead GNSS Data Scientist
@ Lurra Systems | Melbourne
Senior Machine Learning Engineer (MLOps)
@ Promaton | Remote, Europe
[Job - 14823] Senior Data Scientist (Data Analyst Sr)
@ CI&T | Brazil
Data Engineer
@ WorldQuant | Hanoi
ML Engineer / Toronto
@ Intersog | Toronto, Ontario, Canada
Analista de Business Intelligence (Industry Insights)
@ NielsenIQ | Cotia, Brazil