Jan. 17, 2022, 2:10 a.m. | Junyi Li, Tianyi Tang, Wayne Xin Zhao, Jian-Yun Nie, Ji-Rong Wen

cs.CL updates on arXiv.org arxiv.org

Text Generation aims to produce plausible and readable text in human language
from input data. The resurgence of deep learning has greatly advanced this
field by neural generation models, especially the paradigm of pretrained
language models (PLMs). Grounding text generation on PLMs is seen as a
promising direction in both academia and industry. In this survey, we present
the recent advances achieved in the topic of PLMs for text generation. In
detail, we begin with introducing three key points of …

arxiv language language models survey text text generation

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne

Senior Machine Learning Engineer (MLOps)

@ Promaton | Remote, Europe

Software Engineer, Data Tools - Full Stack

@ DoorDash | Pune, India

Senior Data Analyst

@ Artsy | New York City