all AI news
Learning to Watermark LLM-generated Text via Reinforcement Learning
March 19, 2024, 4:41 a.m. | Xiaojun Xu, Yuanshun Yao, Yang Liu
cs.LG updates on arXiv.org arxiv.org
Abstract: We study how to watermark LLM outputs, i.e. embedding algorithmically detectable signals into LLM-generated text to track misuse. Unlike the current mainstream methods that work with a fixed LLM, we expand the watermark design space by including the LLM tuning stage in the watermark pipeline. While prior works focus on token-level watermark that embeds signals into the output, we design a model-level watermark that embeds signals into the LLM weights, and such signals can be …
arxiv cs.ai cs.cr cs.lg generated llm reinforcement reinforcement learning text type via watermark
More from arxiv.org / cs.LG updates on arXiv.org
Jobs in AI, ML, Big Data
Senior Machine Learning Engineer
@ GPTZero | Toronto, Canada
ML/AI Engineer / NLP Expert - Custom LLM Development (x/f/m)
@ HelloBetter | Remote
Doctoral Researcher (m/f/div) in Automated Processing of Bioimages
@ Leibniz Institute for Natural Product Research and Infection Biology (Leibniz-HKI) | Jena
Coding Data Quality Auditor
@ Neuberger Berman | Work At Home-Georgia
Post Graduate (Year-Round) Intern - Market Research Analyst and Agreement Support
@ National Renewable Energy Laboratory | CO - Golden
Retail Analytics Engineering - Sr. Manager (Data)
@ Axalta | Woonsocket-1 CVS Drive