all AI news
[R] Distort, Distract, Decode: Instruction-Tuned Model Can Refine its Response from Noisy Instructions
Nov. 27, 2023, 9:58 p.m. | /u/Queasy_Ad_6423
Machine Learning www.reddit.com
https://preview.redd.it/4dvuvo18113c1.png?width=1916&format=png&auto=webp&s=4b9ba695cb00719c3e80c931f8b52a1008d7db5b
Link: [https://openreview.net/forum?id=IqJ3CU3flr](https://openreview.net/forum?id=IqJ3CU3flr)Abstract:
>While instruction-tuned language models have demonstrated impressive zero-shot generalization, these models often struggle to generate accurate responses when faced with instructions that fall outside their training set. This paper presents Instructive Decoding (ID), a simple yet effective approach that augments the efficacy of instruction-tuned models. Specifically, ID adjusts the logits for nexttoken prediction in a contrastive manner, utilizing predictions generated from a manipulated version of the original instruction, referred to as a noisy instruction. This noisy …
decode decoding generate language language models machinelearning paper refine responses set simple struggle training
More from www.reddit.com / Machine Learning
Jobs in AI, ML, Big Data
Data Architect
@ University of Texas at Austin | Austin, TX
Data ETL Engineer
@ University of Texas at Austin | Austin, TX
Lead GNSS Data Scientist
@ Lurra Systems | Melbourne
Senior Machine Learning Engineer (MLOps)
@ Promaton | Remote, Europe
Software Engineer, Machine Learning (Tel Aviv)
@ Meta | Tel Aviv, Israel
Senior Data Scientist- Digital Government
@ Oracle | CASABLANCA, Morocco