all AI news
$\textit{LinkPrompt}$: Natural and Universal Adversarial Attacks on Prompt-based Language Models
March 26, 2024, 4:51 a.m. | Yue Xu, Wenjie Wang
cs.CL updates on arXiv.org arxiv.org
Abstract: Prompt-based learning is a new language model training paradigm that adapts the Pre-trained Language Models (PLMs) to downstream tasks, which revitalizes the performance benchmarks across various natural language processing (NLP) tasks. Instead of using a fixed prompt template to fine-tune the model, some research demonstrates the effectiveness of searching for the prompt via optimization. Such prompt optimization process of prompt-based learning on PLMs also gives insight into generating adversarial prompts to mislead the model, raising …
abstract adversarial adversarial attacks arxiv attacks benchmarks cs.ai cs.cl language language model language models language model training language processing natural natural language natural language processing nlp paradigm performance processing prompt prompt-based learning research tasks template training type universal
More from arxiv.org / cs.CL updates on arXiv.org
Benchmarking LLMs via Uncertainty Quantification
1 day, 23 hours ago |
arxiv.org
CARE: Extracting Experimental Findings From Clinical Literature
1 day, 23 hours ago |
arxiv.org
Jobs in AI, ML, Big Data
Data Architect
@ University of Texas at Austin | Austin, TX
Data ETL Engineer
@ University of Texas at Austin | Austin, TX
Lead GNSS Data Scientist
@ Lurra Systems | Melbourne
Senior Machine Learning Engineer (MLOps)
@ Promaton | Remote, Europe
AIML - Sr Machine Learning Engineer, Data and ML Innovation
@ Apple | Seattle, WA, United States
Senior Data Engineer
@ Palta | Palta Cyprus, Palta Warsaw, Palta remote