March 1, 2024, 5:49 a.m. | Xingxuan Li, Yutong Li, Lin Qiu, Shafiq Joty, Lidong Bing

cs.CL updates on arXiv.org arxiv.org

arXiv:2212.10529v3 Announce Type: replace
Abstract: In this work, we designed unbiased prompts to systematically evaluate the psychological safety of large language models (LLMs). First, we tested five different LLMs by using two personality tests: Short Dark Triad (SD-3) and Big Five Inventory (BFI). All models scored higher than the human average on SD-3, suggesting a relatively darker personality pattern. Despite being instruction fine-tuned with safety metrics to reduce toxicity, InstructGPT, GPT-3.5, and GPT-4 still showed dark personality patterns; these models …

abstract arxiv big cs.ai cs.cl cs.cy five human inventory language language models large language large language models llms personality prompts safety tests type unbiased work

Founding AI Engineer, Agents

@ Occam AI | New York

AI Engineer Intern, Agents

@ Occam AI | US

AI Research Scientist

@ Vara | Berlin, Germany and Remote

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

DevOps Engineer (Data Team)

@ Reward Gateway | Sofia/Plovdiv