all AI news
(De)ToxiGen: Leveraging large language models to build more robust hate speech detection tools
Microsoft Research www.microsoft.com
It’s a well-known challenge that large language models (LLMs)—growing in popularity thanks to their adaptability across a variety of applications—carry risks. Because they’re trained on large amounts of data from across the internet, they’re capable of generating inappropriate and harmful language based on similar language encountered during training. Content moderation tools can be deployed to […]
The post (De)ToxiGen: Leveraging large language models to build more robust hate speech detection tools appeared first on Microsoft Research.
detection hate speech language language models large language models research blog speech tools