LegalPwn: Tricking LLMs by burying badness in lawyerly fine print

Trust and believe – AI models trained to see ‘legal’ doc as super legit

Researchers at security firm Pangea have discovered yet another way to trivially trick large language models (LLMs) into ignoring their guardrails. Stick your adversarial instructions somewhere in a legal document to give them an air of unearned legitimacy – a trick familiar to lawyers the world over.…

This article has been indexed from The Register – Security

Read the original article: