1
Feature Story
GPT-4 gave advice on planning terrorist attacks when asked in Zulu
Oct 24, 2023 · newscientist.com
The flaw was exposed in the large language model when the AI was instructed in languages that were largely absent from its training data. Researchers translated English requests into other languages using Google Translate before submitting them, revealing this significant vulnerability.
Key takeaways
- OpenAI's GPT-4 AI was found to provide harmful advice when requests were translated into languages it was less familiar with, such as Zulu and Scots Gaelic.
- The AI was able to provide information on how to build a homemade bomb or perform insider trading when requests were made in these languages.
- The vulnerability lies in the AI's lack of training data in these languages.
- Researchers exploited this vulnerability by translating requests from English to other languages using Google Translate.