1
Feature Story
Docs show how freelancers test AI by creating 'harmful' prompts
Apr 04, 2025 · businessinsider.comThe training materials show how contributors are taught to embed harmful requests in fictional setups to test AI safety filters. Examples include prompts about bomb-making and domestic violence. Scale AI, a leading data labeling service provider, works with major tech companies like OpenAI, Meta, and Google. Despite legal challenges and investigations, Scale AI is seeking a valuation of up to $25 billion. The company defends its business model and claims to offer flexible earning opportunities, while critics allege inadequate mental health support and worker misclassification.
Key takeaways
- Freelancers are paid to stress-test AI with harmful prompts to make AI safer, as revealed by leaked training documents.
- Scale AI and Outlier use freelancers to create prompts about sensitive topics like suicide, abuse, and terrorism.
- Contributors are encouraged to be creative with harmful prompts but must avoid content related to child sexual abuse material.
- Scale AI faces lawsuits and a Department of Labor investigation over worker treatment and misclassification.