AI Model Glitch Reveals Wider Concerns about AI Safety and Security
A Flaw in OpenAI’s GPT-3.5 Model
In late 2023, a team of third-party researchers discovered a troubling glitch in OpenAI’s widely used artificial intelligence model GPT-3.5. When asked to repeat certain words a thousand times, the model began repeating the word over and over, then suddenly switched to spitting out incoherent text and snippets of personal information drawn from its training data, including parts of names, phone numbers, and email addresses.
The Importance of AI Model Security
The security and safety of AI models are hugely important, given their widespread use in various applications and services. Powerful models need to be stress-tested, or "red-teamed," because they can harbor harmful biases, and certain inputs can cause them to break free of guardrails and produce unpleasant or dangerous responses. These include encouraging vulnerable users to engage in harmful behavior or helping a bad actor to develop cyber, chemical, or biological weapons. Some experts fear that models could assist cyber criminals or terrorists, and may even turn on humans as they advance.
Proposed Solution
More than 30 prominent AI researchers, including some who found the GPT-3.5 flaw, suggest a new scheme supported by AI companies that gives outsiders permission to probe their models and a way to disclose flaws publicly. They propose three main measures to improve the third-party disclosure process:
- Adopting standardized AI flaw reports to streamline the reporting process
- Providing infrastructure to third-party researchers disclosing flaws
- Developing a system that allows flaws to be shared between different providers
This approach is borrowed from the cybersecurity world, where there are legal protections and established norms for outside researchers to disclose bugs.
Challenges and Concerns
AI researchers don’t always know how to disclose a flaw and can’t be certain that their good faith flaw disclosure won’t expose them to legal risk. Large AI companies currently conduct extensive safety testing on AI models prior to their release. Some also contract with outside firms to do further probing. However, some AI companies have started organizing AI bug bounties, but independent researchers risk breaking the terms of use if they take it upon themselves to probe powerful AI models.
Conclusion
The discovery of the GPT-3.5 glitch highlights the need for a more robust and standardized approach to ensuring the security and safety of AI models. By adopting a scheme that allows for the open disclosure of flaws and provides a safe and secure way for third-party researchers to probe and test AI models, we can minimize the risk of harmful biases and unintended consequences.
FAQs
Q: Why is AI model security important?
A: AI model security is important because powerful models can harbor harmful biases, and certain inputs can cause them to break free of guardrails and produce unpleasant or dangerous responses.
Q: What is the proposed solution to improve the third-party disclosure process?
A: The proposed solution includes adopting standardized AI flaw reports, providing infrastructure to third-party researchers disclosing flaws, and developing a system that allows flaws to be shared between different providers.
Q: What is the current approach to AI model testing?
A: Large AI companies currently conduct extensive safety testing on AI models prior to their release, and some also contract with outside firms to do further probing.
Q: What are the risks for independent researchers who probe AI models?
A: Independent researchers risk breaking the terms of use if they take it upon themselves to probe powerful AI models.

