The new 2025 OWASP Top 10 Risks for Large Language Models (LLMs) highlights critical shifts within AI security. Here's a summary of new, expanded or updated risks that are particularly interesting.
Fergal Glynn
OpenAI had a staggering 300 million weekly active users at the end of 2024. It’s therefore no surprise that the world’s best-known large language model (LLM) chatbot is one of the most well-funded startups of its kind, securing more than $17 billion in funding as of 2024.
Artificial intelligence (AI) and machine learning (ML) tools like OpenAI’s ChatGPT are gaining in popularity, but while OpenAI’s tools promise to revolutionize many industries, they aren’t without risks. In fact, malicious actors frequently try to undermine the AI model with nefarious prompts, edge cases, and more.
That’s why the company invested in the OpenAI Red Teaming Network. This groundbreaking initiative enlists a large group of outside experts to rigorously test and critique OpenAI’s models. OpenAI’s approach has set the standard for responsible AI development.
Learn what the OpenAI Red Teaming Network is, how it works, and why it’s an essential part of building trust in AI systems. Whether you’re a technologist, business leader, or simply curious about the future of AI, understanding this process provides key insights into the intersection of innovation and accountability.
Many organizations use red teaming to test their systems’ defensive capabilities. Google and Microsoft also have their own red team operations, for instance. During a typical red team test, an internal team of ethical hackers uses the strategies of real-world hackers to spot and fix vulnerabilities.
The OpenAI Red Teaming Network takes the power of red teaming one step further. It’s a collaborative initiative that brings external experts from many fields together and asks them to test and critique OpenAI’s systems.
The Red Teaming Network usually includes researchers, social scientists, security specialists, ethicists, academics, and more—such as some of the thought leaders on this list.
Consulting so many experts might sound overwhelming, but this approach allows OpenAI to assess its models from many angles, including:
Since attackers use AI for a range of nefarious purposes, the OpenAI Red Teaming Network’s breadth allows the company to address the multifaceted risks of AI.
Left unchecked, AI and ML models can perpetuate bias, generate malicious responses, and breach sensitive data. OpenAI’s Red Teaming Network plays a critical role in keeping OpenAI’s systems safe and fair.
AI systems can have unintended consequences, such as generating harmful, biased, or misleading content. The biggest benefit of any red teaming exercise is the ability to spot vulnerabilities or gaps in the AI model that the team can address before deploying it.
By simulating adversarial scenarios and edge cases, the network helps identify and address risks in advance, reducing the likelihood of harm.
Unfortunately, AI models with poor safeguards have already led to real-world harm. OpenAI relies on its Red Teaming Network to anticipate accidental or intentional harm that its internal team might miss.
Red teaming ensures that AI systems are robust enough to handle a wide range of user inputs— including malicious or adversarial ones—while maintaining their integrity.
A range of sensitive industries rely on AI technology, including healthcare, government, education, and utilities. Some of these organizations license OpenAI’s model internally, making it necessary to test the model for ethical issues.
The OpenAI Red Teaming Network brings together diverse experts to ensure AI systems are fair, unbiased, and culturally sensitive, minimizing the risk of discrimination or other ethical issues.
AI red teaming is a fast-growing strategy for thoroughly testing AI and ML models before users access them. While internal AI red teaming has many benefits, the OpenAI Red Teaming Network offers a new model for red teaming that works surprisingly well for its LLM, supported by a growing ecosystem of tools and best practices for red teaming efforts.
However, not all organizations have the resources or time to work with a large group of external experts. That’s where Mindgard comes in. Our customized red teaming solutions keep your AI systems secure and reliable through automated AI red teaming. Schedule a no-obligation Mindgard demo today.
While you can apply to join the network, OpenAI often directly recruits experts from a wide range of disciplines, including cybersecurity, AI ethics, sociology, and public policy. Eligibility typically depends on expertise, experience in relevant fields, and a commitment to ethical standards.
Absolutely. OpenAI relies on the network’s findings to refine its AI tools. The Red Teaming Network often leads to changes in the system’s design, additional safety features, adjustments to training data, and updates to usage policies.
There’s no such thing as a completely risk-proof AI model. Organizations can (and should) do their best to mitigate harm, but red teaming can’t eliminate risks entirely.
AI systems are complex and operate in dynamic environments, which means new risks may emerge over time. Still, red teaming is critical for minimizing vulnerabilities and preparing AI to handle a wide range of scenarios.