Retour aux articles
IAOpenAI News

Our updated Preparedness Framework

Sharing our updated framework for measuring and protecting against severe harm from frontier AI capabilities.

Le flux RSS ne fournissait qu'un extrait. FlowMarket a récupéré le contenu public disponible depuis la page originale, sans contourner les contenus réservés.

April 15, 2025

Our updated Preparedness Framework

Sharing our updated framework for measuring and protecting against severe harm from frontier AI capabilities.

OpenAI Preparedness Framework illustration

We’re releasing an update to our Preparedness Framework, our process for tracking and preparing for advanced AI capabilities that could introduce new risks of severe harm. As our models continue to get more capable ⁠ , safety will increasingly depend on having the right real-world safeguards in place.

This update introduces a sharper focus on the specific risks that matter most, stronger requirements for what it means to “sufficiently minimize” those risks in practice, and clearer operational guidance on how we evaluate, govern, and disclose our safeguards. Additionally, we introduce future-facing research categories that allow us to remain at the forefront of understanding emerging capabilities to keep pace with where the technology is headed. We will continue investing deeply in this process by making our preparedness work more actionable, rigorous, and transparent as the technology advances.

We’ve learned a great deal from our own testing, insights from external experts, and lessons from the field. This update reflects that progress. In line with our core safety principles ⁠ , it makes targeted improvements that include:

  • Clear criteria for prioritizing high-risk capabilities. We use a structured risk assessment process to evaluate whether a frontier capability could lead to severe harm and we assign it to a category based on defined criteria. We track capabilities that meet five key criteria that make it a priority for us to prepare in advance: the risk should be plausible, measurable, severe, net new, and instantaneous or irremediable. We measure progress on these capabilities, and build safeguards against the risks that these capabilities create.
  • Sharper capability categories. We've updated our categorization of capabilities to apply these criteria and reflect our current understanding. Tracked Categories: These are established areas where we have mature evaluations and ongoing safeguards. They are Biological and Chemical capabilities, Cybersecurity capabilities, and AI Self-improvement capabilities. We continue to believe some of the most transformative benefits from AI will come from its use in science, engineering, and research - including from capabilities in our Tracked Categories. Investing early in both measurement and safeguards for these dual-use categories will enable us to safely unlock the benefits we anticipate from their use.
  • Research Categories: We’re introducing a set of Research Categories of capability, which are areas that could pose risks of severe harm, that do not yet meet our criteria to be Tracked Categories. We’re working to develop threat models and advanced capability evaluations for these. Current focus areas include Long-range Autonomy, Sandbagging (intentionally underperforming), Autonomous Replication and Adaptation, Undermining Safeguards, and Nuclear and Radiological.
  • Persuasion risks will be handled outside the Preparedness Framework, including via our Model Spec, restricting the use of our tools for political campaigning or lobbying, and our ongoing investigations into misuse of our products (including detecting and disrupting influence operations ⁠ (opens in a new window) ).

We’ll continue to publish our Preparedness findings with each frontier model release, just as we’ve done for GPT‑4o ⁠ , OpenAI o1 ⁠ , Operator ⁠ , o3‑mini ⁠ , deep research ⁠ , and GPT‑4.5 ⁠ , and share new benchmarks to support broader safety efforts across the field.

We’re deeply grateful to internal teams, external researchers and industry peers who’ve contributed invaluable insights to this latest update. The Preparedness Framework remains a living document, and we expect to continue updating it as we learn more.

  • 2025
  • Ethics & Safety
  • Framework

Author

Besoin d'un workflow n8n ou d'aide pour l'installer ?

Après la veille, passez à l'action : trouvez un template n8n ou un créateur capable de l'adapter à vos outils.

Source

OpenAI News - openai.com

Voir la publication originale