Skip to main content

Envisioning is an emerging technology research institute and advisory.

LinkedInInstagramGitHub

2011 — 2026

research
  • Reports
  • Newsletter
  • Methodology
  • Origins
  • Vocab
services
  • Research Sessions
  • Signals Workspace
  • Bespoke Projects
  • Use Cases
  • Signal Scanfree
  • Readinessfree
impact
  • ANBIMAFuture of Brazilian Capital Markets
  • IEEECharting the Energy Transition
  • Horizon 2045Future of Human and Planetary Security
  • WKOTechnology Scanning for Austria
audiences
  • Innovation
  • Strategy
  • Consultants
  • Foresight
  • Associations
  • Governments
resources
  • Pricing
  • Partners
  • How We Work
  • Data Visualization
  • Multi-Model Method
  • FAQ
  • Security & Privacy
about
  • Manifesto
  • Community
  • Events
  • Support
  • Contact
  • Login
ResearchServicesPricingPartnersAbout
ResearchServicesPricingPartnersAbout
  1. Home
  2. Vocab
  3. RSP (Responsible Scaling Policy)

RSP (Responsible Scaling Policy)

A formal framework committing AI labs to safety evaluations before scaling models further.

Year: 2023Generality: 293
Back to Vocab

A Responsible Scaling Policy (RSP) is a formal organizational commitment by an AI developer to evaluate the potential dangers of increasingly capable AI systems before proceeding with further scaling. Rather than treating safety as an afterthought, an RSP establishes concrete thresholds — often called "safety levels" or "capability thresholds" — at which a model must be assessed for dangerous capabilities before additional compute or data is invested. Anthropic introduced the first widely publicized RSP in 2023, and the framework has since influenced how frontier AI labs think about the relationship between model capability and deployment readiness.

In practice, an RSP defines a tiered system of risk levels tied to specific model capabilities. As a model approaches a threshold — such as the ability to meaningfully assist in creating biological weapons, or to autonomously conduct long-horizon cyberattacks — the policy mandates a structured evaluation process. If the model crosses that threshold without adequate safety mitigations in place, the organization commits to halting further scaling until those mitigations are developed and validated. This creates a feedback loop between capability research and safety research, ensuring the two advance in tandem rather than independently.

The policy framework matters because the history of AI development has often seen safety and alignment work lag behind raw capability improvements. RSPs attempt to institutionalize a precautionary principle: the burden of proof shifts from "prove it's dangerous" to "prove it's safe enough to scale." This is a meaningful philosophical shift, particularly for organizations operating at the frontier where the consequences of miscalibration are potentially severe and irreversible.

Critics note that RSPs are self-imposed and largely unverifiable by outside parties, raising questions about accountability and enforcement. Nonetheless, they represent an important step toward operationalizing AI safety commitments in a concrete, auditable form. As governments and standards bodies develop external oversight mechanisms, RSPs may serve as a template for more formal regulatory requirements around frontier model development.

Related

Related

ASL (AI Safety Level)
ASL (AI Safety Level)

A tiered framework for classifying AI risk levels to guide responsible development.

Generality: 322
RSI (Recursive Self-Improvement)
RSI (Recursive Self-Improvement)

AI systems autonomously improving their own capabilities through research and optimization loops

Generality: 525
Responsible AI
Responsible AI

Developing and deploying AI systems that are ethical, fair, transparent, and accountable.

Generality: 834
AI Safety
AI Safety

Research field ensuring AI systems remain beneficial, aligned, and free from catastrophic risk.

Generality: 871
Safety Net
Safety Net

Layered safeguards that prevent, detect, and mitigate harmful AI system outcomes.

Generality: 521
Guardrails
Guardrails

Technical and policy constraints ensuring AI systems behave safely and ethically.

Generality: 694