A tiered framework for classifying AI risk levels to guide responsible development.
AI Safety Levels (ASLs) are a structured classification framework used to assess and categorize the potential risks posed by increasingly capable AI systems. Pioneered by Anthropic as part of its Responsible Scaling Policy, the framework defines discrete tiers—typically numbered ASL-1 through ASL-4 and beyond—each corresponding to a threshold of capability and an associated set of required safety and security measures. The underlying premise is that as AI systems grow more powerful, the potential for catastrophic misuse or unintended harm grows in tandem, and governance protocols must scale accordingly.
The framework operates by establishing concrete, measurable criteria for what capabilities would push a model from one safety level to the next. For example, a model that demonstrates meaningful ability to assist in the creation of biological, chemical, nuclear, or radiological weapons, or that exhibits early signs of autonomous self-replication, might trigger elevation to a higher ASL. Once a threshold is crossed, the organization is committed to implementing specific countermeasures—such as enhanced access controls, red-teaming requirements, or deployment restrictions—before proceeding with further development or release. This creates a binding feedback loop between capability evaluation and safety investment.
ASLs are closely related to, and often used interchangeably with, similar tiered frameworks from other organizations, such as OpenAI's Preparedness Framework and its associated risk classifications. Together, these approaches represent a broader industry movement toward "responsible scaling policies"—formal commitments that tie the pace of AI development to demonstrated safety progress. The goal is to prevent organizations from racing ahead of their own ability to understand and control the systems they build.
The practical importance of ASL frameworks lies in their attempt to operationalize AI safety as a concrete engineering and governance discipline rather than an abstract aspiration. By defining thresholds in advance and committing to specific responses, organizations create accountability structures that can be audited and compared across the industry. Critics note that self-imposed frameworks lack external enforcement, but proponents argue they represent a meaningful first step toward the kind of standardized, internationally recognized AI risk governance that more advanced systems will ultimately require.