San Francisco, CA – Anthropic, a prominent artificial intelligence safety and research company, has activated its highest AI Safety Level 3 (ASL-3) protections for its newly launched Claude Opus 4 model. This unprecedented move comes as internal testing indicated the model could "substantially increase" the ability of individuals with basic STEM knowledge to develop chemical, biological, radiological, and nuclear (CBRN) weapons. The decision underscores the company's cautious approach to AI development, a stance acknowledged by social media user "near" who observed,
"the biorisk classifier in the claude app is pretty bad but no one ever got fired from anthropic for being excessively concerned about biorisk so hey."
The ASL-3 standard, detailed in Anthropic's Responsible Scaling Policy (RSP), mandates enhanced security and deployment measures specifically targeting CBRN misuse. These include "constitutional classifiers," AI systems designed to detect and block harmful CBRN-related queries and outputs, particularly those forming extended workflows. Anthropic clarified that while it hasn't definitively confirmed Opus 4 requires ASL-3, the precautionary activation was necessary due to the model's improved capabilities in handling CBRN-related knowledge, making it difficult to rule out potential risks.
Despite Anthropic's stated commitment to safety, the increased safety measures have drawn mixed reactions from the user community. Many users have expressed frustration over a perceived rise in "refusal rates" for seemingly legitimate queries, impacting the model's general usability. Critics on platforms like Reddit argue that this intense focus on safety might hinder Claude's competitiveness against rivals like OpenAI and DeepSeek, which are seen as prioritizing broader capabilities. However, industry analysts suggest that Anthropic's emphasis on robust safety protocols appeals to enterprise clients, who prioritize legal and reputational risk mitigation.
The activation of ASL-3 also reignites a broader debate within the AI safety community regarding the effectiveness and transparency of biorisk evaluations. Some experts question whether current evaluation methodologies truly capture real-world bioweapon development risks, noting that results are often opaque. Anthropic maintains that its multi-layered "defense in depth" strategy, including bug bounty programs and continuous monitoring, is crucial for mitigating catastrophic misuse, even as the company acknowledges the ongoing challenge of refining these defenses.