News
In an earlier version of its responsible scaling policy, published in October, Anthropic said it would begin sweeping physical offices for hidden devices as part of a ramped-up security effort.
Anthropic announced updates to the "responsible scaling" policy for its AI, including defining which of its model safety levels are powerful enough to need additional security safeguards. In an ...
The policy, referred to as the Responsible Scaling Policy (RSP), is designed specifically to mitigate “catastrophic risks,” or situations where an AI model could directly cause large-scale ...
The company’s Responsible Scaling Policy (RSP) is a public commitment that was originally released in September 2023 and maintained that Anthropic would not “train or deploy models capable of ...
Anthropic, the artificial intelligence company behind the popular Claude chatbot, today announced a sweeping update to its Responsible Scaling Policy (RSP), aimed at mitigating the risks of highly ...
Evaluations will measure the AI Safety Levels defined in Anthropic’s Responsible Scaling Policy, ensuring models are developed and deployed responsibly.
The PDF document, detailing the company's "responsible scaling policy," outlines several procedural changes that it says are needed to monitor the ongoing risks of misuse of AI models.
The Responsible Scaling Policy: A Blueprint for Oversight Anthropic has proposed a Responsible Scaling Policy (RSP) as a foundational approach to AI regulation.
On March 31st, Anthropic announced implementing updates to its “responsible scaling” policy for AI, defining model safety levels to determine the need for additional security.
Holden Karnofsky is a Member of Technical Staff at Anthropic, where he focuses on the design of the company's Responsible Scaling Policy and other aspects of preparing for the possibility of highly ...
Results that may be inaccessible to you are currently showing.
Hide inaccessible results