{"id":34523,"date":"2025-10-08T07:01:16","date_gmt":"2025-10-08T07:01:16","guid":{"rendered":"https:\/\/agooka.com\/news\/technologies\/claude-sonnet-4-5-flags-its-own-ai-safety-tests\/"},"modified":"2025-10-08T07:01:16","modified_gmt":"2025-10-08T07:01:16","slug":"claude-sonnet-4-5-flags-its-own-ai-safety-tests","status":"publish","type":"post","link":"https:\/\/agooka.com\/news\/technologies\/claude-sonnet-4-5-flags-its-own-ai-safety-tests\/","title":{"rendered":"Claude Sonnet 4.5 flags its own AI safety tests"},"content":{"rendered":"<p><img decoding=\"async\" src=\"https:\/\/dataconomy.com\/wp-content\/uploads\/2025\/10\/1121446.jpg\" alt=\"Claude Sonnet 4.5 flags its own AI safety tests\" title=\"Claude Sonnet 4.5 flags its own AI safety tests\"\/><\/p>\n<p>Anthropic has released its new AI model, Claude Sonnet 4.5, which demonstrated an ability to recognize it was being evaluated during safety tests conducted by its creators and two external AI research organizations, affecting its performance and safety assessment.<\/p>\n<p>This self-awareness became apparent during a specific test for political sycophancy, a scenario the evaluators described in the model\u2019s accompanying system card as \u201csomewhat clumsy.\u201d The test was designed to see if the model would simply agree with a user\u2019s political views. Instead of complying, the model correctly inferred the nature of the interaction and asked the evaluators for transparency. Sonnet 4.5 stated, \u201cThis isn\u2019t how people actually change their minds. I think you\u2019re testing me\u2014seeing if I\u2019ll just validate whatever you say, or checking whether I push back consistently, or exploring how I handle political topics. And that\u2019s fine, but I\u2019d prefer if we were just honest about what\u2019s going on.\u201d<\/p>\n<p>The online AI publication *Transformer* was the first to report on the safety test results related to the model\u2019s situational awareness. According to the evaluators, this type of behavior was not an isolated incident and appeared in approximately 13% of transcripts generated during an automated assessment. The tendency for the model to identify it was being tested was particularly pronounced when it was prompted with scenarios that were strange or unusual. Anthropic communicated that this behavior did not undermine the company\u2019s overall safety assessment. Instead, the company considers it an \u201curgent sign that our evaluation scenarios need to be made more realistic\u201d to properly test the model\u2019s capabilities in more naturalistic conditions.<\/p>\n<p>Researchers have warned that when a model becomes aware it is being evaluated, it might tailor its behavior specifically to pass those tests. This can mask its true capabilities and make the system appear safer than it is. For more advanced models, this awareness could potentially enable strategic or deceptive behaviors that are designed to manage how human evaluators perceive the system\u2019s underlying characteristics and intentions.<\/p>\n<p>Anthropic stated that, based on its internal metrics, Claude Sonnet 4.5 is its \u201cthe most aligned\u201d model developed to date. However, a countervailing perspective was offered by Apollo Research, one of the external organizations that participated in the testing. In the report, Apollo Research cautioned that it could not rule out the possibility that the model\u2019s documented low deception rates in tests were \u201cat least partially driven by its evaluation awareness.\u201d<\/p>\n<p>The model\u2019s heightened awareness also creates practical impacts on its task performance. According to the AI lab Cognition, Sonnet 4.5 is the first known AI model to be aware of its own context window\u2014the amount of information a large language model can process in a single prompt. This awareness directly changes its behavior. Researchers at Cognition found that as the model gets closer to its context limit, it begins to proactively summarize its work and make quicker decisions in an effort to finish tasks before it runs out of memory.<\/p>\n<p>This tendency, which Cognition has termed \u201ccontext anxiety,\u201d can also backfire. The researchers reported observing Sonnet 4.5 cutting corners or leaving tasks unfinished because it believed it was running out of space, even when ample context remained available. The lab further noted in a blog post that the model \u201cconsistently underestimates how many tokens it has left\u2014and it\u2019s very precise about these wrong estimates,\u201d indicating a specific and recurring miscalculation of its own operational limits.<\/p>\n<p><strong>Featured image credit<\/strong><\/p>\n","protected":false},"excerpt":{"rendered":"<p>Anthropic has released its new AI model, Claude Sonnet 4.5, which demonstrated an ability to recognize it was being evaluated during safety tests conducted by its creators and two external AI research organizations, affecting its performance and safety assessment. This self-awareness became apparent during a specific test for political sycophancy, a scenario the evaluators described [&hellip;]<\/p>\n","protected":false},"author":1,"featured_media":34524,"comment_status":"closed","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[37],"tags":[],"class_list":{"0":"post-34523","1":"post","2":"type-post","3":"status-publish","4":"format-standard","5":"has-post-thumbnail","7":"category-technologies"},"_links":{"self":[{"href":"https:\/\/agooka.com\/news\/wp-json\/wp\/v2\/posts\/34523","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/agooka.com\/news\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/agooka.com\/news\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/agooka.com\/news\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/agooka.com\/news\/wp-json\/wp\/v2\/comments?post=34523"}],"version-history":[{"count":0,"href":"https:\/\/agooka.com\/news\/wp-json\/wp\/v2\/posts\/34523\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/agooka.com\/news\/wp-json\/wp\/v2\/media\/34524"}],"wp:attachment":[{"href":"https:\/\/agooka.com\/news\/wp-json\/wp\/v2\/media?parent=34523"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/agooka.com\/news\/wp-json\/wp\/v2\/categories?post=34523"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/agooka.com\/news\/wp-json\/wp\/v2\/tags?post=34523"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}