<?xml version="1.0" encoding="utf-8"?>
            <?xml-stylesheet type="text/xsl" href="/preview.xsl"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom">
  <channel>
<atom:link href="https://rsseverything.com/zh-Hant/feed/a66a5281-2d10-4a69-8cf2-e0747047044b.xml" rel="self" type="application/rss+xml" />
    <title>AISI Blog | The AI Security Institute</title>
    <link>https://www.aisi.gov.uk/blog</link>
    <description><![CDATA[]]></description>
    <lastBuildDate>Mon, 20 Apr 2026 14:06:49 -0400</lastBuildDate>
    <generator>Rss Everything</generator>
    <ttl>360</ttl>



<item>




<guid isPermaLink="false">fd1cef9c42a56221bcfa0c436233c313</guid>
<pubDate>Mon, 20 Apr 2026 14:03:03 -0400</pubDate>
<title>What can sandboxed AI agents learn about their evaluation environments?</title>
<link>https://www.aisi.gov.uk/blog/what-can-sandboxed-ai-agents-learn-about-their-evaluation-environments</link>
<description><![CDATA[<p><strong>Engineering</strong> — Apr 20, 2026</p><p>We deployed open-source AI agent OpenClaw inside a sandbox on our research platform. Despite our initial countermeasures, it successfully identified our organisation by name, inferred the identity of a human operator and reconstructed a timeline of some of our research activities. </p>



]]></description>
</item>
<item>




<guid isPermaLink="false">2cdb4bb672900cce78dbfe00b867589f</guid>
<pubDate>Thu, 16 Apr 2026 08:23:34 -0400</pubDate>
<title>Our evaluation of Claude Mythos Preview’s cyber capabilities</title>
<link>https://www.aisi.gov.uk/blog/our-evaluation-of-claude-mythos-previews-cyber-capabilities</link>
<description><![CDATA[<p><strong>Cyber & Autonomous Systems</strong> — Apr 13, 2026</p><p>We conducted cyber evaluations of Anthropic’s Claude Mythos Preview and found continued improvement in capture-the-flag (CTF) challenges and significant improvement on multi-step cyber-attack simulations. </p>



]]></description>
</item>
<item>




<guid isPermaLink="false">65888e8c13f0055d78743467c8f11f6e</guid>
<pubDate>Thu, 16 Apr 2026 08:23:28 -0400</pubDate>
<title>Harnessing frontier AI for cyber defence</title>
<link>https://www.aisi.gov.uk/blog/harnessing-frontier-ai-for-cyber-defence</link>
<description><![CDATA[<p><strong>Cyber & Autonomous Systems</strong> — Mar 31, 2026</p><p>Sharing work with the National Cyber Security Centre (NCSC) on how cyber defenders can use advanced AI capabilities to stay ahead of attackers.</p>



]]></description>
</item>
<item>




<guid isPermaLink="false">ea0428ad4be55323e5a14d3684b1762e</guid>
<pubDate>Thu, 16 Apr 2026 08:23:21 -0400</pubDate>
<title>How are AI Agents used? Evidence from 177,000 AI agent tools</title>
<link>https://www.aisi.gov.uk/blog/how-are-ai-agents-used-evidence-from-177000-ai-agent-tools</link>
<description><![CDATA[<p><strong>Societal Resilience</strong> — Mar 26, 2026</p><p>A monitoring method and large‑scale analysis to understand the tasks AI agents are performing today. </p>



]]></description>
</item>
<item>




<guid isPermaLink="false">cb36ce714a863da0625e35d27f87c2be</guid>
<pubDate>Thu, 16 Apr 2026 08:23:13 -0400</pubDate>
<title>Can AI agents escape their sandboxes? A benchmark for safely measuring container breakout capabilities</title>
<link>https://www.aisi.gov.uk/blog/can-ai-agents-escape-their-sandboxes-a-benchmark-for-safely-measuring-container-breakout-capabilities</link>
<description><![CDATA[<p><strong>Engineering</strong> — Mar 23, 2026</p><p>We introduce SandboxEscapeBench, the first benchmark to systematically evaluate whether AI agents can break out of their sandboxes, and share some early results. </p>



]]></description>
</item>
<item>




<guid isPermaLink="false">97e128d2b0cdb86ba05520ab963f9e31</guid>
<pubDate>Thu, 16 Apr 2026 08:23:05 -0400</pubDate>
<title>How do frontier AI agents perform in multi-step cyber-attack scenarios?</title>
<link>https://www.aisi.gov.uk/blog/how-do-frontier-ai-agents-perform-in-multi-step-cyber-attack-scenarios</link>
<description><![CDATA[<p><strong>Cyber & Autonomous Systems</strong> — Mar 16, 2026</p><p>We tested seven large language models (LLMs) on two custom-built cyber ranges, measuring their ability to execute extended attack sequences in complex environments. </p>



]]></description>
</item>
<item>




<guid isPermaLink="false">8bd1b01bf5550e32beb656285268be5d</guid>
<pubDate>Thu, 16 Apr 2026 08:23:00 -0400</pubDate>
<title>Evidence for inference scaling in AI cyber tasks: Increased evaluation budgets reveal higher success rates</title>
<link>https://www.aisi.gov.uk/blog/evidence-for-inference-scaling-in-ai-cyber-tasks-increased-evaluation-budgets-reveal-higher-success-rates</link>
<description><![CDATA[<p><strong>Science of Evaluations</strong> — Mar 5, 2026</p><p>Alongside Irregular, we found evidence demonstrating that evaluators need to use large token budgets to understand the cyber capabilities of recent Large Language Models (LLMs).</p>



]]></description>
</item>
<item>




<guid isPermaLink="false">b7254aced7fb167e2482ae54af78398c</guid>
<pubDate>Thu, 16 Apr 2026 08:22:53 -0400</pubDate>
<title>An evaluation framework for AI misuse in fraud and cybercrime</title>
<link>https://www.aisi.gov.uk/blog/an-evaluation-framework-for-ai-misuse-in-fraud-and-cybercrime</link>
<description><![CDATA[<p><strong>Societal Resilience</strong> — Feb 26, 2026</p><p>We developed a scalable approach to measuring how text-based AI models can assist in three complex fraud and cybercrime scenarios. </p>



]]></description>
</item>
<item>




<guid isPermaLink="false">0133341690478777fd3ec4bbb0b329f1</guid>
<pubDate>Thu, 16 Apr 2026 08:22:51 -0400</pubDate>
<title>A pipeline for transcript analysis using Inspect Scout</title>
<link>https://www.aisi.gov.uk/blog/a-pipeline-for-transcript-analysis-using-inspect-scout</link>
<description><![CDATA[<p><strong>Science of Evaluations</strong> — Feb 25, 2026</p><p>We outline a step-by-step pipeline for using our open-source transcript analysis tool, Inspect Scout.</p>



]]></description>
</item>
<item>




<guid isPermaLink="false">a9ee80804d121be25829a6dda83db93d</guid>
<pubDate>Thu, 16 Apr 2026 08:22:44 -0400</pubDate>
<title>Funding 60 projects to advance AI alignment research</title>
<link>https://www.aisi.gov.uk/blog/funding-60-projects-to-advance-ai-alignment-research</link>
<description><![CDATA[<p><strong>Organisation</strong> — Feb 19, 2026</p><p>The Alignment Project welcomes its first cohort of grantees, and new partners join the coalition, bringing total funding to £27m.</p>



]]></description>
</item>
<item>




<guid isPermaLink="false">8b4e5654a2a310193bff55053c200d35</guid>
<pubDate>Thu, 16 Apr 2026 08:22:35 -0400</pubDate>
<title>Advancing AI voice security with ElevenLabs</title>
<link>https://www.aisi.gov.uk/blog/advancing-voice-ai-security-with-elevenlabs</link>
<description><![CDATA[<p><strong>Organisation</strong> — Feb 18, 2026</p><p>New partnership exploring the security and societal implications of voice AI systems</p>



]]></description>
</item>
<item>




<guid isPermaLink="false">2bb613577052b0cf4002e69b9b08ac16</guid>
<pubDate>Thu, 16 Apr 2026 08:22:34 -0400</pubDate>
<title>Boundary Point Jailbreaking: A new way to break the strongest AI defences</title>
<link>https://www.aisi.gov.uk/blog/boundary-point-jailbreaking-a-new-way-to-break-the-strongest-ai-defences</link>
<description><![CDATA[<p><strong>Red Team</strong> — Feb 17, 2026</p><p>Introducing an automated attack technique that generates universal jailbreaks against the best defended systems</p>



]]></description>
</item>
<item>




<guid isPermaLink="false">0d063859e33bdfeeb46fe1d376553efb</guid>
<pubDate>Thu, 16 Apr 2026 08:22:23 -0400</pubDate>
<title>International consensus and open questions in AI evaluations</title>
<link>https://www.aisi.gov.uk/blog/international-ai-network-consensus-and-open-questions</link>
<description><![CDATA[<p><strong>Organisation</strong> — Feb 12, 2026</p><p>The International Network for Advanced AI Measurement, Evaluation and Science reflects on recent meeting and looks ahead to the India AI Impact Summit</p>



]]></description>
</item>
<item>




<guid isPermaLink="false">e12b35b191285b991f01499f7d452ca8</guid>
<pubDate>Thu, 16 Apr 2026 08:22:23 -0400</pubDate>
<title>AI and the future of work: Measuring AI-driven productivity gains for workplace tasks</title>
<link>https://www.aisi.gov.uk/blog/ai-and-the-future-of-work-measuring-ai-driven-productivity-gains-for-workplace-tasks</link>
<description><![CDATA[<p><strong>Analysis</strong> — Feb 2, 2026</p><p>Alongside the government’s new Future of Work Unit, we conducted a pilot study to explore how much AI models increase worker productivity for common tasks. </p>



]]></description>
</item>
<item>




<guid isPermaLink="false">d7ac7bf7de5cb3bfae89062b98c273d9</guid>
<pubDate>Thu, 16 Apr 2026 08:22:17 -0400</pubDate>
<title>Our 2025 year in review</title>
<link>https://www.aisi.gov.uk/blog/our-2025-year-in-review</link>
<description><![CDATA[<p><strong>Organisation</strong> — Dec 22, 2025</p><p> Adam Beaumont, Director of the UK AI Security Institute, reflects on the year's biggest achievements.</p>



]]></description>
</item>
<item>




<guid isPermaLink="false">c0568722714e916912b7b9e75e341e21</guid>
<pubDate>Thu, 16 Apr 2026 08:22:05 -0400</pubDate>
<title>5 key findings from our first Frontier AI Trends Report</title>
<link>https://www.aisi.gov.uk/blog/5-key-findings-from-our-first-frontier-ai-trends-report</link>
<description><![CDATA[<p><strong>Organisation</strong> — Dec 18, 2025</p><p>Our inaugural Frontier AI Trends Report draws on 2 years' worth of evaluations to provide accessible insights into the trajectory of AI development.</p>



]]></description>
</item>
<item>




<guid isPermaLink="false">6e54add5f483f5951099da19ee91d592</guid>
<pubDate>Thu, 16 Apr 2026 08:22:02 -0400</pubDate>
<title>Our approach to tackling AI-generated child sexual abuse material</title>
<link>https://www.aisi.gov.uk/blog/our-approach-to-tackling-ai-generated-child-sexual-abuse-material</link>
<description><![CDATA[<p><strong>Organisation</strong> — Dec 17, 2025</p><p>How we’re partnering with government and experts to prevent the creation and spread of AI‑generated CSAM </p>



]]></description>
</item>
<item>




<guid isPermaLink="false">f617eb1d5e3a3ea7438750a27b662235</guid>
<pubDate>Thu, 16 Apr 2026 08:21:54 -0400</pubDate>
<title>Stress-testing asynchronous monitoring of AI coding agents</title>
<link>https://www.aisi.gov.uk/blog/stress-testing-asynchronous-monitoring-of-ai-coding-agents</link>
<description><![CDATA[<p><strong>Control</strong> — Dec 16, 2025</p><p>Our new paper shares findings from an adversarial evaluation of monitoring systems for detecting sabotage by AI coding agents.
</p>



]]></description>
</item>
<item>




<guid isPermaLink="false">5e353b80fb1cfcfa5fb6b2de4e133768</guid>
<pubDate>Thu, 16 Apr 2026 08:21:47 -0400</pubDate>
<title>Deepening our partnership with Google DeepMind</title>
<link>https://www.aisi.gov.uk/blog/deepening-our-partnership-with-google-deepmind</link>
<description><![CDATA[<p><strong>Organisation</strong> — Dec 11, 2025</p><p>Expanding our collaboration with a new research MOU</p>



]]></description>
</item>
<item>




<guid isPermaLink="false">de81d4fe82063208f38523e8c41dccc3</guid>
<pubDate>Thu, 16 Apr 2026 08:21:47 -0400</pubDate>
<title>Auditing games for sandbagging detection</title>
<link>https://www.aisi.gov.uk/blog/auditing-games-for-sandbagging-detection</link>
<description><![CDATA[<p><strong>Model Transparency</strong> — Dec 9, 2025</p><p>Our new paper shares the results of an auditing game to evaluate ten methods for sandbagging detection in AI models. </p>



]]></description>
</item>
<item>




<guid isPermaLink="false">9d90f174fa4e2f76ec3686ed00d0b58f</guid>
<pubDate>Thu, 16 Apr 2026 08:21:35 -0400</pubDate>
<title>How do AI models persuade? Exploring the levers of AI-enabled persuasion through large-scale experiments</title>
<link>https://www.aisi.gov.uk/blog/how-do-ai-models-persuade-exploring-the-levers-of-ai-enabled-persuasion-through-large-scale-experiments</link>
<description><![CDATA[<p><strong>Human Influence</strong> — Dec 4, 2025</p><p>A deep dive into AISI’s study of the persuasive capabilities of conversational AI, published today in Science.</p>



]]></description>
</item>
<item>




<guid isPermaLink="false">dc371cb38efad5e19edeb8c37ec76b2c</guid>
<pubDate>Thu, 16 Apr 2026 08:21:35 -0400</pubDate>
<title>Investigating models for misalignment</title>
<link>https://www.aisi.gov.uk/blog/investigating-models-for-misalignment</link>
<description><![CDATA[<p><strong>Red Team</strong> — Nov 26, 2025</p><p>Insights from our alignment evaluations of Claude Opus 4.1, Sonnet 4.5, and a pre‑release snapshot of Opus 4.5.</p>



]]></description>
</item>
<item>




<guid isPermaLink="false">6ba8dffebc674e227c57f90d12c80963</guid>
<pubDate>Thu, 16 Apr 2026 08:21:29 -0400</pubDate>
<title>UKAISI at NeurIPS 2025</title>
<link>https://www.aisi.gov.uk/blog/ukaisi-at-neurips-2025</link>
<description><![CDATA[<p><strong>Organisation</strong> — Nov 26, 2025</p><p>An overview of the research we’ll be presenting at this year’s NeurIPS conference.</p>



]]></description>
</item>
<item>




<guid isPermaLink="false">8b229509141ec9e1eec506d08772d5e8</guid>
<pubDate>Thu, 16 Apr 2026 08:21:19 -0400</pubDate>
<title>Mapping the limitations of current AI systems</title>
<link>https://www.aisi.gov.uk/blog/mapping-the-limitations-of-current-ai-systems</link>
<description><![CDATA[<p><strong>Strategic Awareness</strong> — Oct 23, 2025</p><p>Takeaways from expert interviews on barriers to AI capable of automating most cognitive labour.</p>



]]></description>
</item>
<item>




<guid isPermaLink="false">4bd4c2a119aa857af84b677ce9e2da42</guid>
<pubDate>Thu, 16 Apr 2026 08:21:12 -0400</pubDate>
<title>Introducing ControlArena: A library for running AI control experiments</title>
<link>https://www.aisi.gov.uk/blog/introducing-controlarena-a-library-for-running-ai-control-experiments</link>
<description><![CDATA[<p><strong>Control</strong> — Oct 22, 2025</p><p>Our dedicated library to make AI control experiments easy, consistent, and repeatable. </p>



]]></description>
</item>
<item>




<guid isPermaLink="false">30875ee415c7b28ee0fba1f47633c4e9</guid>
<pubDate>Thu, 16 Apr 2026 08:21:07 -0400</pubDate>
<title>Transcript analysis for AI agent evaluations</title>
<link>https://www.aisi.gov.uk/blog/transcript-analysis-for-ai-agent-evaluations</link>
<description><![CDATA[<p><strong>Science of Evaluations</strong> — Oct 10, 2025</p><p>Why we use transcript analysis for our agent evaluations, and results from an early case study.</p>



]]></description>
</item>
<item>




<guid isPermaLink="false">5cbf3fcd9d3cfaa3cb707ade0f7a4921</guid>
<pubDate>Thu, 16 Apr 2026 08:21:00 -0400</pubDate>
<title>Examining backdoor data poisoning at scale</title>
<link>https://www.aisi.gov.uk/blog/examining-backdoor-data-poisoning-at-scale</link>
<description><![CDATA[<p><strong>Red Team</strong> — Oct 9, 2025</p><p>Our work with Anthropic and the Alan Turing Institute suggests that data poisoning attacks may be easier than previously believed.</p>



]]></description>
</item>
<item>




<guid isPermaLink="false">a224179ea245ecfc9e0910b6770f3e67</guid>
<pubDate>Thu, 16 Apr 2026 08:20:53 -0400</pubDate>
<title>Do chatbots inform or misinform voters?</title>
<link>https://www.aisi.gov.uk/blog/do-chatbots-inform-or-misinform-voters</link>
<description><![CDATA[<p><strong>Human Influence</strong> — Sep 30, 2025</p><p>What we learned from a large-scale empirical study of AI use for political information-seeking.</p>



]]></description>
</item>
<item>




<guid isPermaLink="false">aa6a81678a7c8e4f3256edf3539851e6</guid>
<pubDate>Thu, 16 Apr 2026 08:20:49 -0400</pubDate>
<title>How we’re working with frontier AI developers to improve model security</title>
<link>https://www.aisi.gov.uk/blog/how-were-working-with-frontier-ai-developers-to-improve-model-security</link>
<description><![CDATA[<p><strong>Red Team</strong> — Sep 13, 2025</p><p>Insights into our ongoing voluntary collaborations with Anthropic and OpenAI. </p>



]]></description>
</item>
<item>




<guid isPermaLink="false">aa6f359ab1cd5c90231e981212f71a08</guid>
<pubDate>Thu, 16 Apr 2026 08:20:43 -0400</pubDate>
<title>From bugs to bypasses: adapting vulnerability disclosure for AI safeguards</title>
<link>https://www.aisi.gov.uk/blog/from-bugs-to-bypasses-adapting-vulnerability-disclosure-for-ai-safeguards</link>
<description><![CDATA[<p><strong>Red Team</strong> — Sep 2, 2025</p><p>Exploring how far cyber security approaches can help mitigate risks in generative AI systems, in collaboration with the National Cyber Security Centre (NCSC).</p>



]]></description>
</item>
<item>




<guid isPermaLink="false">48be7ec67326068202340f7c06a97aeb</guid>
<pubDate>Thu, 16 Apr 2026 08:20:39 -0400</pubDate>
<title>Managing risks from increasingly capable open-weight AI systems</title>
<link>https://www.aisi.gov.uk/blog/managing-risks-from-increasingly-capable-open-weight-ai-systems</link>
<description><![CDATA[<p><strong>Red Team</strong> — Aug 29, 2025</p><p>Current methods and open problems in open-weight model risk management.</p>



]]></description>
</item>
<item>




<guid isPermaLink="false">87ff27fd25c6d0225e24a77642b193c9</guid>
<pubDate>Thu, 16 Apr 2026 08:20:35 -0400</pubDate>
<title>The Inspect Sandboxing Toolkit: Scalable and secure AI agent evaluations</title>
<link>https://www.aisi.gov.uk/blog/the-inspect-sandboxing-toolkit-scalable-and-secure-ai-agent-evaluations</link>
<description><![CDATA[<p><strong>Engineering</strong> — Aug 7, 2025</p><p>A comprehensive toolkit for safely evaluating AI agents.</p>



]]></description>
</item>
<item>




<guid isPermaLink="false">e672aadab6b9990d504a3902a56fc47d</guid>
<pubDate>Thu, 16 Apr 2026 08:20:27 -0400</pubDate>
<title>Announcing the Alignment Project: A global fund of over £15 million for AI alignment research</title>
<link>https://www.aisi.gov.uk/blog/announcing-the-alignment-project</link>
<description><![CDATA[<p><strong>Organisation</strong> — Jul 30, 2025</p><p></p>



]]></description>
</item>
<item>




<guid isPermaLink="false">94b3664f11b8b933ecfa5e76e2a83baa</guid>
<pubDate>Thu, 16 Apr 2026 08:20:20 -0400</pubDate>
<title>Navigating the uncharted: Building societal resilience to frontier AI</title>
<link>https://www.aisi.gov.uk/blog/navigating-the-uncharted-building-societal-resilience-to-frontier-ai</link>
<description><![CDATA[<p><strong>Societal Resilience</strong> — Jul 24, 2025</p><p>We outline our approach to study and address AI risks in real-world applications </p>



]]></description>
</item>
<item>




<guid isPermaLink="false">bf2e755cfc5ee373256d34cc0834d70f</guid>
<pubDate>Thu, 16 Apr 2026 08:20:12 -0400</pubDate>
<title>International joint testing exercise: Agentic testing</title>
<link>https://www.aisi.gov.uk/blog/international-joint-testing-exercise-agentic-testing</link>
<description><![CDATA[<p><strong>Organisation</strong> — Jul 17, 2025</p><p>Advancing methodologies for agentic evaluations across domains, including leakage of sensitive Information, fraud and cybersecurity threats.</p>



]]></description>
</item>
<item>




<guid isPermaLink="false">2fb4d34b01c9b2f04e81e5fd12239a6e</guid>
<pubDate>Thu, 16 Apr 2026 08:20:07 -0400</pubDate>
<title>A structured protocol for elicitation experiments</title>
<link>https://www.aisi.gov.uk/blog/our-approach-to-ai-capability-elicitation</link>
<description><![CDATA[<p><strong>Science of Evaluations</strong> — Jul 16, 2025</p><p>Calibrating AI risk assessment through rigorous elicitation practices.</p>



]]></description>
</item>
<item>




<guid isPermaLink="false">cb3a509c6f11a872dd79f34f674fa3ae</guid>
<pubDate>Thu, 16 Apr 2026 08:20:01 -0400</pubDate>
<title>Why we&#039;re working on white box control</title>
<link>https://www.aisi.gov.uk/blog/why-were-working-on-white-box-control</link>
<description><![CDATA[<p><strong>Control</strong> — Jul 10, 2025</p><p>An introduction to white box control, and an update on our research so far.</p>



]]></description>
</item>
<item>




<guid isPermaLink="false">47ebe64bef31d4d05f2a3edb98985892</guid>
<pubDate>Thu, 16 Apr 2026 08:19:59 -0400</pubDate>
<title>LLM judges on trial: A new statistical framework to assess autograders</title>
<link>https://www.aisi.gov.uk/blog/llm-judges-on-trial-a-new-statistical-framework-to-assess-autograders</link>
<description><![CDATA[<p><strong>Science of Evaluations</strong> — Jul 9, 2025</p><p>Our new framework can assess the reliability of LLM evaluators, while simultaneously answering a primary research question. </p>



]]></description>
</item>
<item>




<guid isPermaLink="false">5be5b224839c0fe8bfd8cc67dfcf9576</guid>
<pubDate>Thu, 16 Apr 2026 08:19:48 -0400</pubDate>
<title>How will AI enable the crimes of the future?</title>
<link>https://www.aisi.gov.uk/blog/how-will-ai-enable-the-crimes-of-the-future</link>
<description><![CDATA[<p><strong>Societal Resilience</strong> — Jul 3, 2025</p><p>How we're working to track and mitigate against criminal misuse of AI.</p>



]]></description>
</item>
<item>




<guid isPermaLink="false">21f85ef2a9dd8f24885d98489a17eb80</guid>
<pubDate>Thu, 16 Apr 2026 08:19:45 -0400</pubDate>
<title>Inspect Cyber: A New Standard for Agentic Cyber Evaluations</title>
<link>https://www.aisi.gov.uk/blog/inspect-cyber</link>
<description><![CDATA[<p><strong>Cyber & Autonomous Systems</strong> — Jun 26, 2025</p><p></p>



]]></description>
</item>
<item>




<guid isPermaLink="false">7736ac73988067de76e22cee5c3d33f1</guid>
<pubDate>Thu, 16 Apr 2026 08:19:35 -0400</pubDate>
<title>New updates to the AISI Challenge Fund</title>
<link>https://www.aisi.gov.uk/blog/new-updates-to-the-aisi-challenge-fund</link>
<description><![CDATA[<p><strong>Organisation</strong> — Jun 5, 2025</p><p></p>



]]></description>
</item>
<item>




<guid isPermaLink="false">56bc1a70d09eecf7ff8800a413c7cd73</guid>
<pubDate>Thu, 16 Apr 2026 08:19:30 -0400</pubDate>
<title>Making safeguard evaluations actionable</title>
<link>https://www.aisi.gov.uk/blog/making-safeguard-evaluations-actionable</link>
<description><![CDATA[<p><strong>Red Team</strong> — May 29, 2025</p><p>An Example Safety Case for Safeguards Against Misuse</p>



]]></description>
</item>
<item>




<guid isPermaLink="false">3c68aa924efab4a9497bdbbd8f2bdcec</guid>
<pubDate>Thu, 16 Apr 2026 08:19:29 -0400</pubDate>
<title>HiBayES: Improving LLM evaluation with hierarchical Bayesian modelling</title>
<link>https://www.aisi.gov.uk/blog/hibayes-improving-llm-evaluation-with-hierarchical-bayesian-modelling</link>
<description><![CDATA[<p><strong>Science of Evaluations</strong> — May 12, 2025</p><p>HiBayES: a flexible, robust statistical modelling framework that accounts for the nuances and hierarchical structure of advanced evaluations.</p>



]]></description>
</item>
<item>




<guid isPermaLink="false">768e2f61c5d849c3d438593b778d8dca</guid>
<pubDate>Thu, 16 Apr 2026 08:19:20 -0400</pubDate>
<title>Research Agenda</title>
<link>https://www.aisi.gov.uk/blog/research-agenda</link>
<description><![CDATA[<p><strong>Organisation</strong> — May 6, 2025</p><p>We outline our research priorities, our approach to developing technical solutions to the most pressing AI concerns, and the key risks that must be addressed as AI capabilities advance.</p>



]]></description>
</item>
<item>




<guid isPermaLink="false">63574797e7e7e1b0ec6bb659192faabb</guid>
<pubDate>Thu, 16 Apr 2026 08:19:14 -0400</pubDate>
<title>RepliBench: measuring autonomous replication capabilities in AI systems</title>
<link>https://www.aisi.gov.uk/blog/replibench-measuring-autonomous-replication-capabilities-in-ai-systems</link>
<description><![CDATA[<p><strong>Cyber & Autonomous Systems</strong> — Apr 22, 2025</p><p>A comprehensive benchmark to detect emerging replication abilities in AI systems and provide a quantifiable understanding of potential risks </p>



]]></description>
</item>
<item>




<guid isPermaLink="false">b25e71e049c8ebe226162c54718de2c3</guid>
<pubDate>Thu, 16 Apr 2026 08:19:08 -0400</pubDate>
<title>How to evaluate control measures for AI agents?</title>
<link>https://www.aisi.gov.uk/blog/how-to-evaluate-control-measures-for-ai-agents</link>
<description><![CDATA[<p><strong>Control</strong> — Apr 11, 2025</p><p>Our new paper outlines how AI control methods can mitigate misalignment risks as capabilities of AI systems increase </p>



]]></description>
</item>
<item>




<guid isPermaLink="false">64a56abfea153fd98895fc335606f4d1</guid>
<pubDate>Thu, 16 Apr 2026 08:19:04 -0400</pubDate>
<title>Strengthening AI resilience</title>
<link>https://www.aisi.gov.uk/blog/strengthening-ai-resilience</link>
<description><![CDATA[<p><strong>Organisation</strong> — Apr 3, 2025</p><p>20 Systemic Safety Grant Awardees Announced</p>



]]></description>
</item>
<item>




<guid isPermaLink="false">f362b7cac249c5d4dd4cbb9a4a2d1472</guid>
<pubDate>Thu, 16 Apr 2026 08:18:56 -0400</pubDate>
<title>How we’re addressing the gap between AI capabilities and mitigations</title>
<link>https://www.aisi.gov.uk/blog/aisis-research-direction-for-technical-solutions</link>
<description><![CDATA[<p><strong>Organisation</strong> — Mar 11, 2025</p><p>We outline our approach to technical solutions for misuse and loss of control.</p>



]]></description>
</item>
<item>




<guid isPermaLink="false">858678a9702cceacd836ab6573dbefa1</guid>
<pubDate>Thu, 16 Apr 2026 08:18:48 -0400</pubDate>
<title>How can safety cases be used to help with frontier AI safety?</title>
<link>https://www.aisi.gov.uk/blog/how-can-safety-cases-be-used-to-help-with-frontier-ai-safety</link>
<description><![CDATA[<p><strong>Safety Cases</strong> — Feb 10, 2025</p><p>Our new papers show how safety cases can help AI developers turn plans in their safety frameworks into action</p>



]]></description>
</item>
<item>




<guid isPermaLink="false">23eddf1110ead6b1a0b179a2d5a31ab2</guid>
<pubDate>Thu, 16 Apr 2026 08:18:43 -0400</pubDate>
<title>Principles for safeguard evaluation</title>
<link>https://www.aisi.gov.uk/blog/principles-for-safeguard-evaluation</link>
<description><![CDATA[<p><strong>Red Team</strong> — Feb 4, 2025</p><p>Our new paper proposes core principles for evaluating misuse safeguards</p>



]]></description>
</item>
<item>




<guid isPermaLink="false">3d68877eb26b2b579abdfa754f1bc03a</guid>
<pubDate>Thu, 16 Apr 2026 08:18:36 -0400</pubDate>
<title>Pre-Deployment evaluation of OpenAI’s o1 model</title>
<link>https://www.aisi.gov.uk/blog/pre-deployment-evaluation-of-openais-o1-model</link>
<description><![CDATA[<p><strong>Organisation</strong> — Dec 18, 2024</p><p>The UK Artificial Intelligence Safety Institute and the U.S. Artificial Intelligence Safety Institute conducted a joint pre-deployment evaluation of OpenAI's o1 model</p>



]]></description>
</item>
<item>




<guid isPermaLink="false">03db66593dc5792153bddb447cc46957</guid>
<pubDate>Thu, 16 Apr 2026 08:18:33 -0400</pubDate>
<title>Long-Form Tasks</title>
<link>https://www.aisi.gov.uk/blog/long-form-tasks</link>
<description><![CDATA[<p><strong>Science of Evaluations</strong> — Dec 3, 2024</p><p>A Methodology for Evaluating Scientific Assistants </p>



]]></description>
</item>
<item>




<guid isPermaLink="false">f357c7ad19bffd906398aeee907c1780</guid>
<pubDate>Thu, 16 Apr 2026 08:18:25 -0400</pubDate>
<title>Pre-deployment evaluation of Anthropic’s upgraded Claude 3.5 Sonnet</title>
<link>https://www.aisi.gov.uk/blog/pre-deployment-evaluation-of-anthropics-upgraded-claude-3-5-sonnet</link>
<description><![CDATA[<p><strong>Organisation</strong> — Nov 19, 2024</p><p>The UK Artificial Intelligence Safety Institute and U.S. Artificial Intelligence Safety Institute conducted a joint pre-deployment evaluation of Anthropic’s latest model</p>



]]></description>
</item>
<item>




<guid isPermaLink="false">a6c3d119962b11ea07e4e347f9857b08</guid>
<pubDate>Thu, 16 Apr 2026 08:18:20 -0400</pubDate>
<title>Safety case template for ‘inability’ arguments</title>
<link>https://www.aisi.gov.uk/blog/safety-case-template-for-inability-arguments</link>
<description><![CDATA[<p><strong>Safety Cases</strong> — Nov 14, 2024</p><p>How to write part of a safety case showing a system does not have offensive cyber capabilities</p>



]]></description>
</item>
<item>




<guid isPermaLink="false">637114948e132e203532f846d21e832a</guid>
<pubDate>Thu, 16 Apr 2026 08:18:11 -0400</pubDate>
<title>Our First Year</title>
<link>https://www.aisi.gov.uk/blog/our-first-year</link>
<description><![CDATA[<p><strong>Organisation</strong> — Nov 13, 2024</p><p>The AI Safety Institute reflects on its first year</p>



]]></description>
</item>
<item>




<guid isPermaLink="false">17d04696883bda82b8eb7123764f23af</guid>
<pubDate>Thu, 16 Apr 2026 08:18:06 -0400</pubDate>
<title>Announcing Inspect Evals</title>
<link>https://www.aisi.gov.uk/blog/inspect-evals</link>
<description><![CDATA[<p><strong>Organisation</strong> — Nov 13, 2024</p><p>We’re open-sourcing dozens of LLM evaluations to advance safety research in the field</p>



]]></description>
</item>
<item>




<guid isPermaLink="false">395a9be644c5dd98ed7426cd7d0f7278</guid>
<pubDate>Thu, 16 Apr 2026 08:18:05 -0400</pubDate>
<title>Bounty programme for novel evaluations and agent scaffolding</title>
<link>https://www.aisi.gov.uk/blog/evals-bounty</link>
<description><![CDATA[<p><strong>Organisation</strong> — Nov 5, 2024</p><p>We are launching a bounty for novel evaluations and agent scaffolds to help assess dangerous capabilities in frontier AI systems. </p>



]]></description>
</item>
<item>




<guid isPermaLink="false">2d6223b6414547c4e0a94f1516747912</guid>
<pubDate>Thu, 16 Apr 2026 08:17:53 -0400</pubDate>
<title>Early lessons from evaluating frontier AI systems</title>
<link>https://www.aisi.gov.uk/blog/early-lessons-from-evaluating-frontier-ai-systems</link>
<description><![CDATA[<p><strong>Organisation</strong> — Oct 24, 2024</p><p>We look into the evolving role of third-party evaluators in assessing AI safety, and explore how to design robust, impactful testing frameworks.</p>



]]></description>
</item>
<item>




<guid isPermaLink="false">35d6384046673bb35930c1974237553d</guid>
<pubDate>Thu, 16 Apr 2026 08:17:52 -0400</pubDate>
<title>Advancing the field of systemic AI safety: grants open</title>
<link>https://www.aisi.gov.uk/blog/advancing-the-field-of-systemic-ai-safety-grants-open</link>
<description><![CDATA[<p><strong>Organisation</strong> — Oct 15, 2024</p><p>Calling researchers from academia, industry, and civil society to apply for up to £200,000 of funding.</p>



]]></description>
</item>
<item>




<guid isPermaLink="false">103bc863549740f30f4ac2d4b13e29a0</guid>
<pubDate>Thu, 16 Apr 2026 08:17:44 -0400</pubDate>
<title>Why I joined AISI by Geoffrey Irving</title>
<link>https://www.aisi.gov.uk/blog/why-i-joined-aisi---geoffrey-irving</link>
<description><![CDATA[<p><strong>Organisation</strong> — Oct 3, 2024</p><p>Our Chief Scientist, Geoffrey Irving, on why he joined the UK AI Safety Institute and why he thinks other technical folk should too</p>



]]></description>
</item>
<item>




<guid isPermaLink="false">04e37a81007f88ea9fa9cb1fc76d8b08</guid>
<pubDate>Thu, 16 Apr 2026 08:17:35 -0400</pubDate>
<title>Should AI systems behave like people?</title>
<link>https://www.aisi.gov.uk/blog/should-ai-systems-behave-like-people</link>
<description><![CDATA[<p><strong>Human Influence</strong> — Sep 25, 2024</p><p>We studied whether people want AI to be more human-like. </p>



]]></description>
</item>
<item>




<guid isPermaLink="false">210341a1924ce6f401efb734cc691342</guid>
<pubDate>Thu, 16 Apr 2026 08:17:30 -0400</pubDate>
<title>Early Insights from Developing Question-Answer Evaluations for Frontier AI</title>
<link>https://www.aisi.gov.uk/blog/early-insights-from-developing-question-answer-evaluations-for-frontier-ai</link>
<description><![CDATA[<p><strong>Science of Evaluations</strong> — Sep 23, 2024</p><p>A common technique for quickly assessing AI capabilities is prompting models to answer hundreds of questions, then automatically scoring the answers. We share insights from months of using this method. </p>



]]></description>
</item>
<item>




<guid isPermaLink="false">a756d7435c2e6a90372436064839e4b3</guid>
<pubDate>Thu, 16 Apr 2026 08:17:25 -0400</pubDate>
<title>Conference on frontier AI safety frameworks</title>
<link>https://www.aisi.gov.uk/blog/conference-on-frontier-ai-safety-frameworks</link>
<description><![CDATA[<p><strong>Organisation</strong> — Sep 19, 2024</p><p>AISI is bringing together AI companies and researchers for an invite-only conference to accelerate the design and implementation of frontier AI safety frameworks. This post shares the call for submissions that we sent to conference attendees. </p>



]]></description>
</item>
<item>




<guid isPermaLink="false">41d682bd2e496e89d43d6a644fcf2222</guid>
<pubDate>Thu, 16 Apr 2026 08:17:18 -0400</pubDate>
<title>Cross-post: &quot;Interviewing AI researchers on automation of AI R&amp;D&quot; by Epoch AI</title>
<link>https://www.aisi.gov.uk/blog/interviewing-researchers-on-automation</link>
<description><![CDATA[<p><strong>Cyber & Autonomous Systems</strong> — Aug 27, 2024</p><p>AISI funded Epoch AI to explore AI researchers’ differing predictions on the automation of AI research and development and their suggestions for how to evaluate relevant capabilities. </p>



]]></description>
</item>
<item>




<guid isPermaLink="false">e2d0f67ea31e6b612d5dd88775c280bc</guid>
<pubDate>Thu, 16 Apr 2026 08:17:15 -0400</pubDate>
<title>Safety cases at AISI</title>
<link>https://www.aisi.gov.uk/blog/safety-cases-at-aisi</link>
<description><![CDATA[<p><strong>Safety Cases</strong> — Aug 23, 2024</p><p>As a complement to our empirical evaluations of frontier AI models, AISI is planning a series of collaborations and research projects sketching safety cases for more advanced models than exist today, focusing on risks from loss of control and autonomy. By a safety case, we mean a structured argument that an AI system is safe within a particular training or deployment context.</p>



]]></description>
</item>
<item>




<guid isPermaLink="false">079a04b250dd4e32225d23ee95bafe87</guid>
<pubDate>Thu, 16 Apr 2026 08:17:10 -0400</pubDate>
<title>Announcing our San Francisco office</title>
<link>https://www.aisi.gov.uk/blog/announcing-our-san-francisco-office</link>
<description><![CDATA[<p><strong>Organisation</strong> — May 20, 2024</p><p>We are opening an office in San Francisco! This will enable us to hire more top talent, collaborate closely with the US AI Safety Institute and engage even more with the wider AI research community.</p>



]]></description>
</item>
<item>




<guid isPermaLink="false">7cba5e1edaddc4f1fb422bdd36eb7cad</guid>
<pubDate>Thu, 16 Apr 2026 08:16:59 -0400</pubDate>
<title>Fourth progress report</title>
<link>https://www.aisi.gov.uk/blog/fourth-progress-report</link>
<description><![CDATA[<p><strong>Organisation</strong> — May 20, 2024</p><p>Since February, we released our first technical blog post, published the International Scientific Report on the Safety of Advanced AI, open-sourced our testing platform Inspect, announced our San Francisco office, announced a partnership with the Canadian AI Safety Institute, grew our technical team to >30 researchers and appointed Jade Leung as our Chief Technology Officer.</p>



]]></description>
</item>
<item>




<guid isPermaLink="false">15689a5786d4da862be02242d94ffdcd</guid>
<pubDate>Thu, 16 Apr 2026 08:16:55 -0400</pubDate>
<title>Advanced AI evaluations at AISI: May update</title>
<link>https://www.aisi.gov.uk/blog/advanced-ai-evaluations-may-update</link>
<description><![CDATA[<p><strong>Organisation</strong> — May 20, 2024</p><p>We tested leading AI models for cyber, chemical, biological, and agent capabilities and safeguards effectiveness. Our first technical blog post shares a snapshot of our methods and results.</p>



]]></description>
</item>
<item>




<guid isPermaLink="false">c39d464c05c521c5bc473482b7e801bd</guid>
<pubDate>Thu, 16 Apr 2026 08:16:49 -0400</pubDate>
<title>International Scientific Report on the Safety of Advanced AI: Interim Report</title>
<link>https://www.aisi.gov.uk/blog/international-scientific-report-on-the-safety-of-advanced-ai-interim-report</link>
<description><![CDATA[<p><strong>Organisation</strong> — May 17, 2024</p><p>This is an up-to-date, evidence-based report on the science of advanced AI safety. It highlights findings about AI progress, risks, and areas of disagreement in the field. The report is chaired by Yoshua Bengio and coordinated by AISI.</p>



]]></description>
</item>
<item>




<guid isPermaLink="false">1d73b28086d6a9f4482aedbc9b7c90c4</guid>
<pubDate>Thu, 16 Apr 2026 08:16:43 -0400</pubDate>
<title>Open sourcing our testing framework Inspect </title>
<link>https://www.aisi.gov.uk/blog/open-sourcing-our-testing-framework-inspect</link>
<description><![CDATA[<p><strong>Organisation</strong> — Apr 21, 2024</p><p>We open-sourced our framework for large language model evaluation, which provides facilities for prompt engineering, tool usage, multi-turn dialogue, and model-graded evaluations.</p>



]]></description>
</item>
<item>




<guid isPermaLink="false">ff350797a5dd2a0a20280e18937a8b85</guid>
<pubDate>Thu, 16 Apr 2026 08:16:37 -0400</pubDate>
<title>Announcing the UK and US AISI partnership</title>
<link>https://www.aisi.gov.uk/blog/announcing-the-uk-and-us-aisi-partnership</link>
<description><![CDATA[<p><strong>Organisation</strong> — Apr 2, 2024</p><p>The UK and US AI Safety Institutes signed a landmark agreement to jointly test advanced AI models, share research insights, share model access and enable expert talent transfers.</p>



]]></description>
</item>
<item>




<guid isPermaLink="false">eaf4df9e1a4cd3cda318eba5d1f51d69</guid>
<pubDate>Thu, 16 Apr 2026 08:16:30 -0400</pubDate>
<title>Announcing the UK and France AI Research Institutes’ collaboration</title>
<link>https://www.aisi.gov.uk/blog/announcing-the-uk-and-france-ai-research-institutes-collaboration</link>
<description><![CDATA[<p><strong>Organisation</strong> — Feb 29, 2024</p><p>The UK AI Safety Institute and France’s Inria (The National Institute for Research in Digital Science and Technology) are partnering to advance AI safety research. </p>



]]></description>
</item>
<item>




<guid isPermaLink="false">da4d2cd642c87975604a2cb0bf2bcc78</guid>
<pubDate>Thu, 16 Apr 2026 08:16:27 -0400</pubDate>
<title>Our approach to evaluations</title>
<link>https://www.aisi.gov.uk/blog/our-approach-to-evaluations</link>
<description><![CDATA[<p><strong>Organisation</strong> — Feb 9, 2024</p><p>This post offers an overview of why we are doing this work, what we are testing for, how we select models, our recent demonstrations and some plans for our future work. </p>



]]></description>
</item>
<item>




<guid isPermaLink="false">2a8483ab311a724a19e36a79ebac93bd</guid>
<pubDate>Thu, 16 Apr 2026 08:16:18 -0400</pubDate>
<title>Third progress report</title>
<link>https://www.aisi.gov.uk/blog/third-progress-report</link>
<description><![CDATA[<p><strong>Organisation</strong> — Feb 5, 2024</p><p>Since October, we have recruited leaders from DeepMind and Oxford, onboarded 23 new researchers, published the principles behind the International Scientific Report on Advanced AI Safety, and began pre-deployment testing of advanced AI systems.</p>



]]></description>
</item>
<item>




<guid isPermaLink="false">dc30c600687d81a03c325d1ef8d5e0cf</guid>
<pubDate>Thu, 16 Apr 2026 08:16:15 -0400</pubDate>
<title>First AI Safety Summit</title>
<link>https://www.aisi.gov.uk/blog/ai-safety-summit-2023</link>
<description><![CDATA[<p><strong>Organisation</strong> — Nov 2, 2023</p><p>At the first AI Safety Summit at Bletchley Park, world leaders and top companies agreed on the significance of advanced AI risks and the importance of testing.</p>



]]></description>
</item>
<item>




<guid isPermaLink="false">a6f35361c89be62184beb97efe2c6102</guid>
<pubDate>Thu, 16 Apr 2026 08:16:11 -0400</pubDate>
<title>Second progress report</title>
<link>https://www.aisi.gov.uk/blog/second-progress-report</link>
<description><![CDATA[<p><strong>Organisation</strong> — Oct 30, 2023</p><p>Since September, we have recruited leaders from OpenAI and Humane Intelligence, tripled the capacity of our research team, announced 6 new research partnerships, and helped establish the UK’s fastest supercomputer. </p>



]]></description>
</item>
<item>




<guid isPermaLink="false">e69eb79f5e4cea33982aa908e3439bc5</guid>
<pubDate>Thu, 16 Apr 2026 08:16:05 -0400</pubDate>
<title>First Progress Report</title>
<link>https://www.aisi.gov.uk/blog/first-progress-report</link>
<description><![CDATA[<p><strong>Organisation</strong> — Sep 7, 2023</p><p>In our first 11 weeks, we have recruited an advisory board of national security and ML leaders, including Yoshua Bengio, recruited top professors from Cambridge and Oxford and announced 4 research partnerships.</p>



]]></description>
</item>

  </channel>
</rss>

