
The UK AI Security Institute (AISI) released its first Frontier AI Trends report on 18 December 2025, summarised in a government factsheet. The report shows state‑of‑the‑art models rapidly improving across cyber, biology, and scientific tasks, sometimes surpassing human experts, while still failing more complex autonomous challenges.
This article aggregates reporting from 4 news sources. The TL;DR is AI-generated from original reporting. Race to AGI's analysis provides editorial context on implications for AGI development.
This is one of the clearest government‑backed snapshots we’ve seen of how fast frontier models are moving. AISI’s data shows capabilities in cyber, chemistry and biology, and advanced software engineering climbing at near‑exponential rates, with some benchmarks doubling roughly every eight months and several models now beating PhD‑level experts at scientific troubleshooting and protocol design. At the same time, the report stresses that fully autonomous, multi‑step cyberattacks still routinely fail, underscoring a gap between impressive narrow skills and robust, end‑to‑end autonomy. ([gov.uk](https://www.gov.uk/government/publications/ai-security-institute-frontier-ai-trends-report-factsheet))
For the race to AGI, the significance is twofold. First, this is a state actor validating that we are firmly in a new performance regime: models are now good enough to materially change the economics of science, cyber operations, and biolab work, even in the hands of non‑experts. Second, the UK is showing what a science‑led oversight regime can look like—embedding measurement infrastructure inside government rather than outsourcing risk assessment to labs. That raises the bar for other countries that want to be taken seriously on AI safety. Expect this report to become a reference point in regulatory debates, investment decks, and lab roadmaps alike.


