NextFin News - This article draws on the featured exchanges and clips assembled in the short documentary "Google's Plan to Dominate Real-World and Physical AI," together with identifiable filmed appearances and company announcements cited in the film. The documentary interleaves Hassabis’s remarks with other voices and material recorded across multiple occasions: interviews and filmed sessions with DeepMind and press organizations (including a Nobel interview recorded during Nobel Week in Stockholm, Sweden on 2024-12-06), DeepMind’s public announcements (AlphaEvolve, unveiled by DeepMind on 2025-05-14), and longer-form filmed interviews released by outlets such as Veritasium (published online 2025-12-25). The principal interviewee throughout the documentary is Demis Hassabis, CEO of Google DeepMind; the assembled footage also includes other on-record speakers and corporate remarks that contextualize his statements.
The following sections present Hassabis’s core statements and observations as they appear in the documentary transcript, arranged by theme and quoted directly where indicated.
Origins and the Transformer breakthrough
Hassabis’s remarks in the film are placed alongside the bigger story of modern generative AI: the Transformer architecture and the wave of large language models that followed. While the documentary traces how an external competitor popularised conversational AI, Hassabis’s perspective underscored a longer internal arc — from research breakthroughs to productised families of models. The film notes Google’s iteration from early chat systems to a multi-model family, and Hassabis’s contributions are presented in that context as part of DeepMind’s broader research effort.
AlphaFold and the scale of scientific acceleration
On AlphaFold, Hassabis describes a dramatic change in the pace of structural biology. He recalls the decades-long human effort that produced roughly 150,000 experimentally solved protein structures and contrasts that with DeepMind’s computational effort to predict structures at scale. In his own words the film records:
"For 50 years the work of tens of thousands of scientists revealed the structures of 150,000 proteins… in a few years your team… found 200 million."
Hassabis frames the release of those predictions as an enabling act for science: the team published the structures and watched the scientific community engage with them in real time. The documentary cites his description of the moment as "humbling and amazing," and places special emphasis on the idea that this is not merely a database but a platform for downstream discovery.
On the implications for drug discovery, the transcript reproduces Hassabis’s expectation that AI-driven workflows can dramatically shorten timelines: "On average it takes 10 years and billions of dollars to design just one drug. We could maybe reduce that down from years to months."
The film also notes the spin‑out of Isomorphic Labs as part of the pathway from structural prediction to drug design.
AlphaEvolve: AI that invents algorithms
The documentary highlights DeepMind’s 2025 announcement of AlphaEvolve and records Hassabis’s account of the system’s purpose: to use large models in an evolutionary framework to generate, test and iterate on code and algorithms automatically. The film quotes his explanation of how AlphaEvolve works at scale within Google:
"AlphaEvolve spins up thousands of code variants, tests them automatically, and keeps mutating the winners to find faster, more efficient solutions."
According to Hassabis’s remarks as included in the transcript, AlphaEvolve has already produced practical optimisations inside Google: new scheduling heuristics for data centres, suggestions for TPU design tweaks, and faster matrix-multiplication kernels that reduce energy and training time without hardware changes. The documentary presents these outcomes as early evidence of agentic systems that can contribute to algorithmic discovery.
World models, simulation and multiworld agents
Hassabis stresses the research value of simulated rehearsal spaces. The film reproduces his and DeepMind’s account of projects that build internal models of environments and enable agents to practise long sequences of behaviour before deployment. The documentary records examples such as Genie (and later versions) that can convert a single image or short clip into an interactive 3D scene, and SIMA, the Scalable Instructable Multiworld Agent, which Hassabis and the researchers describe as learning to follow plain-language instructions across many different virtual worlds and games.
He illustrates the emergent behaviour these systems can develop through iterative self-play and multi-environment training: "At first all the players just run after the ball together like a gaggle of six-year-olds… over time you see strategy and coordination emerging."
The film presents that progression as an argument for using diverse simulated experiences to build perception, planning and instruction-following prior to any physical embodiment.
Robotics, Gemini robotics models and Project Astra
Moving from virtual rehearsal to hardware, Hassabis’s statements in the transcript describe efforts to consolidate vision, language and action into robotics models that generalise across different platforms. The documentary cites example model names and research lines — RT-1, RT-2 and more recent Gemini-powered robotics models — that learn from large collections of demonstrations so they can be applied to different robot arms and early humanoid platforms.
The film also records Hassabis’s articulation of Project Astra, Google’s ambition for an assistant that sees the world through a camera, understands context and responds in real time: "We’ll have a system that really understands everything around you in very nuanced and deep ways, embedded in your everyday life."
The documentary lists envisioned outcomes — glasses that whisper contextual information, robots that follow messy human-level instructions, and cars that drive better than humans — and places Hassabis’s comments as a statement of intent to move AI from screens into everyday physical settings.
Infrastructure, scale and the company’s long game
Throughout the assembled footage the documentary juxtaposes Hassabis’s technical remarks with the wider corporate argument about the importance of infrastructure. The film reiterates that decades of investment in data centres, fibre and custom accelerators (TPUs) give Google and DeepMind a cost and scale advantage when running large models. Hassabis’s remarks on using simulated worlds, combined with AlphaEvolve’s internal optimisations, are presented as part of a strategy to reduce the marginal cost of deploying advanced models.
At the same time the transcript repeats his and others’ acknowledgement of the risks and public scrutiny that come with rapid iteration: early chatbot errors, bias concerns and the need to win over developers, users and regulators are presented as recurring challenges in parallel with technical progress.
Science, safety and the next decade
Closing the interview excerpts, Hassabis frames AlphaFold and related projects as the opening chapter of a broader scientific renaissance. The documentary records his view that AI tools will become core laboratory instruments, enabling breakthroughs across domains: from new medicines to energy and climate work. He reiterates the long-term aspiration that better world models, agentic systems and embodied AI will expand what is possible for science, while also acknowledging that embedding AI into the physical world raises questions about when software becomes a new kind of actor in society.
Below are the principal filmed sources and public announcements cited or referenced in the documentary.
References: DeepMind — AlphaEvolve announcement (2025-05-14); NobelPrize.org — Interview with Demis Hassabis (recorded 2024-12-06); Veritasium — "AlphaFold: The Most Useful Thing AI Has Ever Done" (video published 2025-12-25); AlphaFold — background and database history.
Explore more exclusive insights at nextfin.ai.

