Blockchain CouncilGlobal Technology Council
ai6 min read

Google’s Project Genie

Michael WillsonMichael Willson
Google’s Project Genie

Introduction

Google’s Project Genie is a Google Labs web prototype that lets users generate and explore short, interactive 3D worlds in real time from text prompts and images. It is positioned as an early consumer-facing way to experience Google DeepMind’s “world model” research, specifically the Genie 3 model. If you want to understand what’s actually new here, and what is still research-stage, an AI certification helps because this is less about “AI art” and more about interactive environment generation that reacts to movement and control.

What Project Genie is

Project Genie is described by Google as an experimental research prototype for creating, exploring, and remixing interactive worlds. It is delivered as a web app under Google Labs, which signals its current intent: early access, limited scope, and rapid iteration rather than a finished product.

What powers Project Genie

Google says Project Genie is powered by a stack of models and workflow components:

Genie 3 is the world model generating navigable environments as you move.

Nano Banana Pro is used to preview and refine the visual concept of the world before entering it.

Gemini is part of the overall prototype workflow.

This matters because it clarifies that the experience is not a single model generating everything. It is a pipeline designed to support concept control, interactive navigation, and rapid remixing.

What Genie 3 is showing off

DeepMind describes Genie 3 as a general-purpose world model that uses simple text descriptions to generate photorealistic environments that can be explored in real time. Project Genie is effectively a consumer-friendly wrapper around that idea, letting users interact with a world that extends forward as they move instead of watching a fixed clip.

World sketching

World sketching is the first core capability. Users provide text prompts and can optionally generate or upload images to define the environment and a character. Users also choose how they want to explore, with options such as walking, riding, flying, or driving.

Google says Nano Banana Pro is integrated for tighter control at this stage, including previewing and adjusting the image before entering the world. Users can also choose viewpoint, such as first-person or third-person. This stage exists to reduce mismatch between the imagined world and the generated one by allowing visual iteration before exploration begins.

World exploration

World exploration is the interactive stage. Once inside, the environment is navigable. As the user moves, the system generates the path ahead in real time based on the user’s actions. The camera can be adjusted while moving through the scene.

This is the core “world model” demonstration: generation that is conditioned on movement and interaction rather than being a pre-generated scene.

World remixing

World remixing is the third core capability. Users can take existing worlds and create new variations by modifying prompts. Users can also explore curated worlds through a gallery or randomizer, and export videos of both the generated worlds and their exploration sessions.

This feature matters because it turns Project Genie into a looped workflow. It supports iteration, variation testing, and shareable outputs, which is closer to creative prototyping than one-off generation.

Who can access Project Genie

As of today, Google states access is rolling out to Google AI Ultra subscribers in the US, and users must be 18 or older.

Google’s Labs help page also states Project Genie is only available in the US and currently only for Google AI Ultra subscribers.

Google additionally notes the prototype is not available to Google AI Ultra for Business users.

Google’s AI Ultra plan pricing was announced at $249.99 per month in the US, with an introductory discount for first-time users at launch.

These constraints define the current audience. This is not a broadly available consumer feature. It is a gated prototype.

Known limitations

Google explicitly lists current limitations:

Worlds may not look fully true-to-life and may not closely follow prompts, images, or real-world physics.

Character control can be less reliable or have higher control latency.

Generations are limited to 60 seconds.

Google also says some Genie 3 capabilities previously discussed are not yet included in this prototype, including promptable events that change the world as you explore.

These limitations matter because they define what Project Genie is best for today: short-form exploration and prototyping, not extended gameplay, high-precision simulation, or strict physics-based behavior.

Why Project Genie matters

Google’s stated motivation is broader world model research: systems that can simulate how environments evolve and how actions change them. Google frames this as a step toward more general AI systems and applications ranging from robotics to modeling, animation, and interactive experiences.

Market commentary has focused on potential impact on game creation tools and the economics of game production. Financial coverage has highlighted investor concern after launch, framing Project Genie as a sign that high-level world generation could eventually alter parts of the game production pipeline, even though today’s version is still an early prototype with strict constraints.

The realistic takeaway is that the prototype demonstrates a workflow shift: from creating content to generating interactive environments that can be explored and iterated quickly.

How it connects to earlier “Genie” research

Project Genie sits on top of an ongoing research line. A relevant anchor point is the 2024 paper “Genie: Generative Interactive Environments,” which introduced a generative interactive environment trained in an unsupervised way using unlabeled internet video. That work framed Genie as a foundation world model that can be prompted to create action-controllable worlds.

Project Genie is a productized prototype of that broader direction: interactive worlds that respond to action, not just static outputs.

Conclusion

Google’s Project Genie is an experimental Google Labs web prototype for generating, exploring, and remixing short interactive 3D worlds in real time from text prompts and images. It demonstrates DeepMind’s Genie 3 world model through three core capabilities: world sketching with prompt and image control plus viewpoint and exploration mode selection, world exploration where the environment generates ahead as users move, and world remixing through prompt edits, curated discovery, and video export.

Access is currently restricted to US-based Google AI Ultra subscribers aged 18 and up, excludes AI Ultra for Business users, and is tied to the AI Ultra plan priced at $249.99 per month in the US with an introductory discount at launch. Google also sets explicit constraints: worlds may not fully match prompts or physics, character control may be less reliable or slower, generations are limited to 60 seconds, and some previously discussed Genie 3 capabilities such as promptable world-changing events are not yet present. The significance is not that it replaces existing tools today, but that it makes world model research accessible as an interactive, consumer-facing prototype with clear implications for future creation workflows.

Google’s Project Genie