MITB Banner

Google’s Genie Can Reshape Interactive Video Games Creation 

Google's Genie AI turns sketches and photos into interactive games shows promise to transform video game generation.

Share

Illustration by Pig Minecraft Pixel Video Game Blocks Game Pixels

Listen to this story

Google DeepMind released Genie, an AI model that transforms text descriptions, sketches, and photographs into interactive virtual environments. It uses an architecture with 11 billion parameters and is trained on 200,000 hours of unlabelled Internet videos for understanding and replicating environmental dynamics without manual data labeling.

Tim Rocktäschel, the team lead for Genie wrote on X, “Rather than adding inductive biases, we focus on scale. In an unsupervised way, Genie learns diverse latent actions that control characters in a consistent manner.”

This allowed it to consistently learn a diverse range of character motion, control and action. As a result, “our model can convert any image into a playable 2D world,” explained Rocktäschel.

Genie combines a spatiotemporal video tokenizer that breaks down videos to  understand movement and change over time. Next, the autoregressive dynamics model predicts what will happen next in the virtual environment based on this analysis. Finally a scalable latent action model creates possible actions within the virtual world that weren’t directly shown during training. It essentially ‘imagines’ it and scales to accommodate a wide range of potential interactions. 

Genie is a research so far it is unclear if it will become a real product. But if it does it has applications beyond entertainment in virtual reality, training simulations, architectural design, and urban planning etc. 

Building on DeepMind’s AI contributions, Genie expands into the visual domain, enabling creative expression and interactive experiences. Previously Google DeepMind released DreamerV2 and V3, which focus on learning from interactions within environments to foster planning and goal-oriented behavior. 

Unlike Genie, which observes and learns from video data, Dreamer models require interaction data to learn, making this model distinct in its method of understanding and creating virtual worlds.

Share
Picture of K L Krithika

K L Krithika

K L Krithika is a tech journalist at AIM. Apart from writing tech news, she enjoys reading sci-fi and pondering the impossible technologies, trying not to confuse it with reality.
Related Posts

CORPORATE TRAINING PROGRAMS ON GENERATIVE AI

Generative AI Skilling for Enterprises

Our customized corporate training program on Generative AI provides a unique opportunity to empower, retain, and advance your talent.

Upcoming Large format Conference

May 30 and 31, 2024 | 📍 Bangalore, India

Download the easiest way to
stay informed

Subscribe to The Belamy: Our Weekly Newsletter

Biggest AI stories, delivered to your inbox every week.

AI Courses & Careers

Become a Certified Generative AI Engineer

AI Forum for India

Our Discord Community for AI Ecosystem, In collaboration with NVIDIA. 

Flagship Events

Rising 2024 | DE&I in Tech Summit

April 4 and 5, 2024 | 📍 Hilton Convention Center, Manyata Tech Park, Bangalore

MachineCon GCC Summit 2024

June 28 2024 | 📍Bangalore, India

MachineCon USA 2024

26 July 2024 | 583 Park Avenue, New York

Cypher India 2024

September 25-27, 2024 | 📍Bangalore, India

Cypher USA 2024

Nov 21-22 2024 | 📍Santa Clara Convention Center, California, USA

Data Engineering Summit 2024

May 30 and 31, 2024 | 📍 Bangalore, India

Subscribe to Our Newsletter

The Belamy, our weekly Newsletter is a rage. Just enter your email below.