Everything Unveiled at Google I/O 2025
61 views
May 22, 2025
Watch all the biggest announcements from Google I/O 2025
View Video Transcript
0:00
Hello everyone, good morning. Welcome to Google I.O
0:03
I learned that today is the start of Gemini season. I'm not really sure what the big deal is
0:09
Every day is Gemini season here at Google. A couple of weeks ago, Gemini completed Pokemon Blue
0:16
It earned all eight badges, traveled the victory road, defeated the Elite Four, and the Champion
0:23
bringing us one step closer to achieving API, artificial Pokemon intelligence. introducing Google Beam, a new AI-first video communications platform
0:37
Beam uses a new state-of-the-art video model to transform 2D video streams into a realistic 3D experience
0:46
Behind the scenes, an array of six cameras captures you from different angles
0:52
And with AI, we can merge these video streams together and render you on a 3D light field display
0:59
In collaboration with HP, the first Google Beam devices will be available for early customers later this year
1:08
And today, we are introducing this real-time speech translation directly in Google Meet
1:14
English and Spanish translation is now available for subscribers with more languages rolling out in the next few weeks
1:22
And real-time translation will be coming to enterprises later this year. We also have our research prototype, Project Mariner
1:30
It's an agent that can interact with the web and get stuff done
1:35
Say you want to find an apartment for you and two roommates in Austin
1:39
You've each got a budget of $1,200 a month. You want a washer dryer or at least a laundromat nearby
1:46
Using agent mode, the Gemini app goes to work behind the scenes
1:51
It finds listings from sites like Zillow that match your criteria and uses Project Mariner when needed to adjust very specific filters
2:00
An experimental version of the agent mode in the Gemini app will be coming soon to subscribers
2:06
We are working to bring this to life with something we call personal context
2:11
With your permission, Gemini models can use relevant context across your Google Apps
2:17
in a way that is private, transparent, and fully under your control
2:22
Let me show you an example in Gmail. You might be familiar with our AI-powered smart reply features
2:28
It's amazing how popular they are. Now imagine if those responses could sound like you
2:35
That's the idea behind personalized smart replies. Gemini 2.5 Pro is our most intelligent model ever
2:42
and the best foundation model in the world. Just two weeks ago, we shipped a preview of an updated 2.5 Pro
2:49
so you could get your hands on it and start building with it right away. Today, I'm thrilled to announce that we're releasing an updated version of 2.5 Flash
2:58
The new Flash is better in nearly every dimension, improving across key benchmarks for reasoning, code, and long context
3:06
I'm excited to say that Flash will be generally available in early June with Pro soon after
3:12
But you can go try out the preview now in AI Studio, Vertex AI, and Gemini app
3:18
We are also introducing new previews for text-to-speech. These now have a first-of-its-kind multi-speaker support for two voices
3:29
This means the model can converse in more expressive ways. It can capture the really subtle nuances of how we speak
3:37
This works in over 24 languages, and it can easily go between languages
3:43
You can use this text-to-speech capability starting today in the Gemini API
3:48
Gemini 2 Pro is incredible at coding I got 2 pro to code me a simple web app in Google AI Studio to share some photos and learn more And I going to jump to another tab that I ran right before this keynote
4:02
with the same prompt. And here's what Gemini generates. Whoa. 2.5 Pro is available on your
4:10
favorite IDE platforms and in Google products like Android Studio, Firebase Studio, Gemini Code
4:18
and our asynchronous coding agent, Jules. Jules can tackle complex tasks in large code bases that used to take hours
4:27
Jules is now in public beta, so anyone can sign up at Jules.Google
4:32
Today, we're making 2.5 Pro even better by introducing a new mode we're calling DeepThink
4:39
It pushes model performance to its limits, delivering groundbreaking results. DeepThink uses our latest cutting-edge research in thinking and reasoning
4:49
We're going to make it available to trusted testers via the Gemini API to get their feedback before making it widely available
4:56
Gemini is already the best multimodal foundation model, but we're working hard to extend it to become what we call a world model
5:05
That is a model that can make plans and imagine new experiences by simulating aspects of the world, just like the brain does
5:13
We recently partnered with Aira, a company that assists people in the blind and low vision community to navigate the world by connecting them via video to human visual interpreters
5:24
Using Astra technology, we built a prototype to help more people have access to this type of assistance
5:30
For those who want an end-to-end AI search experience, we are introducing an all-new AI mode
5:37
and I'm excited to share that AI Mode is coming to everyone in the U.S. starting today
5:44
And starting this week, Gemini 2.5, our most intelligent model series, is coming to Search
5:52
With AI Mode, you can ask whatever's on your mind. And as you can see here, Search gets to work
5:58
It generates your response, putting everything together for you, including links to content and creators you might not have otherwise discovered
6:06
and merchants and businesses with useful information like ratings. Soon, AI Mode will be able to make your responses even more helpful
6:15
with personalized suggestions based on your past searches. You can also opt in to connect other Google Apps, starting with Gmail
6:24
We call this personal context, and you'll see when AI Mode is bringing yours in to help
6:30
Deep Search uses the same query fan-out technique you just heard about, but multiplied
6:35
It can issue dozens or even hundreds of searches on your behalf to create an expert-level, fully cited report in just minutes
6:44
Right here, task complete. I get great ticket options with helpful context so I can make an informed decision
6:52
We're taking the next big leap in multimodality by bringing Project Astra's live capabilities into AI mode
6:59
We call this Search Live. And now, using your camera, Search can see what you see
7:05
and give you helpful information as you go back and forth in real time
7:10
With AI mode, we're bringing a new level of intelligence to help you shop with Google
7:16
I want a few new dresses for summer. Search gave me some great options
7:22
But now, I'm faced with the classic online shopping dilemma. I have no clue how these styles will look on me
7:32
So we are introducing a new try-on feature that will help you virtually try on clothes
7:39
so you get a feel for how styles might look on you
7:43
Let me show you a live demo I really like this blue one I click on this button to try it on It asks me to upload a picture which takes me to my camera roll
7:57
I have many pictures here. I'm going to pick one that is full length and a clear view of me
8:03
And off it goes. To do this, we build a custom image generation model specifically trained for fashion
8:13
Wow, and it's back. Our new intelligent capabilities bring the best of shopping together with the best of AI right here in Search
8:23
Our new visual shopping and agent-to-checkout features are rolling out in the coming months
8:28
And you can start trying on looks in labs beginning today. We're launching five things today
8:35
First, let's talk about Gemini Live. People are blown away by how interactive and natural the conversations are
8:43
And it works in over 45 languages, more than 150 countries. And Gemini Live now includes camera and screen sharing, both of which are incredible
8:53
All of it is rolling out free of charge in the Gemini app on Android and iOS today
8:58
Starting today, Deep Research will now let you upload your own files to guide the research agent
9:04
which is one of the top requested features. And soon, we'll let you research across Google Drive and Gmail
9:11
This is where Canvas comes in. It's Gemini's interactive space for co-creation
9:17
Canvas will now let you transform that report with one tap into all kinds of new things
9:23
like a dynamic web page, an infographic, a helpful quiz, even a custom podcast in 45 languages
9:30
We're introducing Gemini in Chrome. This will be your AI assistant that's there for you as you browse the web on your desktop
9:40
The amazing part is that you can use this, and it understands the context of the page that you're on automatically
9:47
So if you have a question, it can be answered. Starting today, we're bringing our latest and most capable image generation model into the Gemini app
9:58
It's called Imagine 4, and it's a big leap forward. Maybe I want to create a poster for a music festival
10:05
We'll make the Chrome Dino the big headliner. Imagine 4 doesn't just get the text and spelling right
10:12
It's actually making creative choices, like using dinosaur bones in the font
10:17
or figuring out the spacing, the font size, the layout that makes it look like this great poster
10:22
I'm excited to announce our new state-of-the-art model, VO3. It's available today
10:30
VO3 comes with native audio generation. Now you prompt it, and your characters can speak
10:37
They left behind a ball today. What manner of magic is that
10:42
The quality is so good, it feels like you're there, on the boat, with this guy
10:49
This ocean, it's a force, a wild, untamed might. Gemini Live capabilities are free and rolling out across Android and iOS
10:58
We recently launched Lyria 2, which can generate high-fidelity music and professional-grade audio
11:03
The music is melodious, with vocals in solos and choirs. As you hear, it makes expressive and rich music
11:17
Learia 2 is available today for enterprises, YouTube creators, and musicians. Two years ago, we pioneered SynthID, which embeds invisible watermarks into generated media
11:28
Our new synth ID detector can identify if an image, audio track, text, or video has synth ID in it
11:34
We're starting to roll this out to early testers today. We built new capabilities for filmmakers
11:40
So when you making a video it will use ingredients you give it characters scenes or styles and keep them consistent Or you can direct Vio giving it precise camera instructions and have it shoot along a specific path
11:54
We've been building a new AI filmmaking tool for creatives, one that combines the best of Vio, Imagine, and Gemini
12:04
We're calling it Flow, and it's launching today. Our hero, the grandpa, is building a flying car with help from a feathered friend
12:13
These are my ingredients, the old man and his car. We make it easy to upload your own images into the tool
12:20
or you can generate them on the fly using Imagine, which is built right in
12:25
With a single prompt, you can describe what you want, including very precise camera controls
12:31
Flow puts everything in place. The character consistency, the scene consistency, it just works
12:39
But what about emerging form factors that could let you experience an AI assistant in new ways
12:47
That's exactly why we're building Android XR. We believe there's not a one-size-fits-all for XR, and you'll use different devices throughout your day
12:56
We built Android XR together as one team with Samsung and optimized it for Snapdragon with Qualcomm
13:03
This is Samsung's Project Wuhan, the first Android XR device. Samsung's Project Wuhan will be available for purchase later this year
13:16
Glasses with Android XR are lightweight and designed for all-day wear, even though they're packed with technology
13:24
A camera and microphones give Gemini the ability to see and hear the world
13:31
Speakers let you listen to the AI, play music, or take calls
13:35
And an optional in-lens display privately shows you helpful information just when you need it
13:42
These glasses work with your phone, giving you access to your apps while keeping your hands free
13:50
Hey, everyone. Right now, you should be seeing exactly what I'm seeing through the lens of my Android XR glasses
13:57
like my delicious coffee over here and that text from Sharama that just came in
14:02
Let's see what he said. All right, it's definitely showtime, so I'm going to launch Gemini and get us going
14:11
Send Sharama text that I'm getting started and silence my notifications, please
14:18
Okay, I've sent that message to him and muted all your notifications
14:22
curtains. Gemini, what was the name of the coffee shop on the cup I had earlier
14:31
Hmm, that might have been Bloomsgiving. From what I can tell, it's a vibrant coffee shop
14:35
on Castro Street. Can you show me the photos of that cafe? I want to check out the vibes
14:42
Definitely. Do these photos from maps help? This is a very risky demo, but we're going to give it
14:49
shop. Nishter and I are going to speak to each other in our mother tongues. Nishter's
14:56
going to speak Hindi. I'm going to speak Farsi very poorly. Asani say. Gentle Monster and Warby Parker will be the first eyewear partners to build glasses
15:28
with Android XR. I can't wait for you to try Android XR for yourself and we'll have lots more
15:35
to share in the months ahead. Thank you so much
#Celebrities & Entertainment News
#Email & Messaging
#Intelligent Personal Assistants
#Software
#Teleconferencing
#Text & Instant Messaging
#Voice & Video Chat