Google I/O 2025: Gemini Ultra, Veo 3, Deep Think, AI Updates

Held in an enthusiastic atmosphere at the Shoreline Amphitheatre in Mountain View, Google's annual developer conference, Google I/O 2025, revealed a massive array of innovations.

Primarily, these advancements centered on artificial intelligence.

The two-day event covered a wide range of products and services, extending from Android, Chrome, and YouTube, all the way to the star of the show: the Gemini AI ecosystem.

Google's AI Updates and Innovations

Gemini Ultra: A Premium Subscription at an Astounding Price

Google announced the launch of "Gemini Ultra," a new subscription plan positioned as the most expensive in the company's history, priced at $249.99 per month.

This premium package offers exclusive access to advanced AI tools.

Among these are the Veo 3 video generator, the Flow editing application, and the "Deep Think" mode within the Gemini 2.5 Pro version.

The package also includes an enormous storage capacity of up to 30 terabytes across Google Drive, Photos, and Gmail services.

Additionally, it provides a YouTube Premium subscription and "agentic" tools powered by the revolutionary Project Mariner.

Deep Think: A New Level of Artificial Comprehension

The "Deep Think" mode represents a powerful addition to AI models' capabilities.

It enables the model to deliberate more deeply on multiple possibilities before delivering an answer, thereby significantly increasing analytical accuracy.

Currently, this feature remains exclusively accessible to a select group of programmers via the Gemini API, pending Google's completion of its security assessments.

Veo 3 and Imagen 4: Evolution and Creativity in Video and Image Generation

Google introduced the third iteration of its Veo video generator. This model can create AI-enhanced sound effects, dialogues, and cinematic environments.

As for Imagen 4, it stands as a new image generation model capable of resolutions up to 2K. It also boasts a speed ten times faster than its predecessor, and it can produce intricate details such as fabric textures, water droplets, and animal fur.

Both these tools will form the core of the new "Flow" platform designed for AI-powered film creation and video editing, catering specifically to content creators.

Gemini Live: Real-time Conversations with Camera and Instant Sharing

Camera and screen-sharing capabilities for "Gemini Live" are set to roll out to all iOS and Android users.

This feature facilitates instant voice conversations with AI while streaming live video from a phone's camera or screen.

Users will be able to interact with Gemini by pointing to real-world objects and receiving immediate responses.

Gemini Nano with On-Device Computer Vision

Google announced the introduction of "visual processing" capabilities into Gemini Nano models.

These are the lightweight versions specifically designed for local operation on smartphones. Now, Gemini Nano possesses the ability to comprehend images, videos, and visual content.

Such advancement paves the way for a digital assistant that understands its surrounding environment and provides assistance even offline.

Pixel phones will be the first to receive these advanced capabilities, with subsequent support planned for newer Android devices.

Stitch and Jules: Smart Tools for Software Development

"Stitch" emerges as a new tool, facilitating application interface design through simple descriptions or even images.

It automatically generates HTML and CSS codes. Additionally, the capabilities of the "Jules" programming assistant have been expanded.

The assistant now supports understanding complex code, creating GitHub pull requests, and managing overdue programming tasks.

Project Mariner: Browsing and Task Execution Without Website Visits

Google revealed significant developments in Project Mariner.

This project allows AI to execute approximately 10 to 12 simultaneous online tasks on behalf of the user, such as purchasing tickets or ordering groceries.

All of this occurs without the need for manual website visits, as it now operates on virtual machines in the cloud.

Currently, this capability is available to AI Ultra subscribers in the United States, and it will be accessible to developers via Gemini API and Vertex AI.

Project Astra and Upcoming Google Glasses

Google disclosed ongoing development for Project Astra, its low-latency multimodal experience. This project will support applications like Search and the Gemini app.

The company also reported collaborations with entities such as Samsung and Warby Parker for developing smart glasses, though no launch date has been specified yet.

AI Mode and Enhanced Search: The Future of Search Begins Now

Google introduced "AI Mode" into its search engine. It enables users to pose complex multi-part questions, and answers are provided in natural language.

Furthermore, an upcoming addition is the "Search Live" feature, which relies on your phone's camera view for instant answers.

Google Meet: Real-time Translation During Meetings

In a revolutionary step, Google added an instant translation feature for voice conversations in Google Meet, all while preserving the original speaker's tone and voice.

Initially, the service will be available in English and Spanish, with other languages to be added soon.

Google also amazed the audience by announcing "Beam" (formerly known as Starline).

This advanced remote conferencing system utilizes a combination of software and hardware, including six cameras and a specialized light field display, to create a three-dimensional representation of speakers as if they were in the same room, complete with millimeter-accurate head tracking and 60 frames per second video streaming.

Lyria RealTime: AI Composes Music

Google presented the Lyria RealTime model for interactive music creation. It is now available via the Gemini API. This model enables users to mix musical styles and control elements like pitch, tempo, and mood, producing audio at 48kHz quality.

NotebookLM: Visual and Auditory Presentation of Complex Information

NotebookLM's application received significant updates. Most notable is the Video Overviews feature, transforming documents and files into concise visual presentations.

Additionally, the "Audio Overviews" feature has been enhanced, allowing customization of audio summary durations, with full support for mobile applications

Wear OS 6 and Android Studio: Improvements for Developers and Users

Google launched the sixth version of Wear OS, introducing a unified design and new customization options.

As for Android Studio, it received new AI features, such as "Journeys" and "Agent Mode," and also includes enhancements for crash detection and intelligent code solution suggestions.

In conclusion, Google I/O 2025 revealed a clear direction. All aspects appear set to become AI-powered, spanning from translation and app development to music, search, and even daily communication.

Related Posts

Did DeepSeek Train Its AI on Gemini Data?
  • June 4, 2025

The AI world is once again stirred by controversy. This time, attention turns to Chinese AI lab DeepSeek and questions about…

Hume EVI 3: New Generation of AI Voices Rivals GPT-4o and Gemini Live
  • June 1, 2025

AI research company Hume AI has unveiled the third generation of its Empathic Voice Interface, known as EVI 3. The company…

Leave a Reply

Your email address will not be published. Required fields are marked *