A new AI model for the agent age

A new AI model for the agent age

A note from Sundar Pichai, CEO of Google and Alphabet:

Information is at the core of human progress. That’s why, for more than 26 years, we’ve focused on our mission of organizing the world’s information and making it accessible and useful. And that’s why we continue to push the boundaries of AI to organize this information across all inputs and make it accessible across all outputs so it can be truly useful to you.

That was our vision when we introduced Gemini 1.0 last December. Gemini 1.0 and 1.5, the first model designed for native multimodality, brought great advances with multimodality and long context to understand and process much more of information in text, video, images, audio and code.

Now millions of developers are building with Gemini. And it helps us reimagine and develop all of our products – including all seven with 2 billion users. NotebookLM is a great example of what multimodality and long context can enable people and why it is loved by so many.

Over the last year, we’ve invested in building more responsive models to better understand the world around you, think multiple steps ahead, and take action on your behalf, under your supervision.

Today we’re excited to introduce our next model era for this new era of agents: the introduction of Gemini 2.0, our most powerful model yet. With new advances in multimodality – such as native image and audio output – and native tool usage, it will allow us to develop new AI agents that bring us closer to our vision of a universal assistant.

Today we are putting 2.0 into the hands of developers and trusted testers. And we’re working quickly to integrate it into our products, most notably Gemini and Search. Starting today, our experimental Gemini 2.0 Flash model is available to all Gemini users. We’re also introducing a new feature called Deep Research, which uses advanced reasoning and long-term context capabilities to act as a research assistant, investigate complex topics, and create reports on your behalf. It is available today in Gemini Advanced.

No product has been more transformed by AI than search. Our AI insights now reach 1 billion people, enabling them to ask entirely new types of questions – quickly becoming one of our most popular search features ever. As a next step, we’re bringing Gemini 2.0’s advanced reasoning capabilities to AI Overviews to tackle more complex topics and multi-step questions, including advanced mathematical equations, multimodal query, and coding. We began limited testing this week and will roll it out more broadly early next year. And over the next year, we will continue to roll out AI Overviews to additional countries and languages.

The advancements of 2.0 are underpinned by decades of investment in our differentiated, full-stack approach to AI innovation. It’s built on custom hardware like Trillium, our sixth-generation TPUs. TPUs supported 100% of Gemini 2.0 training and inference, and today Trillium is generally available to customers so they can build with it too.

While Gemini 1.0 was about organizing and understanding information, Gemini 2.0 is about making it much more useful. I can’t wait to see what this next era brings.

-Sundar


Leave a Reply

Your email address will not be published. Required fields are marked *