Google has launched its new generative AI model ‘Gemini 2.0 Flash’ along with various agent prototypes which utilise its technology.
- OpenAI Launches Google And Microsoft Bing Rival ChatGPT Search
- Honeywell and Google Cloud Deploy AI Agents for Industrial Sector
- Google Signs Nuclear Energy Deal to Power its AI Data Centres
Gemini 2.0 Flash is the successor to 1.5 Flash, Google’s most popular model for developers to date. Most notably, 2.0 Flash is faster than its predecessor and runs at twice the speed of the 1.5 Pro.
Described by the tech giant as an early experimental version of Gemini 2.0 for developers, 2.0 Flash is now available via the Gemini API in Google AI and Vertex AI.
Google and Alphabet CEO Sundar Pichai, introduced Gemini 2.0 via a company blog post: “Over the last year, we have been investing in developing more agentic models, meaning they can understand more about the world around you, think multiple steps ahead, and take action on your behalf, with your supervision.
“Today we’re excited to launch our next era of models built for this new agentic era: introducing Gemini 2.0, our most capable model yet.
“With new advances in multimodality — like native image and audio output — and native tool use, it will enable us to build new AI agents that bring us closer to our vision of a universal assistant.
Gemini 2.0 Flash
As well as new multimodal inputs, such as video, audio, and images, 2.0 Flash also enables multimodal output through image generation, plus text and steerable text-to-speech (TTS) multilingual audio.
It is also able to natively leverage Google Search, code execution and third-party functions.
Google is also releasing a new Multimodal Live API with real-time audio, video-streaming input, and more, to help developers create their applications.
Gemini users will also be able to access a version of 2.0 Flash that has been optimised for chat via the model drop-down on both desktop and mobile.
New Prototypes
Google has been researching the application of AI agents with a series of prototypes to assist with tasks.
To this end, it has updated to Project Astra, a protype delving into the potential capabilities of a universal AI assistant.
Project Mariner will look into the future of human-agent interaction, beginning with a focus on the browsers.
There is also Jules, an AI-powered code agent designed to solve issues, and generate and execute plans on behalf of developers
Pichai helped to define Gemini 2.0: “2.0’s advances are underpinned by decade-long investments in our differentiated full-stack approach to AI innovation.
“It’s built on custom hardware like Trillium, our sixth-generation TPUs. TPUs powered 100% of Gemini 2.0 training and inference, and today Trillium is generally available to customers so they can build with it too.
“If Gemini 1.0 was about organising and understanding information, Gemini 2.0 is about making it much more useful. I can’t wait to see what this next era brings.”
So far, Google’s CEO shared that its first models, Gemini 1.0 and 1.5, are being used by millions of developers. They have also prompted Google itself to review and renew all seven of its core product offerings which collectively serve two billion users.