Google has launched the experimental Gemini 2.0 Flash, the first in its Gemini 2.0 family of AI models.
This version delivers groundbreaking advancements in performance and functionality, building on the success of its predecessor, Gemini 1.5 Flash.
Why it matters
With twice the speed of the Gemini 1.5 Pro, the 2.0 Flash model introduces enhanced multimodal capabilities. It processes and outputs text, images, video, and steerable multilingual text-to-speech audio.
Developers now gain access to advanced tools like native image generation, real-time API integration, and support for complex task execution.
Key features
- Multimodal Inputs and Outputs: Processes and generates a combination of text, images, audio, and video.
- Tool Integration: Enables seamless interaction with tools like Google Search, code execution, and custom user-defined functions.
- Developer Accessibility: Available through the Gemini API on Google AI Studio and Vertex AI, with early-access features for text-to-speech and image generation.
What’s next
Starting today, Gemini users globally can explore the model in the AI assistant platform via the web, with app support coming soon. General availability of Gemini 2.0 Flash is set for January 2025, alongside expanded integration into other Google products.
Exploring the future
Google is pushing the boundaries of AI-assisted experiences with prototypes like:
- Project Astra: A universal AI assistant.
- Project Mariner: Innovations in human-agent interactions through browsers.
- Jules: A code-focused AI agent for developers.
Google aims to refine these tools with input from early testers, paving the way for broader adoption and practical AI applications in daily tasks.