We designed Gemini 2.5 to be a family of hybrid reasoning models that provide amazing performance, while also being on the Pareto Frontier of cost and speed. Today, we’re taking the following step with our 2.5 Pro and Flash models by releasing them as stable and usually available. And we’re bringing you 2.5 Flash-Lite in preview — our most cost-efficient and fastest 2.5 model yet.
Making 2.5 Flash and a couple of.5 Pro generally available
Due to your whole feedback, today we’re releasing stable versions of two.5 Flash and Pro, so you possibly can construct production applications with confidence. Developers like Spline and Rooms and organizations like Snap and SmartBear have already been using the most recent versions in-production for the previous few weeks.
Introducing Gemini 2.5 Flash-Lite
We’re also introducing a preview of the brand new Gemini 2.5 Flash-Lite, our most cost-efficient and fastest 2.5 model yet. You may start constructing with the preview version now, and we’re looking forward to your feedback.
2.5 Flash-Lite has all-around higher quality than 2.0 Flash-Lite on coding, math, science, reasoning and multimodal benchmarks. It excels at high-volume, latency-sensitive tasks like translation and classification, with lower latency than 2.0 Flash-Lite and a couple of.0 Flash on a broad sample of prompts. It comes with the identical capabilities that make Gemini 2.5 helpful, including the power to show considering on at different budgets, connecting to tools like Google Search and code execution, multimodal input, and a 1 million-token context length.
See more details about our 2.5 family of models in the most recent Gemini technical report.
