![Gemini 2.5 Flash]()
Google has announced the release of updated preview versions of Gemini 2.5 Flash and Gemini 2.5 Flash-Lite, now available on Google AI Studio and Vertex AI. These updates focus on delivering higher-quality responses while reducing latency and cost—making them especially attractive for developers building high-throughput and agentic applications.
Key Highlights
Performance and Efficiency
Gemini 2.5 Flash-Lite Updates
Better Instruction Following: Handles complex prompts and system instructions with greater precision.
Reduced Verbosity: Generates concise responses, lowering token usage and latency.
Enhanced Multimodal & Translation: Improved audio transcription, image understanding, and translation accuracy.
Gemini 2.5 Flash Updates
Smarter Tool Use: Significant improvements in agentic applications and multi-step reasoning. Achieved a 5% gain on SWE-Bench Verified benchmark (48.9% ? 54%).
Higher Cost Efficiency: Produces higher-quality outputs while consuming fewer tokens, lowering costs with "thinking" enabled.
Early testers are already seeing strong results. Yichao "Peak" Ji, Co-Founder & Chief Scientist at Manus, highlighted:
“The new Gemini 2.5 Flash model offers a remarkable blend of speed and intelligence. Our evaluation on internal benchmarks revealed a 15% leap in performance for long-horizon agentic tasks. Its outstanding cost-efficiency enables Manus to scale to unprecedented levels—advancing our mission to Extend Human Reach.”
Easier Access with -latest
Aliases
To simplify adoption, Google has introduced -latest
aliases for each model family:
gemini-flash-latest
gemini-flash-lite-latest
These aliases always point to the newest preview versions, letting developers experiment with new features without constantly updating model strings. A 2-week notice will be provided before updates or deprecations.
For production stability, developers should continue using:
gemini-2.5-flash
gemini-2.5-flash-lite