Gemini 1.5 Flash
LegacyGemini 1.5 Flash targets low-latency, cost-efficient multimodal chat and retrieval workloads on the Gemini API and Vertex AI.
Best for:Mobile assistantsCost tier:Flash
Compared to:Gemini 2.0 FlashReplaces:Gemini 1.5 Pro
Multimodal LLM · Release — · See vendor
googlelatencycost
Newer version: Gemini 2.0 Flash
Updated 1 day ago · Verified Apr 2026 · Score 78
Decision summary
Why teams reach for it, where it fits, and what to watch for — before you dive into specs.
Why teams choose it
- Flash vs Pro tradeoffs are task-specific—measure on your eval suite.
- Regional latency varies—instrument client-side.
Best use cases
- Use this when mobile assistants
- Use this when high-QPS customer support
Tradeoffs
- Quality gap vs Pro on hardest reasoning.
- Quota and preview features change frequently.
Technical details
Modalities, benchmarks, and release context.
Modalities
What goes in and what comes out.
- Inputs
- text, image, audio, video
- Outputs
- text
- Capabilities
- long context, multimodal, latency
Release: —·License: See vendor
Benchmarks snapshot
Structured JSON for reproducible comparisons.
No benchmark data yet — see comparisons for relative performance.
Family lineup
Explore other versions in this family after you have the headline on this model.
Current family lineup
Previous versions
Continue exploring
A short set of comparisons, nearby models, and links to go deeper — without repeating the same paths.
Compare with
Learn & build
Tools and curated destinations (max four).