Skip to main content
Google DeepMindReleased February 25, 2025

Gemini 2.0 Flash Lite

#22 Arena RankUndisclosed parameters

Gemini 2.0 Flash Lite holds a solid spot in the Arena rankings at #22. Context window: 0.001K tokens.

Context

1M

Input

$0.075

Key Specifications

🏆

Arena Rank

#22

📐

Context Window

1M

📥

Input Price

per 1M tokens

$0.075

📤

Output Price

per 1M tokens

$0.30

🧠

Parameters

Undisclosed

🔒

Open Source

No

Best For

High-volumelow-cost tasks

About Gemini 2.0 Flash Lite

Gemini 2.0 Flash Lite, developed by Google DeepMind, is the most affordable model in Google's lineup with a 1 million token context window. The model targets extremely high-volume applications where cost minimization is the primary constraint, handling classification, content filtering, routing, and basic summarization tasks competently. At $0.075 per million input tokens and $0.30 per million output tokens, it ranks among the cheapest API-accessible models from any major AI provider. Despite its budget positioning, Flash Lite inherits the massive context window from the Gemini architecture, enabling long-document processing at minimal cost. Gemini 2.0 Flash Lite ranks #22 on the Chatbot Arena leaderboard, demonstrating adequate quality for production workloads that prioritize throughput and cost-efficiency over maximum capability.

Pricing per 1M tokens

Input Tokens

$0.075

Output Tokens

$0.30

Frequently Asked Questions

What is Gemini 2.0 Flash Lite?
Gemini 2.0 Flash Lite, developed by Google DeepMind, is the most affordable model in Google's lineup with a 1 million token context window. The model targets extremely high-volume applications where cost minimization is the primary constraint, handling classification, content filtering, routing, and basic summarization tasks competently. At $0.075 per million input tokens and $0.30 per million output tokens, it ranks among the cheapest API-accessible models from any major AI provider. Despite its budget positioning, Flash Lite inherits the massive context window from the Gemini architecture, enabling long-document processing at minimal cost. Gemini 2.0 Flash Lite ranks #22 on the Chatbot Arena leaderboard, demonstrating adequate quality for production workloads that prioritize throughput and cost-efficiency over maximum capability.
How much does Gemini 2.0 Flash Lite cost?
Gemini 2.0 Flash Lite costs $0.075 per 1M input tokens and $0.30 per 1M output tokens. You pay only for what you use, which keeps costs predictable.
What is Gemini 2.0 Flash Lite's context window?
Gemini 2.0 Flash Lite has a context window of 1M tokens. This determines how much text the model can process in a single request — bigger windows mean longer documents and richer conversation history.
Is Gemini 2.0 Flash Lite open source?
No, Gemini 2.0 Flash Lite is a proprietary model available through Google DeepMind's API. You get managed infrastructure, regular updates, and support as part of the package.
What is Gemini 2.0 Flash Lite best for?
Gemini 2.0 Flash Lite is best suited for: High-volume, low-cost tasks. These use cases play to the model's strengths in capability, speed, and cost within Google DeepMind's lineup.