Google: Gemini 2.0 Flash-Lite
Model Overview
| Property | Value |
|---|---|
| Model ID | google/gemini-2.0-flash-lite |
| Name | Gemini 2.0 Flash-Lite |
| Provider | |
| Released | December 2024 |
Description
Streamlined and ultra-efficient model designed for simple, high-frequency tasks. Gemini 2.0 Flash-Lite prioritizes speed and affordability while maintaining essential multimodal capabilities.
Specifications
| Spec | Value |
|---|---|
| Context Window | 1,000,000 tokens |
| Max Completion | 8,192 tokens |
| Modalities | Text, Image |
Pricing
| Type | Price |
|---|---|
| Input | $0.07 per 1M tokens |
| Output | $0.30 per 1M tokens |
Capabilities
- Vision: Yes
- Tool Use: Yes
- JSON Mode: Yes
- Streaming: Yes
- Function Calling: Yes
- High Throughput: Yes
Use Cases
High-volume inference workloads, cost-sensitive deployments, simple classification and extraction tasks, real-time applications.
Best For
- High-frequency API calls
- Cost-optimized batch processing
- Simple text classification
- Quick data extraction
Integration with LangMart
Gateway Support: Type 2 (Cloud), Type 3 (Self-hosted)
API Usage:
curl -X POST https://api.langmart.ai/v1/chat/completions \
-H "Authorization: Bearer sk-your-api-key" \
-H "Content-Type: application/json" \
-d '{
"model": "google/gemini-2.0-flash-lite",
"messages": [{"role": "user", "content": "Hello"}],
"max_tokens": 2048
}'
Related Models
- google/gemini-2.0-flash - Full-featured Flash model
- google/gemini-2.5-flash-lite - Next generation lite
- google/gemini-2.0-flash-lite-preview - Preview version
Last Updated: December 28, 2025