Requesty
Coding API@us-east5

gemini-2.5-flash

Google's first hybrid reasoning model which supports a 1M token context window and has thinking budgets. Most balanced Gemini model, optimized for low latency use cases.

👁Vision🧠Reasoning🔧Tool callingCaching

Pricing per 1M tokens

Input
$0.30
Output
$2.50
Cache write
$0.55
Cache read
$0.07

Specifications

Context window1.0M tokens
Max output66K tokens
API typechat
AddedMay 20, 2025
Model IDcoding/gemini-2.5-flash@us-east5

Privacy & data

Data retentionNo
Used for trainingNo
Provider location🌍 Global