LLMs
LLMs
Gemini 2.5 Pro Preview API
Google’s advanced AI model is designed for high-stakes reasoning, complex problem-solving, and state-of-the-art code generation.

1RPC.ai
Reasoning
Speed
$1.25
/
$10
Input/Output
1,000,000
Context Window
Gemini 2.5 Pro Preview
Gemini 2.5 Pro Preview was released for public preview on June 5, 2025, as an upgrade to the original 2.5 Pro (I/O edition) that launched in May.
This model brings higher benchmark scores and enhanced abilities for developers and enterprises, rivaling leading frontier models in accuracy, context length, and versatility.
What it’s optimized for
Gemini 2.5 Pro Preview is tailored for:
Complex reasoning and agentic workflows, including high-level coding and automation
Large-context tasks, with robust handling of multimodal (text, audio, image, video) data
Enterprise and developer use cases demanding accuracy, reliability, and fine-grained control over cost/latency
Rigorous coding and front-end/app development, including live, interactive web apps
High-volume, low-latency applications at production scale
Typical use cases
Gemini 2.5 Pro Preview is particularly effective for:
Automated code generation, editing, and large codebase transformations
Interactive web and app development using single-prompt workflows
Large-scale document, video, or audio understanding and summarization
AI agent orchestration and multi-turn tool use (e.g., with function calling, grounding, code execution)
Enterprise-scale summarization, search, categorization, and complex query answering across varied data types
Key characteristics
Top-tier reasoning while maintaining leadership on benchmarks (e.g., LMArena, WebDevArena) and excels at GPQA, HLE, and coding benchmarks
Multimodal native while handling text, code, audio, video, and image inputs; strong in native video understanding (84.8% VideoMME), vision, and audio tasks
Approximately 323 tokens/second output speed; built for high-volume, low-latency workloads
Supports function calling, code execution, structured outputs, and grounding with Google Search as a tool
Better creative writing, output formatting, and response structure, based on developer feedback
Enhanced following of user instructions and safety guardrails for reliable, predictable outputs
Fewer unnecessary rejections and improved handling of nuanced requests
Model architecture
Gemini 2.5 Pro Preview is built on Google’s advanced multimodal transformer architecture, leveraging mixture-of-experts routing and adaptive “thinking budgets” to manage cost, latency, and compute for different workflows.
The model natively supports multimodal reasoning, integrated tool use, and long-sequence context to suit demanding business, development, and research applications. This architecture powers both rapid-response and “extended thinking” scenarios to maximize performance and flexibility for developers.
Why choose 1RPC.ai for Gemini 2.5 Pro Preview
Every call is directly tied to the exact model and version used, ensuring traceability and trust in your outputs
Execution runs inside hardware-backed enclaves, so the relay can’t access or log your request
Connect to multiple AI providers through a single API
Avoid provider lock-in with simple, pay-per-prompt pricing
Privacy by design with our zero-tracking infrastructure that eliminates metadata leakage and protects your activity
Summary
Gemini 2.5 Pro Preview is Google’s flagship multimodal reasoning model as of mid-2025—a leader in accuracy, context, and coding—and a go-to choice for enterprises and developers needing the highest performance across complex, context-heavy, and agentic AI tasks.
Its blend of advanced reasoning, comprehensive tool integration, and full multimodal coverage enables use cases from live app development to large-scale real-world automation, all while maintaining cost-efficiency and high throughput.
Gemini 2.5 Pro Preview
Gemini 2.5 Pro Preview was released for public preview on June 5, 2025, as an upgrade to the original 2.5 Pro (I/O edition) that launched in May.
This model brings higher benchmark scores and enhanced abilities for developers and enterprises, rivaling leading frontier models in accuracy, context length, and versatility.
What it’s optimized for
Gemini 2.5 Pro Preview is tailored for:
Complex reasoning and agentic workflows, including high-level coding and automation
Large-context tasks, with robust handling of multimodal (text, audio, image, video) data
Enterprise and developer use cases demanding accuracy, reliability, and fine-grained control over cost/latency
Rigorous coding and front-end/app development, including live, interactive web apps
High-volume, low-latency applications at production scale
Typical use cases
Gemini 2.5 Pro Preview is particularly effective for:
Automated code generation, editing, and large codebase transformations
Interactive web and app development using single-prompt workflows
Large-scale document, video, or audio understanding and summarization
AI agent orchestration and multi-turn tool use (e.g., with function calling, grounding, code execution)
Enterprise-scale summarization, search, categorization, and complex query answering across varied data types
Key characteristics
Top-tier reasoning while maintaining leadership on benchmarks (e.g., LMArena, WebDevArena) and excels at GPQA, HLE, and coding benchmarks
Multimodal native while handling text, code, audio, video, and image inputs; strong in native video understanding (84.8% VideoMME), vision, and audio tasks
Approximately 323 tokens/second output speed; built for high-volume, low-latency workloads
Supports function calling, code execution, structured outputs, and grounding with Google Search as a tool
Better creative writing, output formatting, and response structure, based on developer feedback
Enhanced following of user instructions and safety guardrails for reliable, predictable outputs
Fewer unnecessary rejections and improved handling of nuanced requests
Model architecture
Gemini 2.5 Pro Preview is built on Google’s advanced multimodal transformer architecture, leveraging mixture-of-experts routing and adaptive “thinking budgets” to manage cost, latency, and compute for different workflows.
The model natively supports multimodal reasoning, integrated tool use, and long-sequence context to suit demanding business, development, and research applications. This architecture powers both rapid-response and “extended thinking” scenarios to maximize performance and flexibility for developers.
Why choose 1RPC.ai for Gemini 2.5 Pro Preview
Every call is directly tied to the exact model and version used, ensuring traceability and trust in your outputs
Execution runs inside hardware-backed enclaves, so the relay can’t access or log your request
Connect to multiple AI providers through a single API
Avoid provider lock-in with simple, pay-per-prompt pricing
Privacy by design with our zero-tracking infrastructure that eliminates metadata leakage and protects your activity
Summary
Gemini 2.5 Pro Preview is Google’s flagship multimodal reasoning model as of mid-2025—a leader in accuracy, context, and coding—and a go-to choice for enterprises and developers needing the highest performance across complex, context-heavy, and agentic AI tasks.
Its blend of advanced reasoning, comprehensive tool integration, and full multimodal coverage enables use cases from live app development to large-scale real-world automation, all while maintaining cost-efficiency and high throughput.
Gemini 2.5 Pro Preview
Gemini 2.5 Pro Preview was released for public preview on June 5, 2025, as an upgrade to the original 2.5 Pro (I/O edition) that launched in May.
This model brings higher benchmark scores and enhanced abilities for developers and enterprises, rivaling leading frontier models in accuracy, context length, and versatility.
What it’s optimized for
Gemini 2.5 Pro Preview is tailored for:
Complex reasoning and agentic workflows, including high-level coding and automation
Large-context tasks, with robust handling of multimodal (text, audio, image, video) data
Enterprise and developer use cases demanding accuracy, reliability, and fine-grained control over cost/latency
Rigorous coding and front-end/app development, including live, interactive web apps
High-volume, low-latency applications at production scale
Typical use cases
Gemini 2.5 Pro Preview is particularly effective for:
Automated code generation, editing, and large codebase transformations
Interactive web and app development using single-prompt workflows
Large-scale document, video, or audio understanding and summarization
AI agent orchestration and multi-turn tool use (e.g., with function calling, grounding, code execution)
Enterprise-scale summarization, search, categorization, and complex query answering across varied data types
Key characteristics
Top-tier reasoning while maintaining leadership on benchmarks (e.g., LMArena, WebDevArena) and excels at GPQA, HLE, and coding benchmarks
Multimodal native while handling text, code, audio, video, and image inputs; strong in native video understanding (84.8% VideoMME), vision, and audio tasks
Approximately 323 tokens/second output speed; built for high-volume, low-latency workloads
Supports function calling, code execution, structured outputs, and grounding with Google Search as a tool
Better creative writing, output formatting, and response structure, based on developer feedback
Enhanced following of user instructions and safety guardrails for reliable, predictable outputs
Fewer unnecessary rejections and improved handling of nuanced requests
Model architecture
Gemini 2.5 Pro Preview is built on Google’s advanced multimodal transformer architecture, leveraging mixture-of-experts routing and adaptive “thinking budgets” to manage cost, latency, and compute for different workflows.
The model natively supports multimodal reasoning, integrated tool use, and long-sequence context to suit demanding business, development, and research applications. This architecture powers both rapid-response and “extended thinking” scenarios to maximize performance and flexibility for developers.
Why choose 1RPC.ai for Gemini 2.5 Pro Preview
Every call is directly tied to the exact model and version used, ensuring traceability and trust in your outputs
Execution runs inside hardware-backed enclaves, so the relay can’t access or log your request
Connect to multiple AI providers through a single API
Avoid provider lock-in with simple, pay-per-prompt pricing
Privacy by design with our zero-tracking infrastructure that eliminates metadata leakage and protects your activity
Summary
Gemini 2.5 Pro Preview is Google’s flagship multimodal reasoning model as of mid-2025—a leader in accuracy, context, and coding—and a go-to choice for enterprises and developers needing the highest performance across complex, context-heavy, and agentic AI tasks.
Its blend of advanced reasoning, comprehensive tool integration, and full multimodal coverage enables use cases from live app development to large-scale real-world automation, all while maintaining cost-efficiency and high throughput.
Like this article? Share it.
Implement
Implement
Get started with an API-friendly relay
Send your first request to verified LLMs with a single code snippet.
import requests
import json
response = requests.post(
url="https://1rpc.ai/v1/chat/completions",
headers={
"Authorization": "Bearer <1RPC_AI_API_KEY>",
"Content-type": "application/json",
},
data=json.dumps ({
"model": "gemini-2.5-pro-preview-06-05",
"messages": [
{
"role": "user",
"content": "What is the meaning of life?"
}
]
})
)Copy and go
Copied!
import requests
import json
response = requests.post(
url="https://1rpc.ai/v1/chat/completions",
headers={
"Authorization": "Bearer <1RPC_AI_API_KEY>",
"Content-type": "application/json",
},
data=json.dumps ({
"model": "gemini-2.5-pro-preview-06-05",
"messages": [
{
"role": "user",
"content": "What is the meaning of life?"
}
]
})
)Copy and go
Copied!
Pricing
Pricing
Estimate Usage Across Any AI Model
Adjust input and output size to estimate token usage and costs.
Token Calculator for Gemini 2.5 Pro Preview
Input (100)
Output (1000 )
$0.0101
Total cost per million tokens