Technical specifications — Gemini 3.1 Pro
| Item | gemini-3-pro (public summary) |
|---|---|
| Provider | |
| Canonical model id | gemini-3-pro (public preview) |
| Input types | Text, Image, Video, Audio, PDF |
| Output types | Text (natural language, structured outputs, function-call payloads) |
| Input token limit (context) | 1,048,576 tokens |
| Output token limit | 65,536 tokens |
| Function-calling / tool use | Supported (function calling, structured outputs, tool integrations) |
| Multimodality | Full multimodal support (images, video, audio, documents) |
| Code execution & agentic flows | Supported (agent mode, code assist, tool orchestration) |
| Knowledge cutoff | January 2025 |
What is Gemini 3.1 Pro ?
Gemini 3.1 Pro is Google’s publicly flagship in the Gemini 3 family, positioned as a state-of-the-art multimodal reasoning model with advanced agentic and developer tooling. The model emphasizes high-capacity context handling (over 1M token inputs), broad media support (images, video, audio, PDF), and deep integrations for tool use, function calling, and code-centric workflows (e.g., Gemini Code Assist and agent modes).
Gemini 3 Pro is presented by Google as optimized for both interactive developer experiences (low-latency coding and agent workflows) and high-fidelity multimodal understanding (interpreting and reasoning across mixed media inputs).
Main Features of Gemini 3.1 Pro
Gemini-3.1 Pro (via its Preview) introduces the following features:
Multimodal Integration
Processes inputs across:
- Natural language
- Images
- Speech/audio
- Video
with a unified token representation for cross-modal reasoning.
Extended Context Window
An exceptionally large context capacity of up to ~1 million tokens enables handling of:
- Long documents
- Multidocument synthesis
- Codebases and transcripts.
This surpasses many competing models that typically support ~32 K–262 K tokens.
Sparse Mixture-of-Experts (MoE) Scaling
Sparse MoE routing allows scaling internal model capacity without proportional compute costs, improving reasoning at scale.
Advanced Reasoning / Planning
Innovations like chain-of-thought training, reinforcement learning from human feedback, and specialized benchmarks make it strong on logical and mathematical tasks.
Supposed benchmarks:
AIME 2025: 100% (with code execution)
SWE-Bench Verified: 83.9%
ARC-AGI-2: 71.8%
LiveCodeBench Pro: 2844 Elo
Terminal-Bench 2.0: 63.5%
MMMLU: 93.6%
Representative enterprise use cases
- End-to-end media pipelines: Ingest video, transcript, and images to produce synchronized summaries, metadata, and structured insight at scale.
- Large-scale code generation and review: Use in IDEs and CI pipelines to auto-generate code, refactor multi-file projects, and produce test suggestions across large codebases.
- Agentic automation: Coordinate multi-tool agents that interact with cloud services, orchestration systems, and internal APIs using structured function calls.
- Research & content production: Draft long-form content (reports, books) that combine text and embedded multimedia with internal cross-references preserved.
How to access Gemini 3.1 Pro API
Step 1: Sign Up for API Key
Log in to cometapi.com. If you are not our user yet, please register first. Sign into your CometAPI console. Get the access credential API key of the interface. Click “Add Token” at the API token in the personal center, get the token key: sk-xxxxx and submit.
Step 2: Send Requests to Gemini 3.1 Pro API
Select the “gemini-3.1-pro” endpoint to send the API request and set the request body. The request method and request body are obtained from our website API doc. Our website also provides Apifox test for your convenience. Replace <YOUR_API_KEY> with your actual CometAPI key from your account. base url is Gemini Generating Content and Chat.
Insert your question or request into the content field—this is what the model will respond to . Process the API response to get the generated answer.
Step 3: Retrieve and Verify Results
Process the API response to get the generated answer. After processing, the API responds with the task status and output data.
See also Gemini 3 Pro API