Supported Models
gpt-4.1
gpt-4o
gemini-2.5-pro
gemini-2.5-flash
deepseek-r1
deepseek-v3-0324
How to Set Models
You don’t need to specify amodel-name
in your API calls. The model for your agent is managed on the backend, so switching models won’t require any code changes.
To change the model for your agent, go to the Agent Settings page in the Vivgrid Console.
Pricing
Pricing is calculated in USD per 1 million tokens. The table below details the cost for input, cached, and output tokens for each model.Model | Input Token | Cached Token | Output Token |
---|---|---|---|
gpt-4.1 | $2.00 | $0.50 | $8.00 |
gpt-4o | $2.50 | $1.25 | $10.00 |
gemini-2.5-pro | $1.25 | $0.31 | $10.00 |
gemini-2.5-flash | $0.30 | $0.08 | $2.50 |
deepseek-r1 | $1.35 | - | $5.40 |
deepseek-v3-0324 | $1.14 | - | $4.56 |
Capabilities
Model | Context Window | Max Output Tokens | Function Calling Support |
---|---|---|---|
gpt-4.1 | 1M tokens | 32K | Yes |
gpt-4o | 128K | 16K | Yes |
gemini-2.5-pro | 1M | 64K | Yes |
gemini-2.5-flash | 1M | 64K | Yes |
deepseek-r1 | 64K | 8K | No |
deepseek-v3-0324 | 128K | 64K | No |
Service Regions
Vivgrid models are globally distributed to reduce latency and support data residency. API requests are routed to the nearest region, and Function Calling Tools are automatically deployed across all regions. The table below outlines the regional availability for each model:Model | Available Regions |
---|---|
gpt-4.1 | US, EU, APAC |
gpt-4o | US, EU, APAC |
gemini-2.5-pro | Global |
gemini-2.5-flash | Global |
deepseek-r1 | APAC |
deepseek-v3-0324 | APAC |