Venice APIs
- API Spec
- Rate Limits
- Error Codes
- Chat
- Images
- Audio
- Models
- API Keys
- API Key Rate Limits
- Characters
List Models
Returns a list of available models supported by the Venice.ai API for both text and image inference.
curl --request GET \
--url https://api.venice.ai/api/v1/models \
--header 'Authorization: Bearer <token>'
{
"object": "list",
"type": "text",
"data": [
{
"id": "llama-3.3-70b",
"type": "text",
"object": "model",
"created": 1699000000,
"owned_by": "venice.ai",
"model_spec": {
"availableContextTokens": 65536,
"constraints": {
"promptCharacterLimit": 2048,
"widthHeightDivisor": 8,
"steps": {
"default": 25,
"max": 50
}
},
"capabilities": {
"optimizedForCode": true,
"supportsFunctionCalling": true,
"supportsResponseSchema": true,
"supportsWebSearch": true,
"supportsReasoning": true,
"supportsVision": true
},
"traits": [
"default_code"
],
"voices": [
"af_alloy",
"af_aoede",
"af_bella",
"af_heart",
"af_jadzia"
],
"modelSource": "https://huggingface.co/meta-llama/Llama-3.3-70B-Instruct",
"beta": false,
"offline": false
}
}
]
}
Postman Collection
For additional examples, please see this Postman Collection.
Authorizations
Bearer authentication header of the form Bearer <token>
, where <token>
is your auth token.
Query Parameters
Filter models by type. Use "all" to get all model types.
image
, text
, embedding
, tts
"text"
Response
list
Type of models returned.
image
, text
, embedding
, tts
"text"
List of available models
Model ID
"llama-3.3-70b"
Model type
image
, text
, embedding
, tts
"text"
Object type
model
"model"
Who runs the model
venice.ai
"venice.ai"
The context length supported by the model. Only applicable for text models.
65536
Constraints that apply to this model.
The maximum supported prompt length.
2048
The requested width and height of the image generation must be divisible by this value.
8
Text model specific capabilities.
Is the LLM optimized for coding?
true
Does the LLM model support function calling?
true
Does the LLM model support response schema? Only models that support function calling can support response_schema.
true
Does the LLM model support web search?
true
Does the model support reasoning with <thinking> blocks of output.
true
Does the LLM support vision?
true
Traits that apply to this model. You can specify a trait to auto-select a model vs. specifying the model ID in your request to avoid breakage as Venice updates and iterates on its models.
["default_code"]
The voices available for this TTS model. Only applicable for TTS models.
[
"af_alloy",
"af_aoede",
"af_bella",
"af_heart",
"af_jadzia"
]
The source of the model, such as a URL to the model repository.
"https://huggingface.co/meta-llama/Llama-3.3-70B-Instruct"
Is this model in beta?
false
Is this model presently offline?
false
Release date on Venice API
1699000000
curl --request GET \
--url https://api.venice.ai/api/v1/models \
--header 'Authorization: Bearer <token>'
{
"object": "list",
"type": "text",
"data": [
{
"id": "llama-3.3-70b",
"type": "text",
"object": "model",
"created": 1699000000,
"owned_by": "venice.ai",
"model_spec": {
"availableContextTokens": 65536,
"constraints": {
"promptCharacterLimit": 2048,
"widthHeightDivisor": 8,
"steps": {
"default": 25,
"max": 50
}
},
"capabilities": {
"optimizedForCode": true,
"supportsFunctionCalling": true,
"supportsResponseSchema": true,
"supportsWebSearch": true,
"supportsReasoning": true,
"supportsVision": true
},
"traits": [
"default_code"
],
"voices": [
"af_alloy",
"af_aoede",
"af_bella",
"af_heart",
"af_jadzia"
],
"modelSource": "https://huggingface.co/meta-llama/Llama-3.3-70B-Instruct",
"beta": false,
"offline": false
}
}
]
}