curl --request GET \
--url https://api.venice.ai/api/v1/models \
--header 'Authorization: Bearer <token>'{
"data": [
{
"created": 1727966436,
"id": "llama-3.2-3b",
"model_spec": {
"availableContextTokens": 131072,
"capabilities": {
"optimizedForCode": false,
"quantization": "fp16",
"supportsFunctionCalling": true,
"supportsReasoning": false,
"supportsResponseSchema": true,
"supportsVision": false,
"supportsWebSearch": true,
"supportsLogProbs": true
},
"constraints": {
"temperature": {
"default": 0.8
},
"top_p": {
"default": 0.9
}
},
"name": "Llama 3.2 3B",
"modelSource": "https://huggingface.co/meta-llama/Llama-3.2-3B",
"offline": false,
"pricing": {
"input": {
"usd": 0.15,
"diem": 0.15
},
"output": {
"usd": 0.6,
"diem": 0.6
}
},
"traits": [
"fastest"
]
},
"object": "model",
"owned_by": "venice.ai",
"type": "text"
}
],
"object": "list",
"type": "text"
}Returns a list of available models supported by the Venice.ai API for both text and image inference.
curl --request GET \
--url https://api.venice.ai/api/v1/models \
--header 'Authorization: Bearer <token>'{
"data": [
{
"created": 1727966436,
"id": "llama-3.2-3b",
"model_spec": {
"availableContextTokens": 131072,
"capabilities": {
"optimizedForCode": false,
"quantization": "fp16",
"supportsFunctionCalling": true,
"supportsReasoning": false,
"supportsResponseSchema": true,
"supportsVision": false,
"supportsWebSearch": true,
"supportsLogProbs": true
},
"constraints": {
"temperature": {
"default": 0.8
},
"top_p": {
"default": 0.9
}
},
"name": "Llama 3.2 3B",
"modelSource": "https://huggingface.co/meta-llama/Llama-3.2-3B",
"offline": false,
"pricing": {
"input": {
"usd": 0.15,
"diem": 0.15
},
"output": {
"usd": 0.6,
"diem": 0.6
}
},
"traits": [
"fastest"
]
},
"object": "model",
"owned_by": "venice.ai",
"type": "text"
}
],
"object": "list",
"type": "text"
}Bearer authentication header of the form Bearer <token>, where <token> is your auth token.
Filter models by type. Use "all" to get all model types.
asr, embedding, image, text, tts, upscale, inpaint, video "text"
OK
List of available models
Show child attributes
Model ID
"venice-uncensored"
Show child attributes
The context length supported by the model. Only applicable for text models.
32768
Is this model in beta?
false
The privacy mode of the model. Private models have zero data retention. Anonymized models Venice can not guarantee privacy on, but requests are not affiliated with a user
private, anonymized "private"
Text model specific capabilities.
Show child attributes
Is the LLM optimized for coding?
true
The quantization type of the running model.
fp4, fp8, fp16, bf16, int8, int4, not-available "fp8"
Does the LLM model support function calling?
true
Does the model support reasoning with
true
Does the LLM model support response schema? Only models that support function calling can support response_schema.
true
Does the LLM support vision?
true
Does the LLM model support web search?
true
Does the LLM model support logprobs parameter?
true
Constraints that apply to this model.
Show child attributes
The maximum supported prompt length.
2048
The requested width and height of the image generation must be divisible by this value.
8
The name of the model.
"Venice Uncensored 1.1"
The source of the model, such as a URL to the model repository.
"https://huggingface.co/cognitivecomputations/Dolphin-Mistral-24B-Venice-Edition"
Is this model presently offline?
false
Pricing details for the model
Show child attributes
Traits that apply to this model. You can specify a trait to auto-select a model vs. specifying the model ID in your request to avoid breakage as Venice updates and iterates on its models.
["default_code"]The voices available for this TTS model. Only applicable for TTS models.
[
"af_alloy",
"af_aoede",
"af_bella",
"af_heart",
"af_jadzia"
]Object type
model "model"
Who runs the model
venice.ai "venice.ai"
Model type
asr, embedding, image, text, tts, upscale, inpaint, video "text"
Release date on Venice API
1699000000
list Type of models returned.
asr, embedding, image, text, tts, upscale, inpaint, video "text"