Quickstart
1
Get your API key
Head to your Venice API Settings and generate a new API key.For a detailed walkthrough with screenshots, check out the API Key guide.
2
Set up your API key
Add your API key to your environment. You can export it in your shell:Or add it to a
.env file in your project:3
Install the SDK
Venice is OpenAI-compatible, so you can use the OpenAI SDK. If you prefer to use cURL or raw HTTP requests, you can skip this step.
4
Send your first request
system- Instructions for how the model should behaveuser- Your prompts or questionsassistant- Previous model responses (for multi-turn conversations)tool- Function calling results (when using tools)
5
Choose your model (optional)
Venice has multiple models for different use cases. Popular choices:
llama-3.3-70b- Balanced performance, great for most use casesqwen3-235b- Most powerful flagship model for complex tasksmistral-31-24b- Vision + function calling supportvenice-uncensored- No content filtering
View All Models
Browse the complete list of models with pricing, capabilities, and context limits
6
Use Venice Parameters
You can choose to enable Venice-specific features like web search using See all available parameters.
venice_parameters:7
Enable streaming (optional)
Stream responses in real-time using
stream=True:8
Customize response behavior (optional)
Control how the model responds with parameters like temperature, max tokens, and more:Check out the Chat Completions docs for more information on all supported parameters.
More Capabilities
Image Generation
Create images from text prompts using diffusion models:images array. Decode the base64 string to save or display the image.
Popular Image Models:
qwen-image- Highest quality image generationvenice-sd35- Default choice, works with all featureshidream- Fast generation for production use
View All Image Models
See all available image models with pricing and capabilities
cfg_scale, negative_prompt, style_preset, seed, variants, and more, check out the Images API Reference.
Image Editing
Modify existing images with AI-powered inpainting using the Qwen-Image model:Image Upscaling
Enhance and upscale images to higher resolutions:Text-to-Speech
Convert text to audio with 60+ multilingual voices:tts-kokoro model supports 60+ multilingual voices including af_sky, af_nova, am_liam, bf_emma, zf_xiaobei, and jm_kumo.
See the TTS API for all voice options.
Embeddings
Generate vector embeddings for semantic search, RAG, and recommendations:Vision (Multimodal)
Analyze images alongside text using vision-capable models likemistral-31-24b:
Function Calling
Define functions that models can call to interact with external tools and APIs:llama-3.3-70b, qwen3-235b, mistral-31-24b, qwen3-4b
Next Steps
Now that you’ve made your first requests, explore more of what Venice API has to offer:Browse Models
Compare all available models with their capabilities, pricing, and context limits
API Reference
Explore detailed API documentation with all endpoints and parameters
Structured Responses
Learn how to get JSON responses with guaranteed schemas
AI Agents Guide
Build autonomous AI agents with Venice API and frameworks like Eliza
Additional Resources
Rate Limiting
Understand rate limits and best practices for production usage
Error Codes
Reference for handling API errors and troubleshooting issues
Postman Collection
Import our complete Postman collection for easy testing
Privacy & Security
Learn about Venice’s privacy-first architecture and data handling
Need Help?
- Discord Community: Join our Discord server for support and discussions
- Documentation: Browse our complete API reference
- Status Page: Check service status at veniceai-status.com
- Twitter: Follow @AskVenice for updates