Overview
The Uno LLM Gateway provides an endpoint at/api/gateway/openai/responses that implements OpenAI’s Responses API specification. This means you can use any OpenAI SDK (Python, JavaScript, Go, etc.) without modifying your code - just change the base URL.
Usage Examples
main.py
Accessing Other Provider Models
The gateway allows you to access models from other providers (like Gemini, Anthropic, etc.) using the same OpenAI SDK. Simply prefix the model name with the provider name followed by a slash:Gemini/gemini-3-flash-preview- Access Google Gemini modelsAnthropic/claude-haiku-4-5- Access Anthropic Claude models
multi_provider.py
Streaming Support
The gateway supports streaming responses via Server-Sent Events (SSE). Use your SDK’s streaming methods as you normally would:streaming.py
Supported Features
The gateway currently supports the Responses API with:- ✅ Text generation - Standard text completions
- ✅ Images - Image generation and processing
- ✅ Tool calling - Function calling capabilities
- ✅ Reasoning - Advanced reasoning models
Authentication
The gateway accepts authentication via theAuthorization header with a Bearer token:
- Virtual Key: Use a virtual key (starts with
sk-amg-) for managed access control - Direct API Key: Use your OpenAI API key directly