Documentation Index
Fetch the complete documentation index at: https://concentrate.ai/docs/llms.txt
Use this file to discover all available pages before exploring further.
Welcome to Concentrate AI
The Concentrate AI Responses API provides a unified interface for interacting with multiple AI model providers. Access GPT-5.4, Claude Opus 4.6, Gemini 3.1, and 120+ other models through a single, normalized API with automatic routing and credit tracking.Quickstart
Get started with your first API request in minutes
API Reference
View detailed endpoint documentation
Claude Code Setup
Use Claude Code with any model in our model fortress
Cursor Setup
Use models on Concentrate in Cursor
Key Features
Unified Interface
Unified Interface
One API format works across all providers. No need to learn different request/response formats for OpenAI, Anthropic, Google, or other providers.
Automatic Routing
Automatic Routing
Use
model: "auto" to automatically select the best model based on cost, performance, or latency. The API intelligently routes your requests based on real-time metrics.Multi-Provider Support
Multi-Provider Support
Access models from OpenAI, Anthropic, Google Vertex, AWS Bedrock, Azure, xAI, Cohere, Mistral, Cloudflare, and Hugging Face through a single endpoint.
Streaming Responses
Streaming Responses
Enable real-time streaming via Server-Sent Events (SSE) for a responsive user experience. Works consistently across all providers.
Credit Tracking
Credit Tracking
Built-in usage tracking and billing integration. Monitor token usage, costs, and set spending limits.
Authentication
All API requests require authentication using an API key. Get your API key from the Concentrate.ai dashboard. Include your API key in theAuthorization header:
Base URL
Supported Models
The API supports 70+ models across multiple providers:- OpenAI
- Anthropic
- Google
- Meta
- Others
- GPT-5.4, GPT-5.3 Codex, GPT-5.2, GPT-5.1
- GPT-5.1 Codex Max, GPT-5.1 Codex Mini
- GPT-5 Mini, GPT-5 Nano
- GPT-4.1, GPT-4o, GPT-4o Mini
- o1 (reasoning model)
Check the Model Fortress page in the app for complete listings and current rates.
Model Selection
You can specify models in three ways:- Model name only:
"gpt-5.4"- Automatic provider routing - Provider prefix:
"openai/gpt-5.4"- Specific provider - Auto routing:
"auto"- Let the API choose based on your criteria
Quick Example
Response Format
All responses follow a normalized format regardless of provider:Error Handling
The API uses standard HTTP status codes:| Status Code | Description |
|---|---|
200 | Successful request |
400 | Bad request - Invalid parameters |
401 | Unauthorized - Invalid API key |
402 | Payment required - Insufficient credits |
424 | Failed dependency - Provider error |
500 | Internal server error |
View Error Examples
See detailed error response formats and troubleshooting
Rate Limits
Rate limits are applied per API key and are based on your subscription tier. Limits are enforced using token bucket algorithm with per-minute windows.Contact support to increase your rate limits or discuss enterprise pricing.
Next Steps
Quickstart Guide
Make your first API call
Create Response
Full endpoint documentation
Streaming
Learn about streaming responses
Auto Routing
Automatic model selection