Quick Start#
1. Get Your API Key#
Log in to the AIone Console and create a new key on the API Keys page.We recommend creating separate keys for each environment or use case to simplify access control, usage tracking, and troubleshooting.2. Send Your First Request#
AIone is fully compatible with the OpenAI Chat Completions API. Simply set base_url to https://api.nexara.net/v1 and use the OpenAI SDK or any compatible client.Streaming is recommended: For interactive use cases, always enable stream: true. This reduces time-to-first-token from 10+ seconds down to 2-3 seconds, significantly improving the user experience.
Python (Streaming - Recommended)#
Python (Non-Streaming)#
Node.js (Streaming - Recommended)#
cURL (Streaming)#
3. Switch Models#
All models are served through the unified v1/chat/completions endpoint. Simply change the model parameter to switch:| Model | model Parameter | Best For |
|---|
| Claude Sonnet 4.6 | claude-sonnet-4-6 | General-purpose; balanced speed and quality |
| Claude Opus 4.6 | claude-opus-4-6 | Strongest reasoning; complex tasks |
| Claude Sonnet 4.5 | claude-sonnet-4-5-20250929 | High-quality general reasoning and code |
| Claude Opus 4.5 | claude-opus-4-5-20251101 | Top-tier complex analysis |
| Claude Haiku 4.5 | claude-haiku-4-5-20251001 | Ultra-fast responses; high concurrency |
| Claude Sonnet 4 | claude-sonnet-4-20250514 | Stable release; broad compatibility |
| GPT-5.4 | gpt-5.4 | OpenAI's current flagship |
| GPT-5.2 | gpt-5.2 | High-quality general and coding tasks |
| GPT-5 | gpt-5 | Existing integration compatibility |
| GPT-5 Mini | gpt-5-mini | High concurrency; low latency |
| GPT-4.1 | gpt-4.1 | Code and complex instruction following |
| GPT-4o | gpt-4o | Multimodal and general tasks |
| Gemini 2.5 Pro | gemini-2.5-pro | Complex reasoning and long context |
| Gemini 2.5 Flash | gemini-2.5-flash | Speed-first scenarios |
| Gemini 2.5 Flash Lite | gemini-2.5-flash-lite | Lower cost; high concurrency |
For the full model list, visit the Models & Pricing page in the console or query the /v1/models endpoint.4. Try the Sandbox#
After logging in, go to the Sandbox page to test any model directly in your browser -- no code required.Modified at 2026-04-04 16:06:55