12 models for chat

Chat & LLM API — every model, one endpoint

GPT-5 · Claude Opus 4.6 · Claude Sonnet 4.5 · Gemini 3 Pro · DeepSeek v3 · Mistral Large. Same /chat/completions request shape, streaming SSE support, up to 200k context.

What it's for

Common workflows.

Chat product backends — agents, copilots, assistants
Content generation — blog posts, product descriptions, emails
Code tasks — GPT-5 Codex, Claude Sonnet for agentic coding
Data extraction and transformation
Translation and localization
Research and analysis — long-context summarization
API call

Ship in 30 seconds.

curl https://aimarcusimage.eu/api/v1/chat/completions \
  -H "Authorization: Bearer sk-aig-..." \
  -H "Content-Type: application/json" \
  -d '{
  "model": "claude-sonnet-4-5",
  "messages": [
    { "role": "user", "content": "Explain streaming SSE to a junior dev in 3 paragraphs." }
  ],
  "stream": true
}'
Models

Pick the model that fits.

All billed from the same credit balance.

FAQ

Questions, answered.

Is it OpenAI-compatible?+

Yes. POST /api/v1/chat/completions with {model, messages, stream}. The response shape matches OpenAI exactly — drop-in replacement for most SDKs.

Does streaming work?+

Yes — SSE streaming for GPT, Claude and most OpenRouter-backed models. Just set stream=true in the payload.

What about long context?+

Gemini 3 Pro supports up to 2M tokens context. Claude Opus 4.6 supports 200k. See each model page for the exact limit.

Ready to start?

Sign up, get $10 free credits, ship your first chat generation in under a minute.