Use the returned task ID to query the task for the final result.
Documentation Index
Fetch the complete documentation index at: https://docs.foxapi.cc/llms.txt
Use this file to discover all available pages before exploring further.
All endpoints require Bearer Token authentication. Add to the request header:
Authorization: Bearer YOUR_API_KEY
YOUR_API_KEY is the API Token (sk-... format).
Model name. Common values:
claude-opus-4-7gemini-2.5-pronemotron-3-nano-omnigpt-5.4 / gpt-5.5 / kimi-k2.6 / gemini-3-pro-preview and others"claude-opus-4-7"
"gemini-2.5-pro"
"nemotron-3-nano-omni"
User prompt, up to 100,000 characters.
100000"Summarize the theory of relativity in two sentences."
Synchronous mode. When true, the endpoint blocks until the upstream completes and returns the full response (if stream=true at the same time, returns an SSE stream); when false, the endpoint returns the task ID immediately, and results are fetched via GET /v1/tasks/{task_id} or the SSE endpoint.
false
Whether to stream. When true, the Submit response includes stream.url pointing to the SSE subscription path; streaming chunks are unified as the OpenAI chat.completion.chunk format.
false
Generation token limit. Optional.
x >= 164
Sampling temperature, range [0, 2]. Optional.
0 <= x <= 20.3
System instruction, prepended to the conversation context. Optional, up to 10,000 characters.
10000"You are a terse assistant."
Whether to include reasoning tokens. Passed through to the upstream; concrete semantics depend on the upstream model (thinking models like gemini-2.5-pro may require true).
Task created (async mode) / full response (sync mode)
Submit response, conforming to the unified task standard shape. results / error are fixed at null during submit; they are returned via GET /v1/tasks/{task_id} after the task completes or fails. In sync=true, stream=false mode, the endpoint directly returns the full OpenAI ChatCompletion JSON (does not follow this shape).
Task ID, formatted as task-llmrouter-{timestamp}-{8random}.
"task-llmrouter-1776874565-yq3szvcu"
llm.generation.task "llm.generation.task"
llm "llm"
The model name submitted by the client (echoed verbatim)
"claude-opus-4-7"
pending "pending"
0
1776874565
Returns {url: ...} when stream=true; null when stream=false.
Fixed at null during submit; returned via GET /v1/tasks/{task_id} after the task completes — results[0] is the full OpenAI ChatCompletion response.
null
Fixed at null during submit; returned via GET /v1/tasks/{task_id} when the task fails.
null