Skip to main content

Documentation Index

Fetch the complete documentation index at: https://docs.dolphy.chat/llms.txt

Use this file to discover all available pages before exploring further.

Endpoint

POST https://dolphy.chat/api/v1/chat/completions
Authorization: Bearer dpy_live_...
This endpoint is fully compatible with OpenAI’s chat completions API. Drop in the official openai SDK by changing baseURL.

Request body

{
  "model": "venice-uncensored",
  "messages": [
    { "role": "system", "content": "You are a helpful assistant." },
    { "role": "user", "content": "Tell me about Mars." }
  ],
  "temperature": 0.8,
  "max_tokens": 512,
  "stream": false
}
FieldTypeNotes
modelstringDefault venice-uncensored
messagesarrayUp to 200 messages, OpenAI shape
temperaturenumber0–2
top_pnumber0–1
max_tokensintUp to 8192
streambooleanSSE streaming if true
stopstring | string[]Stop sequences

Streaming

Set stream: true to get token-by-token SSE chunks (OpenAI format with data: {…}\n\n events ending in data: [DONE]).
const stream = await client.chat.completions.create({
  model: "venice-uncensored",
  messages: [{ role: "user", content: "Tell me a story." }],
  stream: true,
});
for await (const chunk of stream) {
  process.stdout.write(chunk.choices[0]?.delta?.content ?? "");
}

Billing

1 credit per 10,000 tokens (input + output, minimum 1 per call). For streamed responses, the final chunk includes the usage block we use to compute the bill.

Errors

StatusWhen
401Invalid or revoked API key
402Insufficient credits
422Content policy rejection from upstream
429Rate limit (60/min per key)
502Upstream provider error
503Provider temporarily unavailable