PurAI is SHUT DOWN. This website is purely for archival purposes. No such services exist anymore.
Docs
OpenAI
Streaming

OpenAI Chat Completions Streaming

PurAI supports streaming, which powers the "typing" feel of ChatGPT, and can improve response speed. You can just set the stream field to true to use streaming.

Request

Supported URLs

  • /openai/chat/completions
  • /third-party/chat/completions

TS Stream Interface

interface Stream {
	stream?: boolean;
};

Example JSON Request Body

{
    "model": "gpt-3.5-turbo",
    "messages": [
        {
            "role": "user",
            "content": "who are you? answer with 3 words"
        }
    ],
    "stream": true
}

Response

Example Responses (Encoded)

  1. The TypeScript response interface is the same as the previous page's
  2. Look at the highlighted portion of the following code snippets!
  3. Streaming responses doesn't support the cache field!
{
    "id": "chatcmpl-QXlha2FBbmROaXhpZUFyZUF3ZXNvbWUK",
    "object": "chat.completion",
    "created": 0,
    "model": "gpt-3.5-turbo-0301",
    "usage": {
        "prompt_tokens": 0,
        "completion_tokens": 0,
        "total_tokens": 0
    },
    "choices": [
        {
            "index": 0,
            "delta": {
                "content": "AI la"
            },
            "finish_reason": null
        }
    ],
    "provider": "Churchless",
    "overwritten": false,
    "cache": {
        "status": 500,
        "error": {
            "message": "Some of our providers returned with errors. Errors are automatically reported to our developers.",
            "records": [...]
        }
    },
    "calledFunctions": []
}
{
    "id": "chatcmpl-QXlha2FBbmROaXhpZUFyZUF3ZXNvbWUK",
    "object": "chat.completion",
    "created": 0,
    "model": "gpt-3.5-turbo-0301",
    "usage": {
        "prompt_tokens": 0,
        "completion_tokens": 0,
        "total_tokens": 0
    },
    "choices": [
        {
            "index": 0,
            "delta": {
                "content": "nguag"
            },
            "finish_reason": null
        }
    ],
    "provider": "Churchless",
    "overwritten": false,
    "cache": {
        "status": 500,
        "error": {
            "message": "Some of our providers returned with errors. Errors are automatically reported to our developers.",
            "records": [...]
        }
    },
    "calledFunctions": []
}
{
    "id": "chatcmpl-QXlha2FBbmROaXhpZUFyZUF3ZXNvbWUK",
    "object": "chat.completion",
    "created": 0,
    "model": "gpt-3.5-turbo-0301",
    "usage": {
        "prompt_tokens": 0,
        "completion_tokens": 0,
        "total_tokens": 0
    },
    "choices": [
        {
            "index": 0,
            "delta": {
                "content": "e mod"
            },
            "finish_reason": null
        }
    ],
    "provider": "Churchless",
    "overwritten": false,
    "cache": {
        "status": 500,
        "error": {
            "message": "Some of our providers returned with errors. Errors are automatically reported to our developers.",
            "records": [...]
        }
    },
    "calledFunctions": []
}
{
    "id": "chatcmpl-QXlha2FBbmROaXhpZUFyZUF3ZXNvbWUK",
    "object": "chat.completion",
    "created": 0,
    "model": "gpt-3.5-turbo-0301",
    "usage": {
        "prompt_tokens": 0,
        "completion_tokens": 0,
        "total_tokens": 0
    },
    "choices": [
        {
            "index": 0,
            "delta": {
                "content": "el."
            },
            "finish_reason": null
        }
    ],
    "provider": "Churchless",
    "overwritten": false,
    "cache": {
        "status": 500,
        "error": {
            "message": "Some of our providers returned with errors. Errors are automatically reported to our developers.",
            "records": [...]
        }
    },
    "calledFunctions": []
}