Streaming LLM Responses

The streaming endpoint provides real-time delivery of LLM-generated text as it’s being produced, creating a more responsive user experience similar to ChatGPT or Claude.

NOTE: This is handled automatically by components in the Client SDK, should you wish to wish to use that approach.

Streaming Updates Endpoint

POST /api/flowgraph/control/streaming

Key Features

When to use

Request Format

{
  "sessionStepID": string,
  "lastStreamIndex": number
}