Skip to main content

Streams a model-generated response for a chat turn while persisting both the user message and assistant reply.

POST 

/studio/chat_stream

Selects the provider implementation dynamically by options.Provider.
Token usage and estimated cost are computed prior to generation.
Response headers are set to disable buffering and compression for smooth SSE delivery.

Request

Responses

OK