You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
{{ message }}
This repository was archived by the owner on Mar 9, 2026. It is now read-only.
With the introduction of newer reasoning models in v1.7.0, we've seen that while some tweaks have been made, streaming support for models like o1 was initially unavailable. This support varies based on the provider:
OpenAI: Recently added streaming support.
Azure OpenAI: Still lacks streaming support for o1.
Problem
The current implementation assumes that all results are streamed, which is incompatible with some o1 models. They require complete chat completion before results can be read, creating a gap in configuration and logic, especially for Azure OpenAI users.
Proposal
Add a new configuration value:
stream: false (default to true).
Introduce a CLI flag for manual override:
Use --stream false or --no-stream.
Seem reasonable? If so I'll see if I can knock that out.