Skip to content
This repository was archived by the owner on Mar 9, 2026. It is now read-only.
This repository was archived by the owner on Mar 9, 2026. It is now read-only.

O1 model doesn'ts support streaming completions #430

@sheldonhull

Description

@sheldonhull

Background

With the introduction of newer reasoning models in v1.7.0, we've seen that while some tweaks have been made, streaming support for models like o1 was initially unavailable. This support varies based on the provider:

  • OpenAI: Recently added streaming support.
  • Azure OpenAI: Still lacks streaming support for o1.

Problem

The current implementation assumes that all results are streamed, which is incompatible with some o1 models. They require complete chat completion before results can be read, creating a gap in configuration and logic, especially for Azure OpenAI users.

Proposal

  1. Add a new configuration value:
    • stream: false (default to true).
  2. Introduce a CLI flag for manual override:
    • Use --stream false or --no-stream.

Seem reasonable? If so I'll see if I can knock that out.

ref: current limitations of o1 in azure openai

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions