Skip to content

Feature: Groq Batch & OpenAI Flex Processing API #12

@qdrddr

Description

@qdrddr

I'd like to benefit from cheaper batch processing using OpenAI Flex Processing API and Groq Batch API that offers cheaper batch processing.

As a fallback mechanism when the batch is not finished by X minutes, I'd like the batch to be cenceled and fallback to standard LLM processing.

https://developers.openai.com/api/docs/guides/flex-processing

For inspiration
vectorize-io/hindsight#365

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions