Skip to content

Conversation

@chiragbhatt311
Copy link

No description provided.

Copy link
Contributor

@gemini-code-assist gemini-code-assist bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Code Review

This pull request correctly passes the num_samples field to the external vLLM inference engine. The change adds the n parameter to the request payload, which is the standard way to specify the number of samples in OpenAI-compatible APIs. The existing response handling logic is already capable of processing multiple generated sequences, so this change effectively enables the feature. The implementation is straightforward and correct.

@pcmoritz pcmoritz added the tx label Feb 3, 2026
@pcmoritz pcmoritz changed the title Pass num_samples field in payload in vLLM request [tx] Pass num_samples field in payload in vLLM request Feb 3, 2026
@pcmoritz
Copy link
Collaborator

pcmoritz commented Feb 3, 2026

Thanks a lot for fixing this! I'm in the process of preparing a release at the moment and will test and include this PR :)

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants