Skip to content
Open
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
2 changes: 1 addition & 1 deletion specification/inference/_types/CommonTypes.ts
Original file line number Diff line number Diff line change
Expand Up @@ -30,7 +30,7 @@ export class RequestChatCompletion {
*/
messages: Array<Message>
/**
* The ID of the model to use.
* The ID of the model to use. By default, the model ID is set to the value included when creating the inference endpoint.
*/
model?: string
/**
Expand Down
2 changes: 1 addition & 1 deletion specification/inference/completion/CompletionRequest.ts
Original file line number Diff line number Diff line change
Expand Up @@ -56,7 +56,7 @@ export interface Request extends RequestBase {
*/
input: string | Array<string>
/**
* Optional task settings
* Task settings for the individual inference request. These settings are specific to the <task_type> you specified and override the task settings specified when initializing the service.
*/
task_settings?: TaskSettings
}
Expand Down
2 changes: 1 addition & 1 deletion specification/inference/delete/DeleteRequest.ts
Original file line number Diff line number Diff line change
Expand Up @@ -52,7 +52,7 @@ export interface Request extends RequestBase {
}
query_parameters: {
/**
* When true, the endpoint is not deleted and a list of ingest processors which reference this endpoint is returned.
* When true, checks the semantic_text fields and inference processors that reference the endpoint and returns them in a list, but does not delete the endpoint.
* @server_default false
*/
dry_run?: boolean
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -74,7 +74,7 @@ export interface Request extends RequestBase {
*/
service: AnthropicServiceType
/**
* Settings used to install the inference model. These settings are specific to the `watsonxai` service.
* Settings used to install the inference model. These settings are specific to the `anthropic` service.
*/
service_settings: AnthropicServiceSettings
/**
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -73,7 +73,7 @@ export interface Request extends RequestBase {
*/
service: AzureAiStudioServiceType
/**
* Settings used to install the inference model. These settings are specific to the `openai` service.
* Settings used to install the inference model. These settings are specific to the `azureaistudio` service.
*/
service_settings: AzureAiStudioServiceSettings
/**
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -56,7 +56,7 @@ export interface Request extends RequestBase {
*/
input: string | Array<string>
/**
* Optional task settings
* Task settings for the individual inference request. These settings are specific to the <task_type> you specified and override the task settings specified when initializing the service.
*/
task_settings?: TaskSettings
}
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -64,7 +64,7 @@ export interface Request extends RequestBase {
*/
input: string | string[]
/**
* Optional task settings
* Task settings for the individual inference request. These settings are specific to the <task_type> you specified and override the task settings specified when initializing the service.
*/
task_settings?: TaskSettings
}
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -69,7 +69,7 @@ export interface Request extends RequestBase {
*/
input_type?: string
/**
* Optional task settings
* Task settings for the individual inference request. These settings are specific to the <task_type> you specified and override the task settings specified when initializing the service.
*/
task_settings?: TaskSettings
}
Expand Down