Skip to content
Closed
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
6 changes: 6 additions & 0 deletions providers/mistral-ai/codestral-2508.yaml
Original file line number Diff line number Diff line change
@@ -1,15 +1,21 @@
costs:
- input_cost_per_token: 3.e-7
input_cost_per_token_batches: 1.5e-7
output_cost_per_token: 9.e-7
output_cost_per_token_batches: 4.5e-7
region: "*"
features:
- function_calling
- tool_choice
- response_schema
- assistant_prefill
- code
limits:
context_window: 128000
max_input_tokens: 128000
max_output_tokens: 128000
max_tokens: 128000
mode: chat
model: codestral-2508
sources:
- https://docs.mistral.ai/models/codestral-25-08
14 changes: 13 additions & 1 deletion providers/mistral-ai/codestral-embed-2505.yaml
Original file line number Diff line number Diff line change
@@ -1,8 +1,20 @@
costs:
- input_cost_per_token: 1.5e-7
input_cost_per_token_batches: 7.5e-8
region: "*"
limits:
context_window: 8192
max_input_tokens: 8192
max_tokens: 8192
mode: embedding
model: codestral-embed-2505
removeParams:
- max_tokens
- temperature
- top_p
- seed
- stop
- safe_prompt
- stream
sources:
- https://mistral.ai/news/codestral-embed
- https://docs.mistral.ai/capabilities/embeddings/code_embeddings
13 changes: 12 additions & 1 deletion providers/mistral-ai/codestral-embed.yaml
Original file line number Diff line number Diff line change
@@ -1,8 +1,19 @@
costs:
- input_cost_per_token: 1.5e-7
input_cost_per_token_batches: 7.5e-8
region: "*"
limits:
context_window: 8192
max_input_tokens: 8192
max_tokens: 8192
mode: embedding
model: codestral-embed
removeParams:
- max_tokens
- temperature
- top_p
- seed
- stop
- safe_prompt
- stream
sources:
- https://mistral.ai/news/codestral-embed
5 changes: 5 additions & 0 deletions providers/mistral-ai/codestral-latest.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -7,9 +7,14 @@ features:
- tool_choice
- response_schema
- assistant_prefill
- code
limits:
context_window: 128000
max_input_tokens: 128000
max_output_tokens: 128000
max_tokens: 128000
mode: chat
model: codestral-latest
sources:
- https://docs.mistral.ai/models/codestral-25-08
- https://docs.mistral.ai/capabilities/code_generation
8 changes: 7 additions & 1 deletion providers/mistral-ai/devstral-2512.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -7,9 +7,15 @@ features:
- tool_choice
- response_schema
- assistant_prefill
- code
limits:
context_window: 256000
max_input_tokens: 256000
max_output_tokens: 256000
max_tokens: 256000
mode: chat
model: devstral-2512
model: devstral-2-25-12
sources:
- https://docs.mistral.ai/models/devstral-2-25-12
- https://docs.mistral.ai/capabilities/function_calling
- https://docs.mistral.ai/capabilities/vision
7 changes: 4 additions & 3 deletions providers/mistral-ai/devstral-latest.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -8,8 +8,9 @@ features:
- response_schema
- assistant_prefill
limits:
max_input_tokens: 256000
max_output_tokens: 256000
max_tokens: 256000
context_window: 262144
mode: chat
model: devstral-latest
sources:
- https://docs.mistral.ai/models/devstral-2-25-12
- https://mistral.ai/news/devstral-2-vibe-cli
4 changes: 4 additions & 0 deletions providers/mistral-ai/devstral-medium-2507.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -7,9 +7,13 @@ features:
- tool_choice
- response_schema
- assistant_prefill
- code
limits:
context_window: 128000
max_input_tokens: 128000
max_output_tokens: 128000
max_tokens: 128000
mode: chat
model: devstral-medium-2507
sources:
- https://docs.mistral.ai/models/devstral-medium-1-0-25-07
8 changes: 5 additions & 3 deletions providers/mistral-ai/devstral-medium-latest.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -7,9 +7,11 @@ features:
- tool_choice
- response_schema
- assistant_prefill
- code
- tools
limits:
max_input_tokens: 256000
max_output_tokens: 256000
max_tokens: 256000
context_window: 128000
mode: chat
model: devstral-medium-latest
sources:
- https://docs.mistral.ai/models/devstral-medium-1-0-25-07
1 change: 1 addition & 0 deletions providers/mistral-ai/devstral-small-2507.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -7,6 +7,7 @@ features:
- tool_choice
- response_schema
- assistant_prefill
isDeprecated: true
limits:
max_input_tokens: 128000
max_output_tokens: 128000
Expand Down
4 changes: 4 additions & 0 deletions providers/mistral-ai/labs-devstral-small-2512.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -7,9 +7,13 @@ features:
- tool_choice
- response_schema
- assistant_prefill
- code
limits:
context_window: 256000
max_input_tokens: 256000
max_output_tokens: 256000
max_tokens: 256000
mode: chat
model: labs-devstral-small-2512
sources:
- https://docs.mistral.ai/models/devstral-small-2-25-12
5 changes: 5 additions & 0 deletions providers/mistral-ai/labs-mistral-small-creative.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -5,9 +5,14 @@ costs:
features:
- function_calling
- chat
- response_schema
- assistant_prefill
limits:
context_window: 32000
max_input_tokens: 32000
max_output_tokens: 32000
max_tokens: 32000
mode: chat
model: labs-mistral-small-creative
sources:
- https://docs.mistral.ai/models/mistral-small-creative-25-12
10 changes: 10 additions & 0 deletions providers/mistral-ai/magistral-medium-2509.yaml
Original file line number Diff line number Diff line change
@@ -1,15 +1,25 @@
costs:
- input_cost_per_token: 0.000002
input_cost_per_token_batches: 0.000001
output_cost_per_token: 0.000005
output_cost_per_token_batches: 0.0000025
region: "*"
features:
- function_calling
- tool_choice
- response_schema
- assistant_prefill
- system_messages
limits:
context_window: 128000
max_input_tokens: 128000
max_output_tokens: 128000
max_tokens: 128000
mode: chat
model: magistral-medium-2509
sources:
- https://docs.mistral.ai/models/magistral-medium-1-2-25-09
- https://docs.mistral.ai/capabilities/reasoning
- https://docs.mistral.ai/capabilities/vision
- https://docs.mistral.ai/capabilities/batch
thinking: true
13 changes: 10 additions & 3 deletions providers/mistral-ai/magistral-medium-latest.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -4,13 +4,20 @@ costs:
region: "*"
features:
- function_calling
- vision
- image_input
- chat
- pdf
- pdf_input
- doc
- tool_choice
- response_schema
- assistant_prefill
limits:
max_input_tokens: 128000
max_output_tokens: 128000
max_tokens: 4128000
context_window: 128000
mode: chat
model: magistral-medium-latest
sources:
- https://docs.mistral.ai/models/magistral-medium-1-2-25-09
- https://docs.mistral.ai/capabilities/reasoning
thinking: true
12 changes: 10 additions & 2 deletions providers/mistral-ai/magistral-small-2509.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -5,9 +5,17 @@ costs:
features:
- function_calling
- chat
- vision
- image_input
- response_schema
- doc
limits:
context_window: 128000
max_input_tokens: 128000
max_output_tokens: 128000
max_tokens: 128000
mode: chat
model: magistral-small-2509
sources:
- https://docs.mistral.ai/models/magistral-small-1-2-25-09
- https://docs.mistral.ai/capabilities/reasoning
- https://docs.mistral.ai/capabilities/vision
thinking: true
12 changes: 12 additions & 0 deletions providers/mistral-ai/magistral-small-latest.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -4,13 +4,25 @@ costs:
region: "*"
features:
- function_calling
- vision
- image_input
- chat
- pdf
- doc
- tool_choice
- response_schema
- assistant_prefill
limits:
context_window: 128000
max_input_tokens: 128000
max_output_tokens: 128000
max_tokens: 128000
mode: chat
model: magistral-small-latest
params:
- key: max_tokens
maxValue: 128000
sources:
- https://docs.mistral.ai/models/magistral-small-1-2-25-09
- https://docs.mistral.ai/capabilities/reasoning
thinking: true
12 changes: 10 additions & 2 deletions providers/mistral-ai/ministral-14b-2512.yaml
Original file line number Diff line number Diff line change
@@ -1,14 +1,22 @@
costs:
- input_cost_per_image: 0
input_cost_per_token: 2.e-7
- input_cost_per_token: 2.e-7
output_cost_per_token: 2.e-7
region: "*"
features:
- function_calling
- vision
- image_input
- response_schema
- assistant_prefill
- tool_choice
limits:
context_window: 262144
max_input_tokens: 262144
max_output_tokens: 262144
max_tokens: 262144
mode: chat
model: ministral-14b-2512
sources:
- https://docs.mistral.ai/models/ministral-3-14b-25-12
- https://docs.mistral.ai/capabilities/function_calling
- https://docs.mistral.ai/capabilities/vision
10 changes: 7 additions & 3 deletions providers/mistral-ai/ministral-14b-latest.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -6,9 +6,13 @@ costs:
features:
- function_calling
- vision
- image_input
- response_schema
- assistant_prefill
limits:
max_input_tokens: 256000
max_output_tokens: 256000
max_tokens: 256000
context_window: 256000
mode: chat
model: ministral-14b-latest
sources:
- https://docs.mistral.ai/models/ministral-3-14b-25-12
- https://docs.mistral.ai/capabilities/vision
10 changes: 5 additions & 5 deletions providers/mistral-ai/ministral-3b-2512.yaml
Original file line number Diff line number Diff line change
@@ -1,14 +1,14 @@
costs:
- input_cost_per_image: 0
input_cost_per_token: 1.e-7
- input_cost_per_token: 1.e-7
output_cost_per_token: 1.e-7
region: "*"
features:
- function_calling
- vision
- response_schema
limits:
max_input_tokens: 256000
max_output_tokens: 256000
max_tokens: 256000
context_window: 256000
mode: chat
model: ministral-3b-2512
sources:
- https://docs.mistral.ai/models/ministral-3-3b-25-12
1 change: 1 addition & 0 deletions providers/mistral-ai/ministral-3b-latest.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -6,6 +6,7 @@ costs:
features:
- function_calling
- vision
isDeprecated: true
limits:
max_input_tokens: 256000
max_output_tokens: 256000
Expand Down
11 changes: 11 additions & 0 deletions providers/mistral-ai/ministral-8b-2512.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -5,10 +5,21 @@ costs:
region: "*"
features:
- function_calling
- parallel_function_calling
- vision
- image_input
- tool_choice
- response_schema
- assistant_prefill
limits:
context_window: 256000
max_input_tokens: 256000
max_output_tokens: 256000
max_tokens: 256000
mode: chat
model: ministral-8b-2512
sources:
- https://docs.mistral.ai/models/ministral-3-8b-25-12
- https://docs.mistral.ai/capabilities/vision
- https://docs.mistral.ai/capabilities/function_calling
- https://docs.mistral.ai/api
17 changes: 12 additions & 5 deletions providers/mistral-ai/ministral-8b-latest.yaml
Original file line number Diff line number Diff line change
@@ -1,14 +1,21 @@
costs:
- input_cost_per_image: 0
input_cost_per_token: 1.5e-7
- input_cost_per_token: 1.5e-7
output_cost_per_token: 1.5e-7
region: "*"
features:
- function_calling
- parallel_function_calling
- vision
- image_input
- tool_choice
- response_schema
- system_messages
- assistant_prefill
limits:
max_input_tokens: 256000
max_output_tokens: 256000
max_tokens: 256000
context_window: 256000
mode: chat
model: ministral-8b-latest
sources:
- https://docs.mistral.ai/models/ministral-3-8b-25-12
- https://docs.mistral.ai/capabilities/function_calling
- https://docs.mistral.ai/capabilities/vision
1 change: 1 addition & 0 deletions providers/mistral-ai/mistral-embed-2312.yaml
Original file line number Diff line number Diff line change
@@ -1,6 +1,7 @@
costs:
- input_cost_per_token: 1.e-7
region: "*"
isDeprecated: true
limits:
max_input_tokens: 8192
max_tokens: 4096
Expand Down
Loading