Skip to content

Conversation

@github-actions
Copy link
Contributor

@github-actions github-actions bot commented Dec 5, 2025

This PR adds benchmark results for the mistralai/ministral-14b-2512 model.

The following files have been updated:

  • src/benchmark/results.json - Raw benchmark results
  • src/benchmark/validation-results.json - Validation results against human baseline

This PR was automatically generated by the benchmark workflow.

Note: If you don't want to merge this PR, close it and the model will be added to the untested list to prevent re-processing.

@alrocar


Note

Adds benchmark results and validation comparisons for mistralai/ministral-14b-2512, and includes the model in the benchmark config.

  • Benchmark Config:
    • Add mistralai model ministral-14b-2512 to src/benchmark-config.json.
  • Benchmarks:
    • Append extensive raw runs for mistralai/ministral-14b-2512 in src/benchmark/results.json across many SQL tasks (queries, attempts, timings, tokens, errors).
  • Validation:
    • Update src/benchmark/validation-results.json with per-task comparisons for mistralai/ministral-14b-2512, including SQL used, row counts, match metrics, and aggregate stats (model shows no matches).

Written by Cursor Bugbot for commit eb7a7df. This will update automatically on new commits. Configure here.

@vercel
Copy link

vercel bot commented Dec 5, 2025

The latest updates on your projects. Learn more about Vercel for GitHub.

Project Deployment Preview Comments Updated (UTC)
llm-benchmark Ready Ready Preview Comment Dec 5, 2025 0:20am

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant