Evaluate your LLM's response with Prometheus and GPT4 💯
-
Updated
Apr 25, 2025 - Python
Evaluate your LLM's response with Prometheus and GPT4 💯
Deliver safe & effective language models
[ICLR 2025] xFinder: Large Language Models as Automated Evaluators for Reliable Evaluation
This is the repo for the survey of Bias and Fairness in IR with LLMs.
Code and data for "Timo: Towards Better Temporal Reasoning for Language Models" (COLM 2024)
Code and data for Koo et al's ACL 2024 paper "Benchmarking Cognitive Biases in Large Language Models as Evaluators"
LLM-as-judge evals as Semantic Kernel Plugins
Use groq for evaluations
PrometheusLLM is a unique transformer architecture inspired by dignity and recursion. This project aims to explore new frontiers in AI research and welcomes contributions from the community. 🐙🌟
Evaluation system for computer-use agents that uses LLMs to assess agent performance on web browsing and interaction tasks. This judge system reads screenshots, agent trajectories, and final results to provide detailed scoring and feedback.
A tiny, lightweight library for LLM-as-a-Judge evaluations on vLLM-hosted models.
Antibodies for LLMs hallucinations (grouping LLM as a judge, NLI, reward models)
Automated evaluation of llm generated responses on aws
project from UDACITY second module 'agentic workflow' as part of Agentic AI Nanodegree
Add a description, image, and links to the llm-as-evaluator topic page so that developers can more easily learn about it.
To associate your repository with the llm-as-evaluator topic, visit your repo's landing page and select "manage topics."