Projects in Awesome Lists tagged with llm-evaluation-toolkit
A curated list of projects in awesome lists tagged with llm-evaluation-toolkit .
https://github.com/johnsnowlabs/langtest
Deliver safe & effective language models
ai-safety ai-testing artificial-intelligence benchmark-framework benchmarks ethics-in-ai large-language-models llm llm-as-evaluator llm-evaluation-toolkit llm-test llm-testing ml-safety ml-testing mlops model-assessment nlp responsible-ai trustworthy-ai
Last synced: 14 May 2025
https://github.com/JohnSnowLabs/langtest
Deliver safe & effective language models
ai-safety ai-testing artificial-intelligence benchmark-framework benchmarks ethics-in-ai large-language-models llm llm-as-evaluator llm-evaluation-toolkit llm-test llm-testing ml-safety ml-testing mlops model-assessment nlp responsible-ai trustworthy-ai
Last synced: 01 Feb 2025
https://github.com/athina-ai/athina-evals
Python SDK for running evaluations on LLM generated responses
evaluation evaluation-framework evaluation-metrics llm-eval llm-evaluation llm-evaluation-toolkit llm-ops llmops
Last synced: 15 Apr 2025
https://github.com/Re-Align/just-eval
A simple GPT-based evaluation tool for multi-aspect, interpretable assessment of LLMs.
evaluation gpt4 llm llm-eval llm-evaluation llm-evaluation-toolkit
Last synced: 23 Apr 2025
https://github.com/zhuohaoyu/kieval
[ACL'24] A Knowledge-grounded Interactive Evaluation Framework for Large Language Models
acl2024 explainable-ai llm llm-evaluation llm-evaluation-framework llm-evaluation-metrics llm-evaluation-toolkit machine-learning
Last synced: 11 Jan 2025