https://github.com/paradite/16x-eval
Evaluation framework for LLMs and prompts on real world coding tasks in JavaScript, Python and SQL
https://github.com/paradite/16x-eval
eval evaluation llm prompt prompt-engineering
Last synced: 8 months ago
JSON representation
Evaluation framework for LLMs and prompts on real world coding tasks in JavaScript, Python and SQL
- Host: GitHub
- URL: https://github.com/paradite/16x-eval
- Owner: paradite
- Created: 2024-02-29T11:05:03.000Z (almost 2 years ago)
- Default Branch: main
- Last Pushed: 2025-04-28T17:29:03.000Z (8 months ago)
- Last Synced: 2025-04-28T18:33:36.628Z (8 months ago)
- Topics: eval, evaluation, llm, prompt, prompt-engineering
- Language: Python
- Homepage: https://eval.16x.engineer/
- Size: 281 KB
- Stars: 3
- Watchers: 2
- Forks: 1
- Open Issues: 0
-
Metadata Files:
- Readme: README.md
Awesome Lists containing this project
README
# Eval data
Evaluation data for LLMs and prompts on real world coding tasks and writing tasks.
Built by [16x Prompt](https://prompt.16x.engineer/) and [16x Eval](https://eval.16x.engineer/).
## Coding Projects
### Next.js
[emoji-todo](/projects/emoji-todo/) - A simple Next.js TODO app with emojis.
### SQL
[sql](/projects/sql/) - SQL code snippets.
### Python
[python-script](/projects/python-script/) - Python script code snippets.
### Benchmark Visualization
[visualization](/projects/visualization/) - Coding a visualization of benchmark results.
## Writing Projects
### AI Timeline
[ai-timeline](/projects/ai-timeline/) - Writing an AI Timeline.
## Local eval tool
I have decided to build a separate tool to evaluate models and prompts.
See [16x Eval website](https://eval.16x.engineer/) for more information.