Spaces:
Runtime error
Runtime error
Commit
Β·
1ca62af
1
Parent(s):
406a5d4
README
Browse files
README.md
CHANGED
@@ -1,4 +1,4 @@
|
|
1 |
-
---
|
2 |
title: Zeno Evals Hub
|
3 |
emoji: π
|
4 |
colorFrom: pink
|
@@ -7,9 +7,25 @@ sdk: docker
|
|
7 |
pinned: false
|
8 |
license: mit
|
9 |
fullWidth: true
|
10 |
-
---
|
11 |
|
12 |
-
#
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
13 |
|
14 |
### Running
|
15 |
|
|
|
1 |
+
<!-- ---
|
2 |
title: Zeno Evals Hub
|
3 |
emoji: π
|
4 |
colorFrom: pink
|
|
|
7 |
pinned: false
|
8 |
license: mit
|
9 |
fullWidth: true
|
10 |
+
--- -->
|
11 |
|
12 |
+
# Zeno + OpenAI Evals
|
13 |
+
|
14 |
+
OpenAI's [Evals library](https://github.com/openai/evals) is a great resource providing evaluation sets for LLMS.
|
15 |
+
|
16 |
+
This repo provides a hub for exploring these results using the [Zeno](https://zenoml.com) evaluation tool.
|
17 |
+
|
18 |
+
## Add New Evals
|
19 |
+
|
20 |
+
To add new evals, add a new entry to `evals/evals.yaml` with the following fields:
|
21 |
+
|
22 |
+
- `results-file`: The first `.jsonl` result from `oaievals`
|
23 |
+
- `link`: A link to the evals commit for this evaluation
|
24 |
+
- `description`: A succint description of what the evaluation is testing
|
25 |
+
- `second-results-file`: An optional second `.jsonl` result from `oaievals`. Must be the same dataset as the first one.
|
26 |
+
- `functions-file`: An optional Python file with [Zeno functions](https://zenoml.com/docs/api) for the evaluations.
|
27 |
+
|
28 |
+
Make sure you test your evals locally before submitting a PR!
|
29 |
|
30 |
### Running
|
31 |
|