cabreraalex commited on
Commit
1ca62af
Β·
1 Parent(s): 406a5d4
Files changed (1) hide show
  1. README.md +19 -3
README.md CHANGED
@@ -1,4 +1,4 @@
1
- ---
2
  title: Zeno Evals Hub
3
  emoji: πŸƒ
4
  colorFrom: pink
@@ -7,9 +7,25 @@ sdk: docker
7
  pinned: false
8
  license: mit
9
  fullWidth: true
10
- ---
11
 
12
- # Dashboard for exploring OpenAI Evals results
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
13
 
14
  ### Running
15
 
 
1
+ <!-- ---
2
  title: Zeno Evals Hub
3
  emoji: πŸƒ
4
  colorFrom: pink
 
7
  pinned: false
8
  license: mit
9
  fullWidth: true
10
+ --- -->
11
 
12
+ # Zeno + OpenAI Evals
13
+
14
+ OpenAI's [Evals library](https://github.com/openai/evals) is a great resource providing evaluation sets for LLMS.
15
+
16
+ This repo provides a hub for exploring these results using the [Zeno](https://zenoml.com) evaluation tool.
17
+
18
+ ## Add New Evals
19
+
20
+ To add new evals, add a new entry to `evals/evals.yaml` with the following fields:
21
+
22
+ - `results-file`: The first `.jsonl` result from `oaievals`
23
+ - `link`: A link to the evals commit for this evaluation
24
+ - `description`: A succint description of what the evaluation is testing
25
+ - `second-results-file`: An optional second `.jsonl` result from `oaievals`. Must be the same dataset as the first one.
26
+ - `functions-file`: An optional Python file with [Zeno functions](https://zenoml.com/docs/api) for the evaluations.
27
+
28
+ Make sure you test your evals locally before submitting a PR!
29
 
30
  ### Running
31