applied-ai-018 commited on
Commit
7bcece7
·
verified ·
1 Parent(s): d9e27f8

Add files using upload-large-folder tool

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. lm-evaluation/build/lib/lm_eval/tasks/babi/README.md +45 -0
  2. lm-evaluation/build/lib/lm_eval/tasks/babi/babi.yaml +20 -0
  3. lm-evaluation/build/lib/lm_eval/tasks/okapi/arc_multilingual/README.md +47 -0
  4. lm-evaluation/build/lib/lm_eval/tasks/okapi/arc_multilingual/_arc_yaml +23 -0
  5. lm-evaluation/build/lib/lm_eval/tasks/okapi/arc_multilingual/arc_ar.yaml +7 -0
  6. lm-evaluation/build/lib/lm_eval/tasks/okapi/arc_multilingual/arc_bn.yaml +7 -0
  7. lm-evaluation/build/lib/lm_eval/tasks/okapi/arc_multilingual/arc_ca.yaml +7 -0
  8. lm-evaluation/build/lib/lm_eval/tasks/okapi/arc_multilingual/arc_da.yaml +7 -0
  9. lm-evaluation/build/lib/lm_eval/tasks/okapi/arc_multilingual/arc_de.yaml +7 -0
  10. lm-evaluation/build/lib/lm_eval/tasks/okapi/arc_multilingual/arc_es.yaml +7 -0
  11. lm-evaluation/build/lib/lm_eval/tasks/okapi/arc_multilingual/arc_eu.yaml +7 -0
  12. lm-evaluation/build/lib/lm_eval/tasks/okapi/arc_multilingual/arc_fr.yaml +7 -0
  13. lm-evaluation/build/lib/lm_eval/tasks/okapi/arc_multilingual/arc_gu.yaml +7 -0
  14. lm-evaluation/build/lib/lm_eval/tasks/okapi/arc_multilingual/arc_hi.yaml +7 -0
  15. lm-evaluation/build/lib/lm_eval/tasks/okapi/arc_multilingual/arc_hr.yaml +7 -0
  16. lm-evaluation/build/lib/lm_eval/tasks/okapi/arc_multilingual/arc_hu.yaml +7 -0
  17. lm-evaluation/build/lib/lm_eval/tasks/okapi/arc_multilingual/arc_hy.yaml +7 -0
  18. lm-evaluation/build/lib/lm_eval/tasks/okapi/arc_multilingual/arc_id.yaml +7 -0
  19. lm-evaluation/build/lib/lm_eval/tasks/okapi/arc_multilingual/arc_it.yaml +7 -0
  20. lm-evaluation/build/lib/lm_eval/tasks/okapi/arc_multilingual/arc_kn.yaml +7 -0
  21. lm-evaluation/build/lib/lm_eval/tasks/okapi/arc_multilingual/arc_ml.yaml +7 -0
  22. lm-evaluation/build/lib/lm_eval/tasks/okapi/arc_multilingual/arc_mr.yaml +7 -0
  23. lm-evaluation/build/lib/lm_eval/tasks/okapi/arc_multilingual/arc_ne.yaml +7 -0
  24. lm-evaluation/build/lib/lm_eval/tasks/okapi/arc_multilingual/arc_nl.yaml +7 -0
  25. lm-evaluation/build/lib/lm_eval/tasks/okapi/arc_multilingual/arc_pt.yaml +7 -0
  26. lm-evaluation/build/lib/lm_eval/tasks/okapi/arc_multilingual/arc_ro.yaml +7 -0
  27. lm-evaluation/build/lib/lm_eval/tasks/okapi/arc_multilingual/arc_ru.yaml +7 -0
  28. lm-evaluation/build/lib/lm_eval/tasks/okapi/arc_multilingual/arc_sk.yaml +7 -0
  29. lm-evaluation/build/lib/lm_eval/tasks/okapi/arc_multilingual/arc_sr.yaml +7 -0
  30. lm-evaluation/build/lib/lm_eval/tasks/okapi/arc_multilingual/arc_sv.yaml +7 -0
  31. lm-evaluation/build/lib/lm_eval/tasks/okapi/arc_multilingual/arc_ta.yaml +7 -0
  32. lm-evaluation/build/lib/lm_eval/tasks/okapi/arc_multilingual/arc_te.yaml +7 -0
  33. lm-evaluation/build/lib/lm_eval/tasks/okapi/arc_multilingual/arc_uk.yaml +7 -0
  34. lm-evaluation/build/lib/lm_eval/tasks/okapi/arc_multilingual/arc_vi.yaml +7 -0
  35. lm-evaluation/build/lib/lm_eval/tasks/okapi/arc_multilingual/arc_zh.yaml +7 -0
  36. lm-evaluation/build/lib/lm_eval/tasks/okapi/arc_multilingual/utils.py +33 -0
  37. lm-evaluation/build/lib/lm_eval/tasks/okapi/hellaswag_multilingual/hellaswag_ar.yaml +6 -0
  38. lm-evaluation/build/lib/lm_eval/tasks/okapi/hellaswag_multilingual/hellaswag_bn.yaml +6 -0
  39. lm-evaluation/build/lib/lm_eval/tasks/okapi/hellaswag_multilingual/hellaswag_es.yaml +6 -0
  40. lm-evaluation/build/lib/lm_eval/tasks/okapi/hellaswag_multilingual/hellaswag_fr.yaml +6 -0
  41. lm-evaluation/build/lib/lm_eval/tasks/okapi/hellaswag_multilingual/hellaswag_hu.yaml +6 -0
  42. lm-evaluation/build/lib/lm_eval/tasks/okapi/hellaswag_multilingual/hellaswag_te.yaml +6 -0
  43. lm-evaluation/build/lib/lm_eval/tasks/okapi/hellaswag_multilingual/hellaswag_uk.yaml +6 -0
  44. lm-evaluation/build/lib/lm_eval/tasks/okapi/mmlu_multilingual/_default_yaml +17 -0
  45. lm-evaluation/build/lib/lm_eval/tasks/okapi/mmlu_multilingual/_generate_configs.py +27 -0
  46. lm-evaluation/build/lib/lm_eval/tasks/okapi/mmlu_multilingual/m_mmlu_bn.yaml +4 -0
  47. lm-evaluation/build/lib/lm_eval/tasks/okapi/mmlu_multilingual/m_mmlu_ca.yaml +4 -0
  48. lm-evaluation/build/lib/lm_eval/tasks/okapi/mmlu_multilingual/m_mmlu_de.yaml +4 -0
  49. lm-evaluation/build/lib/lm_eval/tasks/okapi/mmlu_multilingual/m_mmlu_en.yaml +4 -0
  50. lm-evaluation/build/lib/lm_eval/tasks/okapi/mmlu_multilingual/m_mmlu_es.yaml +4 -0
lm-evaluation/build/lib/lm_eval/tasks/babi/README.md ADDED
@@ -0,0 +1,45 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # bAbI
2
+
3
+ ### Paper
4
+
5
+ Title: Towards ai-complete question answering: A set of prerequisite toy tasks
6
+ Abstract: https://arxiv.org/abs/1502.05698
7
+
8
+ One long-term goal of machine learning research is to produce methods that are applicable to reasoning and natural language, in particular building an intelligent dialogue agent. To measure progress towards that goal, we argue for the usefulness of a set of proxy tasks that evaluate reading comprehension via question answering. Our tasks measure understanding in several ways: whether a system is able to answer questions via chaining facts, simple induction, deduction and many more. The tasks are designed to be prerequisites for any system that aims to be capable of conversing with a human. We believe many existing learning systems can currently not solve them, and hence our aim is to classify these tasks into skill sets, so that researchers can identify (and then rectify) the failings of their systems. We also extend and improve the recently introduced Memory Networks model, and show it is able to solve some, but not all, of the tasks.
9
+
10
+ Homepage: https://github.com/facebookarchive/bAbI-tasks
11
+
12
+
13
+ ### Citation
14
+
15
+ ```
16
+ @article{weston2015towards,
17
+ title={Towards ai-complete question answering: A set of prerequisite toy tasks},
18
+ author={Weston, Jason and Bordes, Antoine and Chopra, Sumit and Rush, Alexander M and Van Merri{\"e}nboer, Bart and Joulin, Armand and Mikolov, Tomas},
19
+ journal={arXiv preprint arXiv:1502.05698},
20
+ year={2015}
21
+ }
22
+ ```
23
+
24
+ ### Groups and Tasks
25
+
26
+ #### Groups
27
+
28
+ * Not part of a group yet
29
+
30
+ #### Tasks
31
+
32
+ * `babi`
33
+
34
+ ### Checklist
35
+
36
+ For adding novel benchmarks/datasets to the library:
37
+ * [ ] Is the task an existing benchmark in the literature?
38
+ * [ ] Have you referenced the original paper that introduced the task?
39
+ * [ ] If yes, does the original paper provide a reference implementation? If so, have you checked against the reference implementation and documented how to run such a test?
40
+
41
+
42
+ If other tasks on this dataset are already supported:
43
+ * [ ] Is the "Main" variant of this task clearly denoted?
44
+ * [ ] Have you provided a short sentence in a README on what each new variant adds / evaluates?
45
+ * [ ] Have you noted which, if any, published evaluation setups are matched by this variant?
lm-evaluation/build/lib/lm_eval/tasks/babi/babi.yaml ADDED
@@ -0,0 +1,20 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ task: babi
2
+ dataset_path: Muennighoff/babi
3
+ dataset_name: null
4
+ output_type: generate_until
5
+ training_split: train
6
+ validation_split: valid
7
+ test_split: test
8
+ doc_to_text: "Passage: {{passage}}Question: {{question}}\nAnswer:"
9
+ doc_to_target: " {{answer}}"
10
+ target_delimiter: ""
11
+ generation_kwargs:
12
+ until:
13
+ - "\n"
14
+ - "Passage:"
15
+ metric_list:
16
+ - metric: exact_match
17
+ aggregation: mean
18
+ higher_is_better: true
19
+ metadata:
20
+ version: 1.0
lm-evaluation/build/lib/lm_eval/tasks/okapi/arc_multilingual/README.md ADDED
@@ -0,0 +1,47 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Multilingual ARC
2
+
3
+ ### Paper
4
+
5
+ Title: `Okapi: Instruction-tuned Large Language Models in Multiple Languages with Reinforcement Learning from Human Feedback`
6
+
7
+ Abstract: https://arxiv.org/abs/2307.16039
8
+
9
+ A key technology for the development of large language models (LLMs) involves instruction tuning that helps align the models' responses with human expectations to realize impressive learning abilities. Two major approaches for instruction tuning characterize supervised fine-tuning (SFT) and reinforcement learning from human feedback (RLHF), which are currently applied to produce the best commercial LLMs (e.g., ChatGPT). To improve the accessibility of LLMs for research and development efforts, various instruction-tuned open-source LLMs have also been introduced recently, e.g., Alpaca, Vicuna, to name a few. However, existing open-source LLMs have only been instruction-tuned for English and a few popular languages, thus hindering their impacts and accessibility to many other languages in the world. Among a few very recent work to explore instruction tuning for LLMs in multiple languages, SFT has been used as the only approach to instruction-tune LLMs for multiple languages. This has left a significant gap for fine-tuned LLMs based on RLHF in diverse languages and raised important questions on how RLHF can boost the performance of multilingual instruction tuning. To overcome this issue, we present Okapi, the first system with instruction-tuned LLMs based on RLHF for multiple languages. Okapi introduces instruction and response-ranked data in 26 diverse languages to facilitate the experiments and development of future multilingual LLM research. We also present benchmark datasets to enable the evaluation of generative LLMs in multiple languages. Our experiments demonstrate the advantages of RLHF for multilingual instruction over SFT for different base models and datasets. Our framework and resources are released at this https URL.
10
+
11
+ Homepage: `https://github.com/nlp-uoregon/Okapi`
12
+
13
+
14
+ ### Citation
15
+
16
+ ```
17
+ @article{dac2023okapi,
18
+ title={Okapi: Instruction-tuned Large Language Models in Multiple Languages with Reinforcement Learning from Human Feedback},
19
+ author={Dac Lai, Viet and Van Nguyen, Chien and Ngo, Nghia Trung and Nguyen, Thuat and Dernoncourt, Franck and Rossi, Ryan A and Nguyen, Thien Huu},
20
+ journal={arXiv e-prints},
21
+ pages={arXiv--2307},
22
+ year={2023}
23
+ }
24
+ ```
25
+
26
+ ### Groups and Tasks
27
+
28
+ #### Groups
29
+
30
+ - arc_multilingual
31
+
32
+ #### Tasks
33
+
34
+ - `arc_{ar,bn,ca,da,de,es,eu,fr,gu,hi,hr,hu,hy,id,it,kn,ml,mr,ne,nl,pt,ro,ru,sk,sr,sv,ta,te,uk,vi,zh}`
35
+
36
+ ### Checklist
37
+
38
+ For adding novel benchmarks/datasets to the library:
39
+ * [x] Is the task an existing benchmark in the literature?
40
+ * [x] Have you referenced the original paper that introduced the task?
41
+ * [x] If yes, does the original paper provide a reference implementation? If so, have you checked against the reference implementation and documented how to run such a test?
42
+
43
+
44
+ If other tasks on this dataset are already supported:
45
+ * [ ] Is the "Main" variant of this task clearly denoted?
46
+ * [ ] Have you provided a short sentence in a README on what each new variant adds / evaluates?
47
+ * [ ] Have you noted which, if any, published evaluation setups are matched by this variant?
lm-evaluation/build/lib/lm_eval/tasks/okapi/arc_multilingual/_arc_yaml ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ group:
2
+ - arc_multilingual
3
+ dataset_path: null
4
+ dataset_name: null
5
+ output_type: multiple_choice
6
+ training_split: train
7
+ validation_split: validation
8
+ test_split: test
9
+ process_docs: !function utils.process_docs
10
+ doc_to_text: "query"
11
+ doc_to_target: "gold"
12
+ doc_to_choice: "choices"
13
+ should_decontaminate: true
14
+ doc_to_decontamination_query: "query"
15
+ metric_list:
16
+ - metric: acc
17
+ aggregation: mean
18
+ higher_is_better: true
19
+ - metric: acc_norm
20
+ aggregation: mean
21
+ higher_is_better: true
22
+ metadata:
23
+ version: 1.0
lm-evaluation/build/lib/lm_eval/tasks/okapi/arc_multilingual/arc_ar.yaml ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ include: _arc_yaml
2
+ task: arc_ar
3
+ dataset_path: alexandrainst/m_arc
4
+ dataset_name: ar
5
+ training_split: train
6
+ validation_split: validation
7
+ test_split: test
lm-evaluation/build/lib/lm_eval/tasks/okapi/arc_multilingual/arc_bn.yaml ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ include: _arc_yaml
2
+ task: arc_bn
3
+ dataset_path: alexandrainst/m_arc
4
+ dataset_name: bn
5
+ training_split: train
6
+ validation_split: validation
7
+ test_split: test
lm-evaluation/build/lib/lm_eval/tasks/okapi/arc_multilingual/arc_ca.yaml ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ include: _arc_yaml
2
+ task: arc_ca
3
+ dataset_path: alexandrainst/m_arc
4
+ dataset_name: ca
5
+ training_split: train
6
+ validation_split: validation
7
+ test_split: test
lm-evaluation/build/lib/lm_eval/tasks/okapi/arc_multilingual/arc_da.yaml ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ include: _arc_yaml
2
+ task: arc_da
3
+ dataset_path: alexandrainst/m_arc
4
+ dataset_name: da
5
+ training_split: train
6
+ validation_split: validation
7
+ test_split: test
lm-evaluation/build/lib/lm_eval/tasks/okapi/arc_multilingual/arc_de.yaml ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ include: _arc_yaml
2
+ task: arc_de
3
+ dataset_path: alexandrainst/m_arc
4
+ dataset_name: de
5
+ training_split: train
6
+ validation_split: validation
7
+ test_split: test
lm-evaluation/build/lib/lm_eval/tasks/okapi/arc_multilingual/arc_es.yaml ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ include: _arc_yaml
2
+ task: arc_es
3
+ dataset_path: alexandrainst/m_arc
4
+ dataset_name: es
5
+ training_split: train
6
+ validation_split: validation
7
+ test_split: test
lm-evaluation/build/lib/lm_eval/tasks/okapi/arc_multilingual/arc_eu.yaml ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ include: _arc_yaml
2
+ task: arc_eu
3
+ dataset_path: alexandrainst/m_arc
4
+ dataset_name: eu
5
+ training_split: train
6
+ validation_split: validation
7
+ test_split: test
lm-evaluation/build/lib/lm_eval/tasks/okapi/arc_multilingual/arc_fr.yaml ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ include: _arc_yaml
2
+ task: arc_fr
3
+ dataset_path: alexandrainst/m_arc
4
+ dataset_name: fr
5
+ training_split: train
6
+ validation_split: validation
7
+ test_split: test
lm-evaluation/build/lib/lm_eval/tasks/okapi/arc_multilingual/arc_gu.yaml ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ include: _arc_yaml
2
+ task: arc_gu
3
+ dataset_path: alexandrainst/m_arc
4
+ dataset_name: gu
5
+ training_split: train
6
+ validation_split: validation
7
+ test_split: test
lm-evaluation/build/lib/lm_eval/tasks/okapi/arc_multilingual/arc_hi.yaml ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ include: _arc_yaml
2
+ task: arc_hi
3
+ dataset_path: alexandrainst/m_arc
4
+ dataset_name: hi
5
+ training_split: train
6
+ validation_split: validation
7
+ test_split: test
lm-evaluation/build/lib/lm_eval/tasks/okapi/arc_multilingual/arc_hr.yaml ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ include: _arc_yaml
2
+ task: arc_hr
3
+ dataset_path: alexandrainst/m_arc
4
+ dataset_name: hr
5
+ training_split: train
6
+ validation_split: validation
7
+ test_split: test
lm-evaluation/build/lib/lm_eval/tasks/okapi/arc_multilingual/arc_hu.yaml ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ include: _arc_yaml
2
+ task: arc_hu
3
+ dataset_path: alexandrainst/m_arc
4
+ dataset_name: hu
5
+ training_split: train
6
+ validation_split: validation
7
+ test_split: test
lm-evaluation/build/lib/lm_eval/tasks/okapi/arc_multilingual/arc_hy.yaml ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ include: _arc_yaml
2
+ task: arc_hy
3
+ dataset_path: alexandrainst/m_arc
4
+ dataset_name: hy
5
+ training_split: train
6
+ validation_split: validation
7
+ test_split: test
lm-evaluation/build/lib/lm_eval/tasks/okapi/arc_multilingual/arc_id.yaml ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ include: _arc_yaml
2
+ task: arc_id
3
+ dataset_path: alexandrainst/m_arc
4
+ dataset_name: id
5
+ training_split: train
6
+ validation_split: validation
7
+ test_split: test
lm-evaluation/build/lib/lm_eval/tasks/okapi/arc_multilingual/arc_it.yaml ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ include: _arc_yaml
2
+ task: arc_it
3
+ dataset_path: alexandrainst/m_arc
4
+ dataset_name: it
5
+ training_split: train
6
+ validation_split: validation
7
+ test_split: test
lm-evaluation/build/lib/lm_eval/tasks/okapi/arc_multilingual/arc_kn.yaml ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ include: _arc_yaml
2
+ task: arc_kn
3
+ dataset_path: alexandrainst/m_arc
4
+ dataset_name: kn
5
+ training_split: train
6
+ validation_split: validation
7
+ test_split: test
lm-evaluation/build/lib/lm_eval/tasks/okapi/arc_multilingual/arc_ml.yaml ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ include: _arc_yaml
2
+ task: arc_ml
3
+ dataset_path: alexandrainst/m_arc
4
+ dataset_name: ml
5
+ training_split: train
6
+ validation_split: validation
7
+ test_split: test
lm-evaluation/build/lib/lm_eval/tasks/okapi/arc_multilingual/arc_mr.yaml ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ include: _arc_yaml
2
+ task: arc_mr
3
+ dataset_path: alexandrainst/m_arc
4
+ dataset_name: mr
5
+ training_split: train
6
+ validation_split: validation
7
+ test_split: test
lm-evaluation/build/lib/lm_eval/tasks/okapi/arc_multilingual/arc_ne.yaml ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ include: _arc_yaml
2
+ task: arc_ne
3
+ dataset_path: alexandrainst/m_arc
4
+ dataset_name: ne
5
+ training_split: train
6
+ validation_split: validation
7
+ test_split: test
lm-evaluation/build/lib/lm_eval/tasks/okapi/arc_multilingual/arc_nl.yaml ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ include: _arc_yaml
2
+ task: arc_nl
3
+ dataset_path: alexandrainst/m_arc
4
+ dataset_name: nl
5
+ training_split: train
6
+ validation_split: validation
7
+ test_split: test
lm-evaluation/build/lib/lm_eval/tasks/okapi/arc_multilingual/arc_pt.yaml ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ include: _arc_yaml
2
+ task: arc_pt
3
+ dataset_path: alexandrainst/m_arc
4
+ dataset_name: pt
5
+ training_split: train
6
+ validation_split: validation
7
+ test_split: test
lm-evaluation/build/lib/lm_eval/tasks/okapi/arc_multilingual/arc_ro.yaml ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ include: _arc_yaml
2
+ task: arc_ro
3
+ dataset_path: alexandrainst/m_arc
4
+ dataset_name: ro
5
+ training_split: train
6
+ validation_split: validation
7
+ test_split: test
lm-evaluation/build/lib/lm_eval/tasks/okapi/arc_multilingual/arc_ru.yaml ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ include: _arc_yaml
2
+ task: arc_ru
3
+ dataset_path: alexandrainst/m_arc
4
+ dataset_name: ru
5
+ training_split: train
6
+ validation_split: validation
7
+ test_split: test
lm-evaluation/build/lib/lm_eval/tasks/okapi/arc_multilingual/arc_sk.yaml ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ include: _arc_yaml
2
+ task: arc_sk
3
+ dataset_path: alexandrainst/m_arc
4
+ dataset_name: sk
5
+ training_split: train
6
+ validation_split: validation
7
+ test_split: test
lm-evaluation/build/lib/lm_eval/tasks/okapi/arc_multilingual/arc_sr.yaml ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ include: _arc_yaml
2
+ task: arc_sr
3
+ dataset_path: alexandrainst/m_arc
4
+ dataset_name: sr
5
+ training_split: train
6
+ validation_split: validation
7
+ test_split: test
lm-evaluation/build/lib/lm_eval/tasks/okapi/arc_multilingual/arc_sv.yaml ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ include: _arc_yaml
2
+ task: arc_sv
3
+ dataset_path: alexandrainst/m_arc
4
+ dataset_name: sv
5
+ training_split: train
6
+ validation_split: validation
7
+ test_split: test
lm-evaluation/build/lib/lm_eval/tasks/okapi/arc_multilingual/arc_ta.yaml ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ include: _arc_yaml
2
+ task: arc_ta
3
+ dataset_path: alexandrainst/m_arc
4
+ dataset_name: ta
5
+ training_split: train
6
+ validation_split: validation
7
+ test_split: test
lm-evaluation/build/lib/lm_eval/tasks/okapi/arc_multilingual/arc_te.yaml ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ include: _arc_yaml
2
+ task: arc_te
3
+ dataset_path: alexandrainst/m_arc
4
+ dataset_name: te
5
+ training_split: train
6
+ validation_split: validation
7
+ test_split: test
lm-evaluation/build/lib/lm_eval/tasks/okapi/arc_multilingual/arc_uk.yaml ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ include: _arc_yaml
2
+ task: arc_uk
3
+ dataset_path: alexandrainst/m_arc
4
+ dataset_name: uk
5
+ training_split: train
6
+ validation_split: validation
7
+ test_split: test
lm-evaluation/build/lib/lm_eval/tasks/okapi/arc_multilingual/arc_vi.yaml ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ include: _arc_yaml
2
+ task: arc_vi
3
+ dataset_path: alexandrainst/m_arc
4
+ dataset_name: vi
5
+ training_split: train
6
+ validation_split: validation
7
+ test_split: test
lm-evaluation/build/lib/lm_eval/tasks/okapi/arc_multilingual/arc_zh.yaml ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ include: _arc_yaml
2
+ task: arc_zh
3
+ dataset_path: alexandrainst/m_arc
4
+ dataset_name: zh
5
+ training_split: train
6
+ validation_split: validation
7
+ test_split: test
lm-evaluation/build/lib/lm_eval/tasks/okapi/arc_multilingual/utils.py ADDED
@@ -0,0 +1,33 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import re
2
+
3
+ import datasets
4
+
5
+
6
+ def preprocess(text):
7
+ if text is None:
8
+ return " "
9
+ text = text.strip()
10
+ text = text.replace(" [title]", ". ")
11
+ text = re.sub("\\[.*?\\]", "", text)
12
+ text = text.replace(" ", " ")
13
+ return text
14
+
15
+
16
+ def process_docs(dataset: datasets.Dataset) -> datasets.Dataset:
17
+ def _process_doc(doc):
18
+ # breakpoint()
19
+ out_doc = {
20
+ "id": doc["id"],
21
+ "query": "Question: " + preprocess(doc["instruction"]) + "\nAnswer:",
22
+ "choices": [
23
+ preprocess(doc["option_a"]),
24
+ preprocess(doc["option_b"]),
25
+ preprocess(doc["option_c"]),
26
+ preprocess(doc["option_d"]),
27
+ preprocess(doc["option_e"]),
28
+ ],
29
+ "gold": ["A", "B", "C", "D", "E"].index(doc["answer"]),
30
+ }
31
+ return out_doc
32
+
33
+ return dataset.map(_process_doc)
lm-evaluation/build/lib/lm_eval/tasks/okapi/hellaswag_multilingual/hellaswag_ar.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ include: _hellaswag_yaml
2
+ task: hellaswag_ar
3
+ dataset_path: alexandrainst/m_hellaswag
4
+ dataset_name: ar
5
+ training_split: null
6
+ validation_split: val
lm-evaluation/build/lib/lm_eval/tasks/okapi/hellaswag_multilingual/hellaswag_bn.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ include: _hellaswag_yaml
2
+ task: hellaswag_bn
3
+ dataset_path: alexandrainst/m_hellaswag
4
+ dataset_name: bn
5
+ training_split: null
6
+ validation_split: val
lm-evaluation/build/lib/lm_eval/tasks/okapi/hellaswag_multilingual/hellaswag_es.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ include: _hellaswag_yaml
2
+ task: hellaswag_es
3
+ dataset_path: alexandrainst/m_hellaswag
4
+ dataset_name: es
5
+ training_split: null
6
+ validation_split: val
lm-evaluation/build/lib/lm_eval/tasks/okapi/hellaswag_multilingual/hellaswag_fr.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ include: _hellaswag_yaml
2
+ task: hellaswag_fr
3
+ dataset_path: alexandrainst/m_hellaswag
4
+ dataset_name: fr
5
+ training_split: null
6
+ validation_split: val
lm-evaluation/build/lib/lm_eval/tasks/okapi/hellaswag_multilingual/hellaswag_hu.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ include: _hellaswag_yaml
2
+ task: hellaswag_hu
3
+ dataset_path: alexandrainst/m_hellaswag
4
+ dataset_name: hu
5
+ training_split: null
6
+ validation_split: val
lm-evaluation/build/lib/lm_eval/tasks/okapi/hellaswag_multilingual/hellaswag_te.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ include: _hellaswag_yaml
2
+ task: hellaswag_te
3
+ dataset_path: alexandrainst/m_hellaswag
4
+ dataset_name: te
5
+ training_split: null
6
+ validation_split: val
lm-evaluation/build/lib/lm_eval/tasks/okapi/hellaswag_multilingual/hellaswag_uk.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ include: _hellaswag_yaml
2
+ task: hellaswag_uk
3
+ dataset_path: alexandrainst/m_hellaswag
4
+ dataset_name: uk
5
+ training_split: null
6
+ validation_split: val
lm-evaluation/build/lib/lm_eval/tasks/okapi/mmlu_multilingual/_default_yaml ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ group:
2
+ - m_mmlu
3
+ dataset_path: alexandrainst/m_mmlu
4
+ test_split: test
5
+ fewshot_split: train
6
+ fewshot_config:
7
+ sampler: first_n
8
+ output_type: multiple_choice
9
+ doc_to_text: "{{instruction.strip()}}\nA. {{option_a}}\nB. {{option_b}}\nC. {{option_c}}\nD. {{option_d}}\nAnswer:"
10
+ doc_to_choice: ["A", "B", "C", "D"]
11
+ doc_to_target: answer
12
+ metric_list:
13
+ - metric: acc
14
+ aggregation: mean
15
+ higher_is_better: true
16
+ metadata:
17
+ version: 0.0
lm-evaluation/build/lib/lm_eval/tasks/okapi/mmlu_multilingual/_generate_configs.py ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import datasets
2
+ import yaml
3
+ from tqdm import tqdm
4
+
5
+
6
+ def main() -> None:
7
+ dataset_path = "alexandrainst/m_mmlu"
8
+
9
+ for task in tqdm(datasets.get_dataset_infos(dataset_path).keys()):
10
+ file_name = f"m_mmlu_{task}.yaml"
11
+ try:
12
+ with open(f"{file_name}", "w") as f:
13
+ f.write("# Generated by _generate_configs.py\n")
14
+ yaml.dump(
15
+ {
16
+ "include": "_default_yaml",
17
+ "task": f"{dataset_path.split('/')[-1]}_{task}",
18
+ "dataset_name": task,
19
+ },
20
+ f,
21
+ )
22
+ except FileExistsError:
23
+ pass
24
+
25
+
26
+ if __name__ == "__main__":
27
+ main()
lm-evaluation/build/lib/lm_eval/tasks/okapi/mmlu_multilingual/m_mmlu_bn.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by _generate_configs.py
2
+ dataset_name: bn
3
+ include: _default_yaml
4
+ task: m_mmlu_bn
lm-evaluation/build/lib/lm_eval/tasks/okapi/mmlu_multilingual/m_mmlu_ca.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by _generate_configs.py
2
+ dataset_name: ca
3
+ include: _default_yaml
4
+ task: m_mmlu_ca
lm-evaluation/build/lib/lm_eval/tasks/okapi/mmlu_multilingual/m_mmlu_de.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by _generate_configs.py
2
+ dataset_name: de
3
+ include: _default_yaml
4
+ task: m_mmlu_de
lm-evaluation/build/lib/lm_eval/tasks/okapi/mmlu_multilingual/m_mmlu_en.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by _generate_configs.py
2
+ dataset_name: en
3
+ include: _default_yaml
4
+ task: m_mmlu_en
lm-evaluation/build/lib/lm_eval/tasks/okapi/mmlu_multilingual/m_mmlu_es.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by _generate_configs.py
2
+ dataset_name: es
3
+ include: _default_yaml
4
+ task: m_mmlu_es