Datasets:

Modalities:
Text
Formats:
text
ArXiv:
Libraries:
Datasets
License:
datatune commited on
Commit
aca77fa
·
verified ·
1 Parent(s): f8affe9

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +85 -2
README.md CHANGED
@@ -4,7 +4,7 @@ license: mit
4
  # LogiQA2.0
5
  Logiqa2.0 dataset - logical reasoning in MRC and NLI tasks
6
 
7
- > This repository contains the datasets for our paper [LogiQA2.0 - An Improved Dataset for Logic Reasoning in Question Answering and Textual Inference](https://ieeexplore.ieee.org/abstract/document/10174688)
8
 
9
  ## How to cite
10
  ```
@@ -43,4 +43,87 @@ The MRC part of LogiQA2.0 dataset can be found in the `/logiqa/DATA/LOGIQA` fold
43
 
44
  `dev_zh.txt`: dev split of the Chinese version of dataset in json lines.
45
 
46
- `test_zh.txt`: test split of the Chinese version of dataset in json lines.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
4
  # LogiQA2.0
5
  Logiqa2.0 dataset - logical reasoning in MRC and NLI tasks
6
 
7
+ > This is the official repository for the LogiQA 2.0 datasets in our paper [LogiQA2.0 - An Improved Dataset for Logic Reasoning in Question Answering and Textual Inference](https://ieeexplore.ieee.org/abstract/document/10174688)
8
 
9
  ## How to cite
10
  ```
 
43
 
44
  `dev_zh.txt`: dev split of the Chinese version of dataset in json lines.
45
 
46
+ `test_zh.txt`: test split of the Chinese version of dataset in json lines.
47
+
48
+ `train_fol.zip` is the training data with AMR and FOL annotations. The file is too big so we compressed it.
49
+
50
+ `dev_fol.jsonl` is the dev data with AMR and FOL annotations.
51
+
52
+ `test_fol.jsonl` is the test data with AMR and FOL annotations.
53
+
54
+
55
+ An example:
56
+ ```
57
+ {"id": 10471, "answer": 0, "text": "The medieval Arabs had many manuscripts of the ancient Greek. When needed, they translate them into Arabic. Medieval Arab philosophers were very interested in Aristotle's Theory of Poetry, which was obviously not shared by Arab poets, because a poet interested in it must want to read Homer's poems. Aristotle himself often quotes Homer's poems. However, Homer's poems were not translated into Arabic until modern times.", "question": "Which of the following options, if true, strongly supports the above argument?", "options": ["Some medieval Arab translators have manuscripts of Homer poems in ancient Greek.", "Aristotle's Theory of Poetry is often quoted and commented by modern Arab poets.", "In Aristotle's Theory of Poetry, most of the content is related to drama, and medieval Arabs also wrote plays and performed them.", "A series of medieval Arab stories, such as Arab Night, are very similar to some parts of Homer's epic."], "type": {"Sufficient Conditional Reasoning": true, "Necessry Condtional Reasoning": true, "Conjunctive Reasoning": true}}
58
+ ```
59
+ An example of the Chinese dataset:
60
+ ```
61
+ {"id": 8018, "answer": 0, "text": "常春藤通常指美国东部的八所大学。常春藤一词一直以来是美国名校的代名词,这八所大学不仅历史悠久,治学严谨,而且教学质量极高。这些学校的毕业生大多成为社会精英,他们中的多数人年薪超过20万美元,有很多政界领袖来自常春藤,更有为数众多的科学家毕业于长春藤。", "question": "根据以上条件,下面那个选项一定为真:", "options": ["A.有些社会精英年薪超过20万美金", "B.有些政界领袖年薪不足20万美元", "C.有些科学家年薪超过20万美元", "D.有些政界领袖是社会精英"]}
62
+ ```
63
+
64
+ ### NLI
65
+ The NLI part of LogiQA2.0 dataset can be found in the `/logiqa2nli/DATA/QA2NLI` folder.
66
+
67
+ `train.txt`: train split of the dataset in json lines
68
+
69
+ `dev.txt`: dev split of the dataset in json lines
70
+
71
+ `test.txt`: test split of the dataset in json lines
72
+
73
+ An example:
74
+ ```
75
+ {"label": "not entailed", "major_premise": ["Among the employees of a software company, there are three Cantonese, one Beijinger and three northerners"], "minor_premise": " Four are only responsible for software development and two are only responsible for product sales", "conclusion": "There may be at least 7 people and at most 12 people."}
76
+ ```
77
+ ## Annotations
78
+ The translation and annotation work is outsourced to [Speechocean](en.speechocean.com), the project fund is provided by Microsoft Research Asia
79
+ ### Translation
80
+
81
+ | Final Report | |
82
+ | --- | --- |
83
+ | provider | Speechocean |
84
+ | Project Duration | 2021/10/20-2021/12/3 |
85
+ | Actual Working Hour | 667 hours |
86
+ | Cost | 45000 RMB |
87
+
88
+ Translation style/method:
89
+
90
+ 1. Maintain a unified style, and the translated English questions need to inherit the logic of the original questions;
91
+
92
+ 2. The pronoun in the question need to be unique, the translation needs to be unique and consistent without ambiguity;
93
+
94
+ 3. The translated English conforms to the form of a proper question, that is, it is a clear question from the perspective of the respondent;
95
+
96
+ ### Label consistency
97
+ The label credibility is mannually verified after the translation was done to maintain the truthfulness of the original text. 3 workers run a consistency test on each example, if 2 or more workers give different answer compared to the original answer, the translation would be redone to guareentee the label is correct.
98
+
99
+ ### Additional annotations
100
+ Reasoning types of each question is assigned by a total of 5 workers, each of them corresponds to one reasoning type. We give the description of reasoning types (which can be found in our paper) to the workers. The reasoning types of each question is a collection of 5 workers' decision.
101
+ ## Baseline Guidance
102
+ ### Requirements
103
+ * python 3.6+
104
+ * pytorch 1.0+
105
+ * transformers 2.4.1
106
+ * sklearn
107
+ * tqdm
108
+ * tensorboardX
109
+
110
+ We recommend to use conda to manage virtual environments:
111
+
112
+ ```
113
+ conda env update --name logiqa --file requirements.yml
114
+ ```
115
+ ### Logiqa
116
+ The folder `logiqa` contains both the code and data to run baseline experiments of LogiQA2.0 MRC.
117
+
118
+ To fine-tune the dataset, type following command from the terminal in your :computer:
119
+ ```
120
+ bash logiqa.sh
121
+ ```
122
+ ### Logiqa2nli
123
+ The folder `logiqa2nli` contains both the code and data to run baseline experiments of LogiQA2.0 NLI.
124
+
125
+ To fine-tune the dataset, type following command from the terminal in your :computer:
126
+ ```
127
+ bash qa2nli.sh
128
+ ```
129
+ Note: `./scripts` contains the scripts for running other NLI benchmarks.