The dataset viewer is not available for this split.
Error code: FeaturesError
Exception: ArrowInvalid
Message: Schema at index 1 was different:
training_data_buffer_size: int64
last_training_data: int64
last_proof_id: string
external_theorems_used_cnt: int64
local_theorems_used_cnt: int64
total_proof_step_cnt: int64
data_filename_prefix: string
data_filename_suffix: string
lemma_ref_filename_prefix: string
lemma_ref_filename_suffix: string
num_theorems: int64
vs
training_data: list<item: struct<proof_id: string, all_useful_defns_theorems: list<item: null>, goal_description: null, start_goals: list<item: struct<hypotheses: list<item: string>, goal: string, relevant_defns: list<item: null>, used_theorems_local: list<item: null>, used_theorems_external: list<item: null>, possible_useful_theorems_external: list<item: null>, possible_useful_theorems_local: list<item: null>>>, end_goals: list<item: struct<hypotheses: list<item: string>, goal: string, relevant_defns: list<item: null>, used_theorems_local: list<item: null>, used_theorems_external: list<item: null>, possible_useful_theorems_external: list<item: null>, possible_useful_theorems_local: list<item: null>>>, proof_steps: list<item: string>, simplified_goals: list<item: null>, addition_state_info: struct<>, file_path: string, project_id: string, theorem_name: string>>
Traceback: Traceback (most recent call last):
File "/src/services/worker/src/worker/job_runners/split/first_rows.py", line 228, in compute_first_rows_from_streaming_response
iterable_dataset = iterable_dataset._resolve_features()
File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/iterable_dataset.py", line 3339, in _resolve_features
features = _infer_features_from_batch(self.with_format(None)._head())
File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/iterable_dataset.py", line 2096, in _head
return next(iter(self.iter(batch_size=n)))
File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/iterable_dataset.py", line 2300, in iter
for key, example in iterator:
File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/iterable_dataset.py", line 1856, in __iter__
for key, pa_table in self._iter_arrow():
File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/iterable_dataset.py", line 1878, in _iter_arrow
yield from self.ex_iterable._iter_arrow()
File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/iterable_dataset.py", line 504, in _iter_arrow
yield new_key, pa.Table.from_batches(chunks_buffer)
File "pyarrow/table.pxi", line 4116, in pyarrow.lib.Table.from_batches
File "pyarrow/error.pxi", line 154, in pyarrow.lib.pyarrow_internal_check_status
File "pyarrow/error.pxi", line 91, in pyarrow.lib.check_status
pyarrow.lib.ArrowInvalid: Schema at index 1 was different:
training_data_buffer_size: int64
last_training_data: int64
last_proof_id: string
external_theorems_used_cnt: int64
local_theorems_used_cnt: int64
total_proof_step_cnt: int64
data_filename_prefix: string
data_filename_suffix: string
lemma_ref_filename_prefix: string
lemma_ref_filename_suffix: string
num_theorems: int64
vs
training_data: list<item: struct<proof_id: string, all_useful_defns_theorems: list<item: null>, goal_description: null, start_goals: list<item: struct<hypotheses: list<item: string>, goal: string, relevant_defns: list<item: null>, used_theorems_local: list<item: null>, used_theorems_external: list<item: null>, possible_useful_theorems_external: list<item: null>, possible_useful_theorems_local: list<item: null>>>, end_goals: list<item: struct<hypotheses: list<item: string>, goal: string, relevant_defns: list<item: null>, used_theorems_local: list<item: null>, used_theorems_external: list<item: null>, possible_useful_theorems_external: list<item: null>, possible_useful_theorems_local: list<item: null>>>, proof_steps: list<item: string>, simplified_goals: list<item: null>, addition_state_info: struct<>, file_path: string, project_id: string, theorem_name: string>>Need help to make the dataset viewer work? Make sure to review how to configure the dataset viewer, and open a discussion for direct support.
π ProofWalaDataset
The ProofWalaDataset is a multilingual dataset of formal theorem proving traces collected from multiple interactive theorem prover (ITP) ecosystems. It provides a structured view of proof steps, goals, hypotheses, and theorem names from diverse mathematical and program verification libraries.
This dataset is intended for researchers and practitioners working on:
- Automated theorem proving
- Formal code generation
- Machine learning for logic
- Proof step prediction
- Multi-language transfer in formal systems
π Dataset Structure
The dataset is organized into the following ITP families:
lean/coq/GeoCoq/math-comp/multilingual/(cross-formal-language hybrid)
Each family includes standard splits:train/, test/, and eval/, each containing multiple JSON files.
Each JSON file contains a top-level key: "training_data" with a list of proof records.
π Each record contains
| Field | Description |
|---|---|
proof_id |
Unique identifier for the proof trace |
goal_description |
Optional natural language description of the proof |
start_goals |
List of starting goals (each with goal and hypotheses) |
end_goals |
Final goals after applying proof steps |
proof_steps |
List of applied proof tactics (inv, rewrite, etc.) |
simplified_goals |
Simplified representations of goals (if any) |
all_useful_defns_theorems |
Set of useful definitions or theorems (static analysis) |
addition_state_info |
Optional additional metadata about the proof context |
file_path |
Source file where the proof appears |
project_id |
The ITP project or repository path (e.g., CompCert) |
theorem_name |
Name of the theorem being proved |
For convenience, structured fields such as start_goals[*].goal, start_goals[*].hypotheses, end_goals[*].goal, and end_goals[*].hypotheses are exposed directly through the Croissant metadata.
π§ Use Cases
- Pretraining and finetuning LLMs for formal verification
- Evaluating proof search strategies
- Building cross-language proof translators
- Fine-grained proof tactic prediction
π Format
- Data format: JSON
- Schema described via Croissant metadata (
croissant.json) - Fully validated using mlcroissant
- Downloads last month
- 11