Alvinn-aai commited on
Commit
9eba8d6
Β·
1 Parent(s): daa3ab0

submission extraction

Browse files
Files changed (1) hide show
  1. src/about.py +34 -21
src/about.py CHANGED
@@ -45,35 +45,48 @@ To reproduce our results, here is the commands you can run:
45
  """
46
 
47
  EVALUATION_QUEUE_TEXT = """
48
- ## Some good practices before submitting a model
49
-
50
- ### 1) Make sure you can load your model and tokenizer using AutoClasses:
51
- ```python
52
- from transformers import AutoConfig, AutoModel, AutoTokenizer
53
- config = AutoConfig.from_pretrained("your model name", revision=revision)
54
- model = AutoModel.from_pretrained("your model name", revision=revision)
55
- tokenizer = AutoTokenizer.from_pretrained("your model name", revision=revision)
 
 
 
 
56
  ```
57
- If this step fails, follow the error messages to debug your model before submitting it. It's likely your model has been improperly uploaded.
58
 
59
- Note: make sure your model is public!
60
- Note: if your model needs `use_remote_code=True`, we do not support this option yet but we are working on adding it, stay posted!
 
 
 
61
 
62
- ### 2) Convert your model weights to [safetensors](https://huggingface.co/docs/safetensors/index)
63
- It's a new format for storing weights which is safer and faster to load and use. It will also allow us to add the number of parameters of your model to the `Extended Viewer`!
 
 
 
 
64
 
65
- ### 3) Make sure your model has an open license!
66
- This is a leaderboard for Open LLMs, and we'd love for as many people as possible to know they can use your model πŸ€—
67
 
68
- ### 4) Fill up your model card
69
- When we add extra information about models to the leaderboard, it will be automatically taken from the model card
 
 
 
 
 
 
 
70
 
71
- ## In case of model failure
72
- If your model is displayed in the `FAILED` category, its execution stopped.
73
- Make sure you have followed the above steps first.
74
- If everything is done, check you can launch the EleutherAIHarness on your model locally, using the above command without modifications (you can add `--limit` to limit the number of examples per task).
75
  """
76
 
 
77
  CITATION_BUTTON_LABEL = """πŸ“š How to cite FormulaOne"""
78
  CITATION_BUTTON_TEXT = r"""
79
  @misc{beniamini2025formulaonemeasuringdepthalgorithmic,
 
45
  """
46
 
47
  EVALUATION_QUEUE_TEXT = """
48
+ ## πŸ§ͺ Submitting to the FormulaOne Leaderboard
49
+
50
+ This leaderboard evaluates systems on the FormulaOne core dataset. Submissions consist of a .jsonl file with solution code for each problem.
51
+
52
+ ### πŸ“ 1. Format Your Submission File
53
+
54
+ Your submission must be a .jsonl file with one entry per problem:
55
+
56
+ ```json
57
+ {"problem_id": 1, "solution": "<your Python code here>"}
58
+ {"problem_id": 2, "solution": "<your Python code here>"}
59
+ ...
60
  ```
 
61
 
62
+ - problem_id: Must match the official list of FormulaOne core problems.
63
+ - solution: A Python code implementing the required callback functions.
64
+
65
+ πŸ“„ Full list of problem_ids:
66
+ View the [FormulaOne core dataset](https://github.com/double-ai/formulaone-dataset-release/dataset/formulaone) for the complete list of problem IDs.
67
 
68
+ ⚠️ Validation Rules:
69
+ Submissions must:
70
+ β€’ Contain exactly two columns: ["problem_id", "solution"]
71
+ β€’ Include all required problems (no missing/unknown IDs)
72
+ β€’ Provide solutions as Python strings
73
+ β€’ Avoid duplicates
74
 
 
 
75
 
76
+ ### πŸ“€ 2. Submit via the Web UI
77
+ 1. Upload your .jsonl file.
78
+ 2. Fill in:
79
+ β€’ System Name
80
+ β€’ Organization
81
+ β€’ System Type
82
+ 3. Click Submit.
83
+
84
+ ### ⏱️ After Submission
85
 
86
+ Submissions are validated and evaluated within ~24 hours. Results will appear on the leaderboard once processed.
 
 
 
87
  """
88
 
89
+
90
  CITATION_BUTTON_LABEL = """πŸ“š How to cite FormulaOne"""
91
  CITATION_BUTTON_TEXT = r"""
92
  @misc{beniamini2025formulaonemeasuringdepthalgorithmic,