jacklangerman commited on
Commit
19f4382
·
verified ·
1 Parent(s): 979fca8

Upload folder using huggingface_hub

Browse files
Files changed (2) hide show
  1. README.md +12 -2
  2. script.py +22 -10
README.md CHANGED
@@ -1,4 +1,14 @@
1
- # My Cool Submission 2025
2
 
3
- This repo contains a submission to the [S23DR Challenge](https://huggingface.co/spaces/usm3d/S23DR) (part of the [USM3D](https://usm3d.github.io/) workshop at CVPR2025). It was prepared by [jacklangerman](https://huggingface.co/jacklangerman).
 
 
 
 
 
 
 
 
 
 
4
 
 
1
+ # Empty solution example for the S23DR competition
2
 
3
+ This repo provides a minimalistic example of a valid, but empty submission to S23DR competition.
4
+ We recommend you take a look at [this example](https://huggingface.co/usm3d/handcrafted_baseline_submission),
5
+ which implements some primitive algorithms and provides useful I/O and visualization functions.
6
+
7
+ This example seeks to simply provide minimal code which succeeds at reading the dataset and producing a solution (in this case two vertices at the origin and edge of zero length connecting them).
8
+
9
+ `script.py` - is the main file which is run by the competition space. It should produce `submission.parquet` as the result of the run. Please see the additional comments in the `script.py` file.
10
+
11
+ ---
12
+ license: apache-2.0
13
+ ---
14
 
script.py CHANGED
@@ -56,18 +56,30 @@ if __name__ == "__main__":
56
  data_path = data_path_local
57
 
58
  print(data_path)
59
- print([str(p) for p in data_path.rglob('*validation*.arrow')])
60
 
61
  # dataset = load_dataset(params['dataset'], trust_remote_code=True, use_auth_token=params['token'])
62
- dataset = load_dataset(
63
- "arrow",
64
- data_files={
65
- "validation": [str(p) for p in data_path.rglob('*validation*.arrow')],
66
- "test": [str(p) for p in data_path.rglob('*test*.arrow')],
67
- },
68
- trust_remote_code=True,
69
- # streaming=True
70
- )
 
 
 
 
 
 
 
 
 
 
 
 
71
  print(dataset, flush=True)
72
  # dataset = load_dataset('webdataset', data_files={)
73
 
 
56
  data_path = data_path_local
57
 
58
  print(data_path)
59
+ print([str(p) for p in data_path.rglob('*validation*.(arrow|tar)')])
60
 
61
  # dataset = load_dataset(params['dataset'], trust_remote_code=True, use_auth_token=params['token'])
62
+ data_files = {
63
+ "validation": [str(p) for p in [*data_path.rglob('*validation*.arrow')]+[*data_path.rglob('*validation*.tar')]],
64
+ "test": [str(p) for p in [*data_path.rglob('*test*.arrow')]+[*data_path.rglob('*test*.tar')]],
65
+ }
66
+ try:
67
+ dataset = load_dataset(
68
+ "arrow",
69
+ data_files=data_files,
70
+ trust_remote_code=True,
71
+ # streaming=True
72
+ )
73
+ print('load with arrow')
74
+ except:
75
+ dataset = load_dataset(
76
+ "webdataset",
77
+ data_files=data_files,
78
+ trust_remote_code=True,
79
+ # streaming=True
80
+ )
81
+ print('load with webdataset')
82
+
83
  print(dataset, flush=True)
84
  # dataset = load_dataset('webdataset', data_files={)
85