Update README.md
Browse files
README.md
CHANGED
|
@@ -65,7 +65,7 @@ configs:
|
|
| 65 |
|
| 66 |
The wa-hls4ml projects dataset, comprized of the Vivado/Vitis projects of neural networks converted into HLS Code via hls4ml. Projects are complete, and include all logs, HLS Code, VHDL Code, Intermediete Representations, and source keras models.
|
| 67 |
|
| 68 |
-
This is a companion dataset to the [wa-hls4ml](fastmachinelearning/wa-hls4ml) dataset. There is a reference CSV for each model type that contains a reference to each individual model name, the artifacts file for that model, and which batch archive file contains that project archive for that model.
|
| 69 |
|
| 70 |
**PLEASE NOTE:** The dataset is currently incomplete, and is in the process of being cleaned and uploaded.
|
| 71 |
|
|
@@ -160,7 +160,7 @@ The resulting projects, reports, logs, and a JSON file containing the resource/l
|
|
| 160 |
The data, excluding the projects and logs, were then further processed into a collection of JSON files, distributed alongside this paper and described below.
|
| 161 |
|
| 162 |
- **Repository:** [fastmachinelearning/wa-hls4ml-paper](https://github.com/fastmachinelearning/wa-hls4ml-paper)
|
| 163 |
-
- **Paper [optional]:**
|
| 164 |
|
| 165 |
---
|
| 166 |
|
|
@@ -181,18 +181,7 @@ As this dataset is generated using the hls4ml and Vivado/Vitis tools, it should
|
|
| 181 |
|
| 182 |
## Dataset Structure
|
| 183 |
|
| 184 |
-
|
| 185 |
-
Each sample contains the model architecture, hls4ml conversion parameters, the latency and resource usage numbers for that network post-logic synthesis, and associated metadata.
|
| 186 |
-
In addition to the training, validation, and test sets, the dataset also includes 887 samples representing the successful logic synthesis of the exemplar models with varying conversion parameters.
|
| 187 |
-
The dataset as a whole is split, distributed, and intended to be used as follows:
|
| 188 |
-
|
| 189 |
-
- **Training set**: The set of 478,220 samples intended to be used for training a given estimator.
|
| 190 |
-
- **Validation set**: The set of 102,472 samples intended to be used during training for validation purposes.
|
| 191 |
-
- **Test set**: The set of 102,484 samples intended to be used for testing and generating results for a given estimator.
|
| 192 |
-
- **Exemplar test set**: The set of 887 samples, comprising the models described in the benchmark, intended to be used for testing and generating results for a given estimator.
|
| 193 |
-
|
| 194 |
-
Within each subset, excluding the exemplar test set, the data is further grouped as follows.
|
| 195 |
-
These categories explain the composition of our dataset but have no bearing on how a given estimator should be trained.
|
| 196 |
|
| 197 |
- **2_20 (rule4ml)**: The updated rule4ml dataset, containing fully-connected neural networks that were randomly generated with layer counts between 2 and 20 layers, using hls4ml resource and latency strategies.
|
| 198 |
- **2_layer**: A subset containing 2-layer deep fully-connected neural networks generated via a grid search using hls4ml resource and io_parallel strategies.
|
|
@@ -253,7 +242,7 @@ We intend to continuously update this dataset, addressing this imbalance over ti
|
|
| 253 |
|
| 254 |
Appropriate measures should be taken to weight the data to account for the dataset imbalance at training time.
|
| 255 |
|
| 256 |
-
## Citation
|
| 257 |
|
| 258 |
Paper currently in review.
|
| 259 |
|
|
|
|
| 65 |
|
| 66 |
The wa-hls4ml projects dataset, comprized of the Vivado/Vitis projects of neural networks converted into HLS Code via hls4ml. Projects are complete, and include all logs, HLS Code, VHDL Code, Intermediete Representations, and source keras models.
|
| 67 |
|
| 68 |
+
This is a companion dataset to the [wa-hls4ml](https://huggingface.co/datasets/fastmachinelearning/wa-hls4ml) dataset. There is a reference CSV for each model type that contains a reference to each individual model name, the artifacts file for that model, and which batch archive file contains that project archive for that model.
|
| 69 |
|
| 70 |
**PLEASE NOTE:** The dataset is currently incomplete, and is in the process of being cleaned and uploaded.
|
| 71 |
|
|
|
|
| 160 |
The data, excluding the projects and logs, were then further processed into a collection of JSON files, distributed alongside this paper and described below.
|
| 161 |
|
| 162 |
- **Repository:** [fastmachinelearning/wa-hls4ml-paper](https://github.com/fastmachinelearning/wa-hls4ml-paper)
|
| 163 |
+
- **Paper [optional]:** In Review
|
| 164 |
|
| 165 |
---
|
| 166 |
|
|
|
|
| 181 |
|
| 182 |
## Dataset Structure
|
| 183 |
|
| 184 |
+
Within each subset, excluding the exemplar test set, the data is grouped as follows.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 185 |
|
| 186 |
- **2_20 (rule4ml)**: The updated rule4ml dataset, containing fully-connected neural networks that were randomly generated with layer counts between 2 and 20 layers, using hls4ml resource and latency strategies.
|
| 187 |
- **2_layer**: A subset containing 2-layer deep fully-connected neural networks generated via a grid search using hls4ml resource and io_parallel strategies.
|
|
|
|
| 242 |
|
| 243 |
Appropriate measures should be taken to weight the data to account for the dataset imbalance at training time.
|
| 244 |
|
| 245 |
+
## Citation
|
| 246 |
|
| 247 |
Paper currently in review.
|
| 248 |
|