loubnabnl HF Staff commited on
Commit
7dca969
·
1 Parent(s): 6d0e573

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +8 -8
README.md CHANGED
@@ -16,7 +16,7 @@ pinned: false
16
 
17
  <li>
18
  <p>
19
- Interactive blog where we compare different code models and explain how they are trained and evaluated: <a
20
  href="https://huggingface.co/spaces/loubnabnl/code-generation-models"
21
  class="underline">Code generation with 🤗</a
22
  >
@@ -25,19 +25,19 @@ pinned: false
25
 
26
  <li>
27
  <p>
28
- Spaces: code generation with: <a href="https://huggingface.co/codeparrot/codeparrot">CodeParrot</a> (1.5B), <a href="https://huggingface.co/facebook/incoder-6B">InCoder</a> (6B) and <a href="https://github.com/salesforce/CodeGen">CodeGen</a> (6B)
29
  </p>
30
  </li>
31
 
32
  <li>Models: CodeParrot (1.5B) and CodeParrot-small (110M), each repo has different ongoing experiments in the branches.</li>
33
 
34
  <li>Datasets:<ul>
35
- <li><a href="https://huggingface.co/datasets/codeparrot/codeparrot-clean">codeparrot-clean</a>, dataset on which we trained and evaluated CodeParrot, the splits are available under <a href="https://huggingface.co/datasets/codeparrot/codeparrot-clean-train">codeparrot-clean-train</a> and <a href="https://huggingface.co/datasets/codeparrot/codeparrot-clean-valid">codeparrot-clean-valid</a>.</li>
36
- <li>A more filtered version of codeparrot-clean under <a href="https://huggingface.co/datasets/codeparrot/codeparrot-train-more-filtering">codeparrot-train-more-filtering</a> and <a href="https://huggingface.co/datasets/codeparrot/codeparrot-valid-more-filtering">codeparrot-train-more-filtering</a>.</li>
37
- <li>CodeParrot dataset after near deduplication since initially only exact match deduplication was performed, it&#39;s available under <a href="https://huggingface.co/datasets/codeparrot/codeparrot-train-near-deduplication">codeparrot-train-near-deduplication</a> and <a href="https://huggingface.co/datasets/codeparrot/codeparrot-valid-near-deduplication">codeparrot-train-near-deduplication</a>.</li>
38
- <li><a href="https://huggingface.co/datasets/codeparrot/github-code">GitHub-Code</a>, a 1TB dataset of 32 programming languages with 60 from GitHub files.</li>
39
- <li><a href="https://huggingface.co/datasets/codeparrot/github-jupyter">GitHub-Jupyter</a>, a 16.3GB dataset of Jupyter Notebooks from BigQuery GitHub.</li>
40
- <li><a href="https://huggingface.co/datasets/codeparrot/apps">APPS</a>, a benchmark for code generation with 10000 problems.</li>
41
  </ul>
42
  </li>
43
  </ul>
 
16
 
17
  <li>
18
  <p>
19
+ Interactive blog: where we compare different code models and explain how they are trained and evaluated <a
20
  href="https://huggingface.co/spaces/loubnabnl/code-generation-models"
21
  class="underline">Code generation with 🤗</a
22
  >
 
25
 
26
  <li>
27
  <p>
28
+ Spaces: code generation with: <a ref="https://huggingface.co/codeparrot/codeparrot" class="underline">CodeParrot (1.5B)</a>, <a href="https://huggingface.co/facebook/incoder-6B" class="underline">InCoder</a> (6B) and <a href="https://github.com/salesforce/CodeGen" class="underline">CodeGen</a> (6B)
29
  </p>
30
  </li>
31
 
32
  <li>Models: CodeParrot (1.5B) and CodeParrot-small (110M), each repo has different ongoing experiments in the branches.</li>
33
 
34
  <li>Datasets:<ul>
35
+ <li><a href="https://huggingface.co/datasets/codeparrot/codeparrot-clean" class="underline">codeparrot-clean</a>, dataset on which we trained and evaluated CodeParrot, the splits are available under <a href="https://huggingface.co/datasets/codeparrot/codeparrot-clean-train" class="underline">codeparrot-clean-train</a> and <a href="https://huggingface.co/datasets/codeparrot/codeparrot-clean-valid" class="underline">codeparrot-clean-valid</a>.</li>
36
+ <li>A more filtered version of codeparrot-clean under <a href="https://huggingface.co/datasets/codeparrot/codeparrot-train-more-filtering" class="underline">codeparrot-train-more-filtering</a> and <a href="https://huggingface.co/datasets/codeparrot/codeparrot-valid-more-filtering" class="underline">codeparrot-train-more-filtering</a>.</li>
37
+ <li>CodeParrot dataset after near deduplication since initially only exact match deduplication was performed, it's available under <a href="https://huggingface.co/datasets/codeparrot/codeparrot-train-near-deduplication" class="underline">codeparrot-train-near-deduplication</a> and <a href="https://huggingface.co/datasets/codeparrot/codeparrot-valid-near-deduplication" class="underline">codeparrot-train-near-deduplication</a>.</li>
38
+ <li><a href="https://huggingface.co/datasets/codeparrot/github-code" class="underline">GitHub-Code</a>, a 1TB dataset of 32 programming languages with 60 from GitHub files.</li>
39
+ <li><a href="https://huggingface.co/datasets/codeparrot/github-jupyter" class="underline">GitHub-Jupyter</a>, a 16.3GB dataset of Jupyter Notebooks from BigQuery GitHub.</li>
40
+ <li><a href="https://huggingface.co/datasets/codeparrot/apps" class="underline">APPS</a>, a benchmark for code generation with 10000 problems.</li>
41
  </ul>
42
  </li>
43
  </ul>