rajabmondal commited on
Commit
da27032
·
verified ·
1 Parent(s): efadb81

updated model summary

Browse files
Files changed (1) hide show
  1. README.md +3 -3
README.md CHANGED
@@ -26,13 +26,13 @@ widget:
26
 
27
  ## Model Summary
28
 
29
- The JavaCoder models are 1B parameter models trained on 80+ programming languages from [The Stack (v1.2)](https://huggingface.co/datasets/bigcode/the-stack), with opt-out requests excluded. The model uses [Multi Query Attention](https://arxiv.org/abs/1911.02150), [a context window of 8192 tokens](https://arxiv.org/abs/2205.14135), and was trained using the [Fill-in-the-Middle objective](https://arxiv.org/abs/2207.14255) on 1 trillion tokens.
30
 
31
- - **Repository:**
32
  - **Project Website:**
33
  - **Paper:**
34
  - **Point of Contact:**
35
- - **Languages:** 80+ Programming languages
36
 
37
  ## Use
38
 
 
26
 
27
  ## Model Summary
28
 
29
+ The Narrow Transformer (NT) model NT-Java-1.1B is an open-source specialized code model built on StarCoderBase, designed for code completion tasks in Java programming. The model is a decoder-only transformer with Multi-Query-Attention and learned absolute positional embeddings and was finetuned for Java subset of the training data (starcoderdata) which is ~22B tokens and with a context of 8192 tokens.
30
 
31
+ - **Repository:** [bigcode/Megatron-LM](https://github.com/bigcode-project/Megatron-LM)
32
  - **Project Website:**
33
  - **Paper:**
34
  - **Point of Contact:**
35
+ - **Languages:** Java
36
 
37
  ## Use
38