update new info
Browse files- README.md +19 -17
- config.json +1 -1
- tokenizer_config.json +6 -6
README.md
CHANGED
@@ -1,14 +1,12 @@
|
|
1 |
---
|
2 |
license: apache-2.0
|
3 |
-
datasets:
|
4 |
-
- togethercomputer/RedPajama-Data-1T
|
5 |
language:
|
6 |
- en
|
7 |
-
|
8 |
-
-
|
9 |
pipeline_tag: text-generation
|
10 |
---
|
11 |
-
#
|
12 |
|
13 |
<!-- Provide a quick summary of what the model is/does. -->
|
14 |
|
@@ -22,20 +20,18 @@ This modelcard aims to be a base template for new models. It has been generated
|
|
22 |
|
23 |
|
24 |
|
25 |
-
- **Developed by:**
|
26 |
-
- **
|
27 |
-
- **
|
28 |
-
- **
|
29 |
-
- **
|
30 |
-
- **Finetuned from model [optional]:** [More Information Needed]
|
31 |
|
32 |
-
### Model Sources
|
33 |
|
34 |
<!-- Provide the basic links for the model. -->
|
35 |
|
36 |
-
- **Repository:**
|
37 |
-
- **
|
38 |
-
- **Demo [optional]:** [More Information Needed]
|
39 |
|
40 |
## Uses
|
41 |
|
@@ -75,7 +71,13 @@ Users (both direct and downstream) should be made aware of the risks, biases and
|
|
75 |
|
76 |
Use the code below to get started with the model.
|
77 |
|
78 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
79 |
|
80 |
## Training Details
|
81 |
|
@@ -83,7 +85,7 @@ Use the code below to get started with the model.
|
|
83 |
|
84 |
<!-- This should link to a Data Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
|
85 |
|
86 |
-
|
87 |
|
88 |
### Training Procedure
|
89 |
|
|
|
1 |
---
|
2 |
license: apache-2.0
|
|
|
|
|
3 |
language:
|
4 |
- en
|
5 |
+
tags:
|
6 |
+
- hawk
|
7 |
pipeline_tag: text-generation
|
8 |
---
|
9 |
+
# Hawk-demo
|
10 |
|
11 |
<!-- Provide a quick summary of what the model is/does. -->
|
12 |
|
|
|
20 |
|
21 |
|
22 |
|
23 |
+
- **Developed by:** Rexopia
|
24 |
+
- **Reach me:** [email protected]
|
25 |
+
- **Language(s) (NLP):** English
|
26 |
+
- **License:** Apache license 2.0
|
27 |
+
- **Pretrained model [optional]:** True
|
|
|
28 |
|
29 |
+
### Model Sources
|
30 |
|
31 |
<!-- Provide the basic links for the model. -->
|
32 |
|
33 |
+
- **Github Repository:** Coming soon
|
34 |
+
- **Demo version:** True
|
|
|
35 |
|
36 |
## Uses
|
37 |
|
|
|
71 |
|
72 |
Use the code below to get started with the model.
|
73 |
|
74 |
+
```python
|
75 |
+
from transformers import AutoModelForCausalLM, AutoTokenizer
|
76 |
+
|
77 |
+
tokenizer = AutoTokenizer.from_pretrained("Rexopia/HawkLM-demo", trust_remote_code=True)
|
78 |
+
|
79 |
+
model = AutoModelForCausalLM.from_pretrained("Rexopia/HawkLM-demo", device_map="auto", trust_remote_code=True)
|
80 |
+
```
|
81 |
|
82 |
## Training Details
|
83 |
|
|
|
85 |
|
86 |
<!-- This should link to a Data Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
|
87 |
|
88 |
+
We sampled from Redpajama 1T datasets without any Arxiv and GitHub tags.
|
89 |
|
90 |
### Training Procedure
|
91 |
|
config.json
CHANGED
@@ -12,7 +12,7 @@
|
|
12 |
"embd_pdrop": 0.0,
|
13 |
"eos_token_id": 65535,
|
14 |
"initializer_range": 0.02,
|
15 |
-
"layer_norm_epsilon": 1e-
|
16 |
"model_type": "hawk",
|
17 |
"n_embd": 1024,
|
18 |
"n_head": 16,
|
|
|
12 |
"embd_pdrop": 0.0,
|
13 |
"eos_token_id": 65535,
|
14 |
"initializer_range": 0.02,
|
15 |
+
"layer_norm_epsilon": 1e-06,
|
16 |
"model_type": "hawk",
|
17 |
"n_embd": 1024,
|
18 |
"n_head": 16,
|
tokenizer_config.json
CHANGED
@@ -1,6 +1,12 @@
|
|
1 |
{
|
2 |
"add_bos_token": false,
|
3 |
"add_prefix_space": false,
|
|
|
|
|
|
|
|
|
|
|
|
|
4 |
"bos_token": {
|
5 |
"__type": "AddedToken",
|
6 |
"content": "<s>",
|
@@ -36,11 +42,5 @@
|
|
36 |
"normalized": true,
|
37 |
"rstrip": false,
|
38 |
"single_word": false
|
39 |
-
},
|
40 |
-
"auto_map": {
|
41 |
-
"AutoTokenizer": [
|
42 |
-
"tokenization_hawk.HawkTokenizer",
|
43 |
-
null
|
44 |
-
]
|
45 |
}
|
46 |
}
|
|
|
1 |
{
|
2 |
"add_bos_token": false,
|
3 |
"add_prefix_space": false,
|
4 |
+
"auto_map": {
|
5 |
+
"AutoTokenizer": [
|
6 |
+
"tokenization_hawk.HawkTokenizer",
|
7 |
+
null
|
8 |
+
]
|
9 |
+
},
|
10 |
"bos_token": {
|
11 |
"__type": "AddedToken",
|
12 |
"content": "<s>",
|
|
|
42 |
"normalized": true,
|
43 |
"rstrip": false,
|
44 |
"single_word": false
|
|
|
|
|
|
|
|
|
|
|
|
|
45 |
}
|
46 |
}
|