Rexopia commited on
Commit
bec1dde
·
1 Parent(s): 6d0b7dd

update new info

Browse files
Files changed (3) hide show
  1. README.md +19 -17
  2. config.json +1 -1
  3. tokenizer_config.json +6 -6
README.md CHANGED
@@ -1,14 +1,12 @@
1
  ---
2
  license: apache-2.0
3
- datasets:
4
- - togethercomputer/RedPajama-Data-1T
5
  language:
6
  - en
7
- metrics:
8
- - accuracy
9
  pipeline_tag: text-generation
10
  ---
11
- # Model Card for Model ID
12
 
13
  <!-- Provide a quick summary of what the model is/does. -->
14
 
@@ -22,20 +20,18 @@ This modelcard aims to be a base template for new models. It has been generated
22
 
23
 
24
 
25
- - **Developed by:** [More Information Needed]
26
- - **Shared by [optional]:** [More Information Needed]
27
- - **Model type:** [More Information Needed]
28
- - **Language(s) (NLP):** [More Information Needed]
29
- - **License:** [More Information Needed]
30
- - **Finetuned from model [optional]:** [More Information Needed]
31
 
32
- ### Model Sources [optional]
33
 
34
  <!-- Provide the basic links for the model. -->
35
 
36
- - **Repository:** [More Information Needed]
37
- - **Paper [optional]:** [More Information Needed]
38
- - **Demo [optional]:** [More Information Needed]
39
 
40
  ## Uses
41
 
@@ -75,7 +71,13 @@ Users (both direct and downstream) should be made aware of the risks, biases and
75
 
76
  Use the code below to get started with the model.
77
 
78
- [More Information Needed]
 
 
 
 
 
 
79
 
80
  ## Training Details
81
 
@@ -83,7 +85,7 @@ Use the code below to get started with the model.
83
 
84
  <!-- This should link to a Data Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
85
 
86
- [More Information Needed]
87
 
88
  ### Training Procedure
89
 
 
1
  ---
2
  license: apache-2.0
 
 
3
  language:
4
  - en
5
+ tags:
6
+ - hawk
7
  pipeline_tag: text-generation
8
  ---
9
+ # Hawk-demo
10
 
11
  <!-- Provide a quick summary of what the model is/does. -->
12
 
 
20
 
21
 
22
 
23
+ - **Developed by:** Rexopia
24
+ - **Reach me:** [email protected]
25
+ - **Language(s) (NLP):** English
26
+ - **License:** Apache license 2.0
27
+ - **Pretrained model [optional]:** True
 
28
 
29
+ ### Model Sources
30
 
31
  <!-- Provide the basic links for the model. -->
32
 
33
+ - **Github Repository:** Coming soon
34
+ - **Demo version:** True
 
35
 
36
  ## Uses
37
 
 
71
 
72
  Use the code below to get started with the model.
73
 
74
+ ```python
75
+ from transformers import AutoModelForCausalLM, AutoTokenizer
76
+
77
+ tokenizer = AutoTokenizer.from_pretrained("Rexopia/HawkLM-demo", trust_remote_code=True)
78
+
79
+ model = AutoModelForCausalLM.from_pretrained("Rexopia/HawkLM-demo", device_map="auto", trust_remote_code=True)
80
+ ```
81
 
82
  ## Training Details
83
 
 
85
 
86
  <!-- This should link to a Data Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
87
 
88
+ We sampled from Redpajama 1T datasets without any Arxiv and GitHub tags.
89
 
90
  ### Training Procedure
91
 
config.json CHANGED
@@ -12,7 +12,7 @@
12
  "embd_pdrop": 0.0,
13
  "eos_token_id": 65535,
14
  "initializer_range": 0.02,
15
- "layer_norm_epsilon": 1e-05,
16
  "model_type": "hawk",
17
  "n_embd": 1024,
18
  "n_head": 16,
 
12
  "embd_pdrop": 0.0,
13
  "eos_token_id": 65535,
14
  "initializer_range": 0.02,
15
+ "layer_norm_epsilon": 1e-06,
16
  "model_type": "hawk",
17
  "n_embd": 1024,
18
  "n_head": 16,
tokenizer_config.json CHANGED
@@ -1,6 +1,12 @@
1
  {
2
  "add_bos_token": false,
3
  "add_prefix_space": false,
 
 
 
 
 
 
4
  "bos_token": {
5
  "__type": "AddedToken",
6
  "content": "<s>",
@@ -36,11 +42,5 @@
36
  "normalized": true,
37
  "rstrip": false,
38
  "single_word": false
39
- },
40
- "auto_map": {
41
- "AutoTokenizer": [
42
- "tokenization_hawk.HawkTokenizer",
43
- null
44
- ]
45
  }
46
  }
 
1
  {
2
  "add_bos_token": false,
3
  "add_prefix_space": false,
4
+ "auto_map": {
5
+ "AutoTokenizer": [
6
+ "tokenization_hawk.HawkTokenizer",
7
+ null
8
+ ]
9
+ },
10
  "bos_token": {
11
  "__type": "AddedToken",
12
  "content": "<s>",
 
42
  "normalized": true,
43
  "rstrip": false,
44
  "single_word": false
 
 
 
 
 
 
45
  }
46
  }