madatnlp commited on
Commit
edc41c8
·
1 Parent(s): fcb5398

Training in progress epoch 0

Browse files
Files changed (3) hide show
  1. README.md +3 -3
  2. config.json +21 -0
  3. tf_model.h5 +2 -2
README.md CHANGED
@@ -13,8 +13,8 @@ probably proofread and complete it, then remove this comment. -->
13
 
14
  This model is a fine-tuned version of [klue/roberta-large](https://huggingface.co/klue/roberta-large) on an unknown dataset.
15
  It achieves the following results on the evaluation set:
16
- - Train Loss: nan
17
- - Validation Loss: nan
18
  - Epoch: 0
19
 
20
  ## Model description
@@ -41,7 +41,7 @@ The following hyperparameters were used during training:
41
 
42
  | Train Loss | Validation Loss | Epoch |
43
  |:----------:|:---------------:|:-----:|
44
- | nan | nan | 0 |
45
 
46
 
47
  ### Framework versions
 
13
 
14
  This model is a fine-tuned version of [klue/roberta-large](https://huggingface.co/klue/roberta-large) on an unknown dataset.
15
  It achieves the following results on the evaluation set:
16
+ - Train Loss: 2.1016
17
+ - Validation Loss: 1.8401
18
  - Epoch: 0
19
 
20
  ## Model description
 
41
 
42
  | Train Loss | Validation Loss | Epoch |
43
  |:----------:|:---------------:|:-----:|
44
+ | 2.1016 | 1.8401 | 0 |
45
 
46
 
47
  ### Framework versions
config.json CHANGED
@@ -11,8 +11,28 @@
11
  "hidden_act": "gelu",
12
  "hidden_dropout_prob": 0.1,
13
  "hidden_size": 1024,
 
 
 
 
 
 
 
 
 
 
14
  "initializer_range": 0.02,
15
  "intermediate_size": 4096,
 
 
 
 
 
 
 
 
 
 
16
  "layer_norm_eps": 1e-05,
17
  "max_position_embeddings": 514,
18
  "model_type": "roberta",
@@ -20,6 +40,7 @@
20
  "num_hidden_layers": 24,
21
  "pad_token_id": 1,
22
  "position_embedding_type": "absolute",
 
23
  "tokenizer_class": "BertTokenizer",
24
  "transformers_version": "4.19.2",
25
  "type_vocab_size": 1,
 
11
  "hidden_act": "gelu",
12
  "hidden_dropout_prob": 0.1,
13
  "hidden_size": 1024,
14
+ "id2label": {
15
+ "0": "LABEL_0",
16
+ "1": "LABEL_1",
17
+ "2": "LABEL_2",
18
+ "3": "LABEL_3",
19
+ "4": "LABEL_4",
20
+ "5": "LABEL_5",
21
+ "6": "LABEL_6",
22
+ "7": "LABEL_7"
23
+ },
24
  "initializer_range": 0.02,
25
  "intermediate_size": 4096,
26
+ "label2id": {
27
+ "LABEL_0": 0,
28
+ "LABEL_1": 1,
29
+ "LABEL_2": 2,
30
+ "LABEL_3": 3,
31
+ "LABEL_4": 4,
32
+ "LABEL_5": 5,
33
+ "LABEL_6": 6,
34
+ "LABEL_7": 7
35
+ },
36
  "layer_norm_eps": 1e-05,
37
  "max_position_embeddings": 514,
38
  "model_type": "roberta",
 
40
  "num_hidden_layers": 24,
41
  "pad_token_id": 1,
42
  "position_embedding_type": "absolute",
43
+ "problem_type": "multi_label_classification",
44
  "tokenizer_class": "BertTokenizer",
45
  "transformers_version": "4.19.2",
46
  "type_vocab_size": 1,
tf_model.h5 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7757df9898b81b2ba94921e0143a3bb190724a71231680ba725124cdaf9473e7
3
- size 1347187352
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:367263510d918f93df0dabbe849d017daab6fc2b7f962d858174d64f2c048ce8
3
+ size 1347211928