zdaniar commited on
Commit
ec9027d
·
1 Parent(s): 924d201

Training in progress epoch 0

Browse files
README.md CHANGED
@@ -1,18 +1,22 @@
1
  ---
2
  license: apache-2.0
3
  tags:
4
- - generated_from_trainer
5
  model-index:
6
- - name: my_awesome_eli5_mlm_model
7
  results: []
8
  ---
9
 
10
- <!-- This model card has been generated automatically according to the information the Trainer had access to. You
11
- should probably proofread and complete it, then remove this comment. -->
12
 
13
- # my_awesome_eli5_mlm_model
14
 
15
- This model is a fine-tuned version of [distilroberta-base](https://huggingface.co/distilroberta-base) on the None dataset.
 
 
 
 
16
 
17
  ## Model description
18
 
@@ -31,17 +35,19 @@ More information needed
31
  ### Training hyperparameters
32
 
33
  The following hyperparameters were used during training:
34
- - learning_rate: 2e-05
35
- - train_batch_size: 8
36
- - eval_batch_size: 8
37
- - seed: 42
38
- - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
39
- - lr_scheduler_type: linear
40
- - num_epochs: 3
 
 
41
 
42
  ### Framework versions
43
 
44
  - Transformers 4.26.1
45
- - Pytorch 1.13.1+cu116
46
  - Datasets 2.10.1
47
  - Tokenizers 0.13.2
 
1
  ---
2
  license: apache-2.0
3
  tags:
4
+ - generated_from_keras_callback
5
  model-index:
6
+ - name: zdaniar/my_awesome_eli5_mlm_model
7
  results: []
8
  ---
9
 
10
+ <!-- This model card has been generated automatically according to the information Keras had access to. You should
11
+ probably proofread and complete it, then remove this comment. -->
12
 
13
+ # zdaniar/my_awesome_eli5_mlm_model
14
 
15
+ This model is a fine-tuned version of [distilroberta-base](https://huggingface.co/distilroberta-base) on an unknown dataset.
16
+ It achieves the following results on the evaluation set:
17
+ - Train Loss: 2.0455
18
+ - Validation Loss: 1.8457
19
+ - Epoch: 0
20
 
21
  ## Model description
22
 
 
35
  ### Training hyperparameters
36
 
37
  The following hyperparameters were used during training:
38
+ - optimizer: {'name': 'AdamWeightDecay', 'learning_rate': 2e-05, 'decay': 0.0, 'beta_1': 0.9, 'beta_2': 0.999, 'epsilon': 1e-07, 'amsgrad': False, 'weight_decay_rate': 0.01}
39
+ - training_precision: float32
40
+
41
+ ### Training results
42
+
43
+ | Train Loss | Validation Loss | Epoch |
44
+ |:----------:|:---------------:|:-----:|
45
+ | 2.0455 | 1.8457 | 0 |
46
+
47
 
48
  ### Framework versions
49
 
50
  - Transformers 4.26.1
51
+ - TensorFlow 2.11.0
52
  - Datasets 2.10.1
53
  - Tokenizers 0.13.2
config.json CHANGED
@@ -19,7 +19,6 @@
19
  "num_hidden_layers": 6,
20
  "pad_token_id": 1,
21
  "position_embedding_type": "absolute",
22
- "torch_dtype": "float32",
23
  "transformers_version": "4.26.1",
24
  "type_vocab_size": 1,
25
  "use_cache": true,
 
19
  "num_hidden_layers": 6,
20
  "pad_token_id": 1,
21
  "position_embedding_type": "absolute",
 
22
  "transformers_version": "4.26.1",
23
  "type_vocab_size": 1,
24
  "use_cache": true,
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
runs/Mar09_10-26-04_84cc66ed7119/events.out.tfevents.1678357570.84cc66ed7119.125.2 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cb173a36a72e09224223a4eebe9b9270afc9fbe58352b331391a9c584762c59e
3
- size 40
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:436b20e3cab687e527c9f3d5b7865763e6b177ecf2f8391312b4f157d9025f56
3
+ size 3840
special_tokens_map.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": "<s>",
3
+ "cls_token": "<s>",
4
+ "eos_token": "</s>",
5
+ "mask_token": {
6
+ "content": "<mask>",
7
+ "lstrip": true,
8
+ "normalized": false,
9
+ "rstrip": false,
10
+ "single_word": false
11
+ },
12
+ "pad_token": "</s>",
13
+ "sep_token": "</s>",
14
+ "unk_token": "<unk>"
15
+ }
tf_model.h5 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6dc86138eaa01945922003b5fbcee1fe0f900da151fed1775fc627a0992969d6
3
+ size 484838716
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,16 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": false,
3
+ "bos_token": "<s>",
4
+ "cls_token": "<s>",
5
+ "eos_token": "</s>",
6
+ "errors": "replace",
7
+ "mask_token": "<mask>",
8
+ "model_max_length": 512,
9
+ "name_or_path": "distilroberta-base",
10
+ "pad_token": "<pad>",
11
+ "sep_token": "</s>",
12
+ "special_tokens_map_file": null,
13
+ "tokenizer_class": "RobertaTokenizer",
14
+ "trim_offsets": true,
15
+ "unk_token": "<unk>"
16
+ }
vocab.json ADDED
The diff for this file is too large to render. See raw diff