taicheng commited on
Commit
42a32be
·
verified ·
1 Parent(s): 05aeb7f

Model save

Browse files
README.md CHANGED
@@ -3,15 +3,10 @@ library_name: transformers
3
  license: apache-2.0
4
  base_model: alignment-handbook/zephyr-7b-sft-full
5
  tags:
6
- - alignment-handbook
7
- - trl
8
- - dpo
9
- - generated_from_trainer
10
  - trl
11
  - dpo
 
12
  - generated_from_trainer
13
- datasets:
14
- - HuggingFaceH4/ultrafeedback_binarized
15
  model-index:
16
  - name: zephyr-7b-align-scan-0.0-0.2-polynomial-3
17
  results: []
@@ -22,17 +17,17 @@ should probably proofread and complete it, then remove this comment. -->
22
 
23
  # zephyr-7b-align-scan-0.0-0.2-polynomial-3
24
 
25
- This model is a fine-tuned version of [alignment-handbook/zephyr-7b-sft-full](https://huggingface.co/alignment-handbook/zephyr-7b-sft-full) on the HuggingFaceH4/ultrafeedback_binarized dataset.
26
  It achieves the following results on the evaluation set:
27
- - Loss: 0.7011
28
- - Rewards/chosen: -1.4864
29
- - Rewards/rejected: -2.0925
 
 
30
  - Rewards/accuracies: 0.3313
31
- - Rewards/margins: 0.6061
32
- - Logps/rejected: -89.9617
33
- - Logps/chosen: -80.7659
34
- - Logits/rejected: -2.4878
35
- - Logits/chosen: -2.5051
36
 
37
  ## Model description
38
 
@@ -51,7 +46,7 @@ More information needed
51
  ### Training hyperparameters
52
 
53
  The following hyperparameters were used during training:
54
- - learning_rate: 2.1445950529764786e-07
55
  - train_batch_size: 8
56
  - eval_batch_size: 8
57
  - seed: 42
@@ -67,10 +62,10 @@ The following hyperparameters were used during training:
67
 
68
  ### Training results
69
 
70
- | Training Loss | Epoch | Step | Validation Loss | Rewards/chosen | Rewards/rejected | Rewards/accuracies | Rewards/margins | Logps/rejected | Logps/chosen | Logits/rejected | Logits/chosen |
71
- |:-------------:|:------:|:----:|:---------------:|:--------------:|:----------------:|:------------------:|:---------------:|:--------------:|:------------:|:---------------:|:-------------:|
72
- | 0.5502 | 1.0417 | 100 | 0.6403 | 0.2781 | 0.0281 | 0.3254 | 0.2500 | -81.0099 | -73.3173 | -2.5124 | -2.5290 |
73
- | 0.377 | 2.0833 | 200 | 0.6467 | -0.1914 | -0.6247 | 0.3313 | 0.4333 | -83.7654 | -75.2992 | -2.5078 | -2.5251 |
74
 
75
 
76
  ### Framework versions
 
3
  license: apache-2.0
4
  base_model: alignment-handbook/zephyr-7b-sft-full
5
  tags:
 
 
 
 
6
  - trl
7
  - dpo
8
+ - alignment-handbook
9
  - generated_from_trainer
 
 
10
  model-index:
11
  - name: zephyr-7b-align-scan-0.0-0.2-polynomial-3
12
  results: []
 
17
 
18
  # zephyr-7b-align-scan-0.0-0.2-polynomial-3
19
 
20
+ This model is a fine-tuned version of [alignment-handbook/zephyr-7b-sft-full](https://huggingface.co/alignment-handbook/zephyr-7b-sft-full) on an unknown dataset.
21
  It achieves the following results on the evaluation set:
22
+ - Logits/chosen: -2.5251
23
+ - Logits/rejected: -2.5078
24
+ - Logps/chosen: -75.2992
25
+ - Logps/rejected: -83.7654
26
+ - Loss: 0.6467
27
  - Rewards/accuracies: 0.3313
28
+ - Rewards/chosen: -0.1914
29
+ - Rewards/margins: 0.4333
30
+ - Rewards/rejected: -0.6247
 
 
31
 
32
  ## Model description
33
 
 
46
  ### Training hyperparameters
47
 
48
  The following hyperparameters were used during training:
49
+ - learning_rate: 7.526744872300726e-07
50
  - train_batch_size: 8
51
  - eval_batch_size: 8
52
  - seed: 42
 
62
 
63
  ### Training results
64
 
65
+ | Training Loss | Epoch | Step | Logits/chosen | Logits/rejected | Logps/chosen | Logps/rejected | Validation Loss | Rewards/accuracies | Rewards/chosen | Rewards/margins | Rewards/rejected |
66
+ |:-------------:|:------:|:----:|:-------------:|:---------------:|:------------:|:--------------:|:---------------:|:------------------:|:--------------:|:---------------:|:----------------:|
67
+ | 0.5502 | 1.0417 | 100 | -2.5290 | -2.5124 | -73.3173 | -81.0099 | 0.6403 | 0.3254 | 0.2781 | 0.2500 | 0.0281 |
68
+ | 0.377 | 2.0833 | 200 | -2.5251 | -2.5078 | -75.2992 | -83.7654 | 0.6467 | 0.3313 | -0.1914 | 0.4333 | -0.6247 |
69
 
70
 
71
  ### Framework versions
all_results.json CHANGED
@@ -14,9 +14,9 @@
14
  "eval_samples_per_second": 17.543,
15
  "eval_steps_per_second": 0.553,
16
  "total_flos": 0.0,
17
- "train_loss": 0.4920133708251847,
18
- "train_runtime": 3526.2726,
19
  "train_samples": 6113,
20
- "train_samples_per_second": 5.201,
21
- "train_steps_per_second": 0.082
22
  }
 
14
  "eval_samples_per_second": 17.543,
15
  "eval_steps_per_second": 0.553,
16
  "total_flos": 0.0,
17
+ "train_loss": 0.0,
18
+ "train_runtime": 0.0237,
19
  "train_samples": 6113,
20
+ "train_samples_per_second": 774670.33,
21
+ "train_steps_per_second": 12165.606
22
  }
config.json CHANGED
@@ -22,6 +22,6 @@
22
  "tie_word_embeddings": false,
23
  "torch_dtype": "bfloat16",
24
  "transformers_version": "4.44.2",
25
- "use_cache": true,
26
  "vocab_size": 32000
27
  }
 
22
  "tie_word_embeddings": false,
23
  "torch_dtype": "bfloat16",
24
  "transformers_version": "4.44.2",
25
+ "use_cache": false,
26
  "vocab_size": 32000
27
  }
runs/Sep24_14-18-45_qa-a100-004.crc.nd.edu/events.out.tfevents.1727202087.qa-a100-004.crc.nd.edu.1892028.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:54068d221111f0d30718ebac97582b857e36f8cc2c89291b53b7f207ece8f67b
3
+ size 6828
train_results.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "epoch": 3.0,
3
  "total_flos": 0.0,
4
- "train_loss": 0.4920133708251847,
5
- "train_runtime": 3526.2726,
6
  "train_samples": 6113,
7
- "train_samples_per_second": 5.201,
8
- "train_steps_per_second": 0.082
9
  }
 
1
  {
2
  "epoch": 3.0,
3
  "total_flos": 0.0,
4
+ "train_loss": 0.0,
5
+ "train_runtime": 0.0237,
6
  "train_samples": 6113,
7
+ "train_samples_per_second": 774670.33,
8
+ "train_steps_per_second": 12165.606
9
  }
trainer_state.json CHANGED
@@ -479,10 +479,10 @@
479
  "epoch": 3.0,
480
  "step": 288,
481
  "total_flos": 0.0,
482
- "train_loss": 0.4920133708251847,
483
- "train_runtime": 3526.2726,
484
- "train_samples_per_second": 5.201,
485
- "train_steps_per_second": 0.082
486
  }
487
  ],
488
  "logging_steps": 10,
 
479
  "epoch": 3.0,
480
  "step": 288,
481
  "total_flos": 0.0,
482
+ "train_loss": 0.0,
483
+ "train_runtime": 0.0237,
484
+ "train_samples_per_second": 774670.33,
485
+ "train_steps_per_second": 12165.606
486
  }
487
  ],
488
  "logging_steps": 10,
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:492c64dc58da7ba228f8deac2fac8a1ed145c20f522d9c6826444661dccf8d7b
3
  size 7672
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d3768ab389c39deb7c4853b0e98898373bab62f7663a87fd250647e7bee6999f
3
  size 7672