zerozeroz commited on
Commit
292d0d1
·
verified ·
1 Parent(s): d3a3b56

Model save

Browse files
README.md ADDED
@@ -0,0 +1,67 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: transformers
3
+ model_name: CodeLlama-7b-hf_mbxpall
4
+ tags:
5
+ - generated_from_trainer
6
+ - trl
7
+ - grpo
8
+ licence: license
9
+ ---
10
+
11
+ # Model Card for CodeLlama-7b-hf_mbxpall
12
+
13
+ This model is a fine-tuned version of [None](https://huggingface.co/None).
14
+ It has been trained using [TRL](https://github.com/huggingface/trl).
15
+
16
+ ## Quick start
17
+
18
+ ```python
19
+ from transformers import pipeline
20
+
21
+ question = "If you had a time machine, but could only go to the past or the future once and never return, which would you choose and why?"
22
+ generator = pipeline("text-generation", model="zerozeroz/CodeLlama-7b-hf_mbxpall", device="cuda")
23
+ output = generator([{"role": "user", "content": question}], max_new_tokens=128, return_full_text=False)[0]
24
+ print(output["generated_text"])
25
+ ```
26
+
27
+ ## Training procedure
28
+
29
+
30
+
31
+
32
+ This model was trained with GRPO, a method introduced in [DeepSeekMath: Pushing the Limits of Mathematical Reasoning in Open Language Models](https://huggingface.co/papers/2402.03300).
33
+
34
+ ### Framework versions
35
+
36
+ - TRL: 0.14.0
37
+ - Transformers: 4.48.1
38
+ - Pytorch: 2.5.1+cu121
39
+ - Datasets: 3.1.0
40
+ - Tokenizers: 0.21.0
41
+
42
+ ## Citations
43
+
44
+ Cite GRPO as:
45
+
46
+ ```bibtex
47
+ @article{zhihong2024deepseekmath,
48
+ title = {{DeepSeekMath: Pushing the Limits of Mathematical Reasoning in Open Language Models}},
49
+ author = {Zhihong Shao and Peiyi Wang and Qihao Zhu and Runxin Xu and Junxiao Song and Mingchuan Zhang and Y. K. Li and Y. Wu and Daya Guo},
50
+ year = 2024,
51
+ eprint = {arXiv:2402.03300},
52
+ }
53
+
54
+ ```
55
+
56
+ Cite TRL as:
57
+
58
+ ```bibtex
59
+ @misc{vonwerra2022trl,
60
+ title = {{TRL: Transformer Reinforcement Learning}},
61
+ author = {Leandro von Werra and Younes Belkada and Lewis Tunstall and Edward Beeching and Tristan Thrush and Nathan Lambert and Shengyi Huang and Kashif Rasul and Quentin Gallouédec},
62
+ year = 2020,
63
+ journal = {GitHub repository},
64
+ publisher = {GitHub},
65
+ howpublished = {\url{https://github.com/huggingface/trl}}
66
+ }
67
+ ```
all_results.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "total_flos": 0.0,
3
+ "train_loss": 0.011399921192697879,
4
+ "train_runtime": 6781.3987,
5
+ "train_samples": 1687,
6
+ "train_samples_per_second": 1.493,
7
+ "train_steps_per_second": 0.249
8
+ }
config.json CHANGED
@@ -25,6 +25,6 @@
25
  "tie_word_embeddings": false,
26
  "torch_dtype": "bfloat16",
27
  "transformers_version": "4.48.1",
28
- "use_cache": false,
29
  "vocab_size": 32016
30
  }
 
25
  "tie_word_embeddings": false,
26
  "torch_dtype": "bfloat16",
27
  "transformers_version": "4.48.1",
28
+ "use_cache": true,
29
  "vocab_size": 32016
30
  }
generation_config.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 1,
4
+ "eos_token_id": 2,
5
+ "transformers_version": "4.48.1"
6
+ }
model-00001-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6a20f752ee9233c5577edffa2bf29d9050bbbe2f95ae424e3a58420a0b6e2c35
3
  size 4939116424
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3a3c098b745de81c3b94ec6abe50ad6b5ee8639f403217356140e9953aa12d5d
3
  size 4939116424
model-00002-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:39c693e45d8863112549d87c711a5bc56eec687b982a84b5d76d3b1d7c93ff9b
3
  size 4947390880
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fa5c1b823c6ba6d810f3dbf15033405262c89f674fd23fdc6951f51b803063e1
3
  size 4947390880
model-00003-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cc58ecf9bad5ba002d6c8ef8038e7db70144f46aa44bec2c87b4109ae2fcaa8e
3
  size 3590619888
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2066d27170c45bf65f7e620982cac6b5b9c9c704919796ff9b63181537996d8b
3
  size 3590619888
runs/May18_08-47-19_fdse/events.out.tfevents.1747558109.fdse.4130810.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b382d7bb75236089a46f50fc577d7fb32117f6d98766749436149f45e30a400f
3
- size 649949
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5404e0a3dc6badb5791d202f081dbabcdcab96af310ba7e545ccbc9519caa70e
3
+ size 860160
runs/May20_01-48-23_fdse/events.out.tfevents.1747705849.fdse.3296835.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4e22afc0bb9b1f584839abf08662a0b9b7bb4c50ff30e5283018fb62d56100c6
3
+ size 163939
train_results.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "total_flos": 0.0,
3
+ "train_loss": 0.011399921192697879,
4
+ "train_runtime": 6781.3987,
5
+ "train_samples": 1687,
6
+ "train_samples_per_second": 1.493,
7
+ "train_steps_per_second": 0.249
8
+ }
trainer_state.json ADDED
The diff for this file is too large to render. See raw diff
 
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:abca653864f94e3029474175d954aabb8195fb92bf52b22300d7d574dd35a493
3
  size 6840
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ef75073f82c16c90d8efd947584a8dbb2fde7179ab757751158228225c13823a
3
  size 6840