Upload 13 files
Browse files- Modelfile +10 -9
- adapter_config.json +6 -6
- adapter_model.safetensors +1 -1
- codellama-7b-qml.gguf +2 -2
- optimizer.pt +1 -1
- rng_state.pth +1 -1
- scheduler.pt +1 -1
- trainer_state.json +0 -0
- training_args.bin +1 -1
Modelfile
CHANGED
@@ -1,12 +1,13 @@
|
|
1 |
FROM codellama:7b-code
|
2 |
|
3 |
-
# Base model name and adapter
|
4 |
-
ADAPTER ./codellama-7b-qml.gguf
|
5 |
-
|
6 |
-
# Parameters optimized for code generation
|
7 |
-
PARAMETER temperature 0
|
8 |
PARAMETER num_predict 500
|
9 |
-
PARAMETER top_p
|
|
|
10 |
PARAMETER stop "<SUF>"
|
11 |
PARAMETER stop "<PRE>"
|
12 |
PARAMETER stop "</PRE>"
|
@@ -17,9 +18,9 @@ PARAMETER stop "<MID>"
|
|
17 |
PARAMETER stop "</MID>"
|
18 |
PARAMETER stop "##"
|
19 |
|
20 |
-
|
21 |
-
# Template format for code interactions
|
22 |
-
TEMPLATE """{{if .Prompt}}{{ .Prompt }}{{end}}{{if .Response}}{{ .Response }}{{end}}"""
|
23 |
|
24 |
LICENSE """
|
25 |
Qt AI Model Use Terms and Conditions
|
|
|
1 |
FROM codellama:7b-code
|
2 |
|
3 |
+
# Base model name and adapter
|
4 |
+
ADAPTER ./codellama-7b-qml.gguf
|
5 |
+
|
6 |
+
# Parameters optimized for code generation
|
7 |
+
PARAMETER temperature 0
|
8 |
PARAMETER num_predict 500
|
9 |
+
PARAMETER top_p 1
|
10 |
+
PARAMETER repeat_penalty 1.05
|
11 |
PARAMETER stop "<SUF>"
|
12 |
PARAMETER stop "<PRE>"
|
13 |
PARAMETER stop "</PRE>"
|
|
|
18 |
PARAMETER stop "</MID>"
|
19 |
PARAMETER stop "##"
|
20 |
|
21 |
+
|
22 |
+
# Template format for code interactions
|
23 |
+
TEMPLATE """{{if .Prompt}}{{ .Prompt }}{{end}}{{if .Response}}{{ .Response }}{{end}}"""
|
24 |
|
25 |
LICENSE """
|
26 |
Qt AI Model Use Terms and Conditions
|
adapter_config.json
CHANGED
@@ -14,7 +14,7 @@
|
|
14 |
"loftq_config": {},
|
15 |
"lora_alpha": 64,
|
16 |
"lora_bias": false,
|
17 |
-
"lora_dropout": 0.
|
18 |
"megatron_config": null,
|
19 |
"megatron_core": "megatron.core",
|
20 |
"modules_to_save": null,
|
@@ -23,14 +23,14 @@
|
|
23 |
"rank_pattern": {},
|
24 |
"revision": null,
|
25 |
"target_modules": [
|
26 |
-
"v_proj",
|
27 |
-
"o_proj",
|
28 |
"gate_proj",
|
29 |
-
"
|
30 |
-
"down_proj",
|
31 |
"q_proj",
|
|
|
32 |
"up_proj",
|
33 |
-
"
|
|
|
|
|
34 |
],
|
35 |
"task_type": "CAUSAL_LM",
|
36 |
"use_dora": false,
|
|
|
14 |
"loftq_config": {},
|
15 |
"lora_alpha": 64,
|
16 |
"lora_bias": false,
|
17 |
+
"lora_dropout": 0.02,
|
18 |
"megatron_config": null,
|
19 |
"megatron_core": "megatron.core",
|
20 |
"modules_to_save": null,
|
|
|
23 |
"rank_pattern": {},
|
24 |
"revision": null,
|
25 |
"target_modules": [
|
|
|
|
|
26 |
"gate_proj",
|
27 |
+
"k_proj",
|
|
|
28 |
"q_proj",
|
29 |
+
"o_proj",
|
30 |
"up_proj",
|
31 |
+
"v_proj",
|
32 |
+
"down_proj",
|
33 |
+
"lm_head"
|
34 |
],
|
35 |
"task_type": "CAUSAL_LM",
|
36 |
"use_dora": false,
|
adapter_model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 648936760
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:126dbda0a3f5d86df24daa2176ec0302363d5c88d3213674c58f7778640a4842
|
3 |
size 648936760
|
codellama-7b-qml.gguf
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9298cedefd9ae4a2b11b0eb13d3beacb7e65b34192895ba844a630efc1a57e32
|
3 |
+
size 648907040
|
optimizer.pt
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 325614754
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:40d81f632c3097acbf16d2f345f186c95a2ca66d4a7d08db30fd475354b95520
|
3 |
size 325614754
|
rng_state.pth
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 14244
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7bd1594d76d68b1d4a1950f6c0e9f5e68990863544f80e019ccf7a2e05bdcb69
|
3 |
size 14244
|
scheduler.pt
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 1064
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a1e14f542310c644a4bc29da20a8f89a9ba2f396d841056f00df29c8bfcd1650
|
3 |
size 1064
|
trainer_state.json
CHANGED
The diff for this file is too large to render.
See raw diff
|
|
training_args.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 5816
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:975fd47280f92ed577c218a35385d271f6263cfd5c676dc0079baed0c6ff516a
|
3 |
size 5816
|