qt-spyro-hf commited on
Commit
d5d0407
·
verified ·
1 Parent(s): aba7f7e

Upload 13 files

Browse files
Modelfile CHANGED
@@ -1,12 +1,13 @@
1
  FROM codellama:7b-code
2
 
3
- # Base model name and adapter
4
- ADAPTER ./codellama-7b-qml.gguf
5
-
6
- # Parameters optimized for code generation
7
- PARAMETER temperature 0.2
8
  PARAMETER num_predict 500
9
- PARAMETER top_p 0.9
 
10
  PARAMETER stop "<SUF>"
11
  PARAMETER stop "<PRE>"
12
  PARAMETER stop "</PRE>"
@@ -17,9 +18,9 @@ PARAMETER stop "<MID>"
17
  PARAMETER stop "</MID>"
18
  PARAMETER stop "##"
19
 
20
-
21
- # Template format for code interactions
22
- TEMPLATE """{{if .Prompt}}{{ .Prompt }}{{end}}{{if .Response}}{{ .Response }}{{end}}"""
23
 
24
  LICENSE """
25
  Qt AI Model Use Terms and Conditions
 
1
  FROM codellama:7b-code
2
 
3
+ # Base model name and adapter
4
+ ADAPTER ./codellama-7b-qml.gguf
5
+
6
+ # Parameters optimized for code generation
7
+ PARAMETER temperature 0
8
  PARAMETER num_predict 500
9
+ PARAMETER top_p 1
10
+ PARAMETER repeat_penalty 1.05
11
  PARAMETER stop "<SUF>"
12
  PARAMETER stop "<PRE>"
13
  PARAMETER stop "</PRE>"
 
18
  PARAMETER stop "</MID>"
19
  PARAMETER stop "##"
20
 
21
+
22
+ # Template format for code interactions
23
+ TEMPLATE """{{if .Prompt}}{{ .Prompt }}{{end}}{{if .Response}}{{ .Response }}{{end}}"""
24
 
25
  LICENSE """
26
  Qt AI Model Use Terms and Conditions
adapter_config.json CHANGED
@@ -14,7 +14,7 @@
14
  "loftq_config": {},
15
  "lora_alpha": 64,
16
  "lora_bias": false,
17
- "lora_dropout": 0.1,
18
  "megatron_config": null,
19
  "megatron_core": "megatron.core",
20
  "modules_to_save": null,
@@ -23,14 +23,14 @@
23
  "rank_pattern": {},
24
  "revision": null,
25
  "target_modules": [
26
- "v_proj",
27
- "o_proj",
28
  "gate_proj",
29
- "lm_head",
30
- "down_proj",
31
  "q_proj",
 
32
  "up_proj",
33
- "k_proj"
 
 
34
  ],
35
  "task_type": "CAUSAL_LM",
36
  "use_dora": false,
 
14
  "loftq_config": {},
15
  "lora_alpha": 64,
16
  "lora_bias": false,
17
+ "lora_dropout": 0.02,
18
  "megatron_config": null,
19
  "megatron_core": "megatron.core",
20
  "modules_to_save": null,
 
23
  "rank_pattern": {},
24
  "revision": null,
25
  "target_modules": [
 
 
26
  "gate_proj",
27
+ "k_proj",
 
28
  "q_proj",
29
+ "o_proj",
30
  "up_proj",
31
+ "v_proj",
32
+ "down_proj",
33
+ "lm_head"
34
  ],
35
  "task_type": "CAUSAL_LM",
36
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ec4924dfac7c38ed7e78d23d166c1bac46cc7cde6ef2ec3d70c741a4d5b11ea7
3
  size 648936760
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:126dbda0a3f5d86df24daa2176ec0302363d5c88d3213674c58f7778640a4842
3
  size 648936760
codellama-7b-qml.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:87a033f1b70147091d4a52d8d7a3938ca182b5efa701418adaddcaf7331dce4e
3
- size 648907072
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9298cedefd9ae4a2b11b0eb13d3beacb7e65b34192895ba844a630efc1a57e32
3
+ size 648907040
optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8da06080bb75f0fe64cb3bb6897ff13b9b2f17a551222d1a4883a09ca52cf17e
3
  size 325614754
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:40d81f632c3097acbf16d2f345f186c95a2ca66d4a7d08db30fd475354b95520
3
  size 325614754
rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:054177611f2412e52a8cd0a4d0895e2cb9e4fef94374378f0bf6c4059d506903
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7bd1594d76d68b1d4a1950f6c0e9f5e68990863544f80e019ccf7a2e05bdcb69
3
  size 14244
scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ef1fe48b040bf7678711805e0ccda18391a9ad1db687676a750066fbac4eecfb
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a1e14f542310c644a4bc29da20a8f89a9ba2f396d841056f00df29c8bfcd1650
3
  size 1064
trainer_state.json CHANGED
The diff for this file is too large to render. See raw diff
 
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e81583ff738f437f9b8ba61f8cd63306401c4b51ce22ce038811cf0a2a0f493e
3
  size 5816
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:975fd47280f92ed577c218a35385d271f6263cfd5c676dc0079baed0c6ff516a
3
  size 5816