sadkins65 commited on
Commit
63f3a45
·
verified ·
1 Parent(s): 517610b

Upload folder using huggingface_hub

Browse files
Files changed (3) hide show
  1. config.json +20 -3
  2. model.safetensors +2 -2
  3. recipe.yaml +12 -1
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "/nm/drive0/sadkins/.cache/huggingface/hub/models--TinyLlama--TinyLlama-1.1B-intermediate-step-1431k-3T/snapshots/036fa4651240b9a1487f709833b9e4b96b4c1574",
3
  "architectures": [
4
  "LlamaForCausalLM"
5
  ],
@@ -23,7 +23,12 @@
23
  "quantization_config": {
24
  "config_groups": {
25
  "group_0": {
26
- "input_activations": {
 
 
 
 
 
27
  "block_structure": null,
28
  "group_size": null,
29
  "num_bits": 8,
@@ -32,6 +37,18 @@
32
  "strategy": "tensor",
33
  "symmetric": true,
34
  "type": "int"
 
 
 
 
 
 
 
 
 
 
 
 
35
  },
36
  "output_activations": null,
37
  "targets": [
@@ -50,7 +67,7 @@
50
  }
51
  },
52
  "format": "fakequant",
53
- "global_compression_ratio": 1.4375595368095078,
54
  "ignore": [
55
  "model.layers.0.mlp.down_proj",
56
  "lm_head"
 
1
  {
2
+ "_name_or_path": "/root/.cache/huggingface/hub/models--TinyLlama--TinyLlama-1.1B-intermediate-step-1431k-3T/snapshots/036fa4651240b9a1487f709833b9e4b96b4c1574",
3
  "architectures": [
4
  "LlamaForCausalLM"
5
  ],
 
23
  "quantization_config": {
24
  "config_groups": {
25
  "group_0": {
26
+ "input_activations": null,
27
+ "output_activations": null,
28
+ "targets": [
29
+ "Embedding"
30
+ ],
31
+ "weights": {
32
  "block_structure": null,
33
  "group_size": null,
34
  "num_bits": 8,
 
37
  "strategy": "tensor",
38
  "symmetric": true,
39
  "type": "int"
40
+ }
41
+ },
42
+ "group_1": {
43
+ "input_activations": {
44
+ "block_structure": null,
45
+ "group_size": null,
46
+ "num_bits": 8,
47
+ "observer": "minmax",
48
+ "observer_kwargs": {},
49
+ "strategy": "tensor",
50
+ "symmetric": false,
51
+ "type": "int"
52
  },
53
  "output_activations": null,
54
  "targets": [
 
67
  }
68
  },
69
  "format": "fakequant",
70
+ "global_compression_ratio": 1.4416825559554713,
71
  "ignore": [
72
  "model.layers.0.mlp.down_proj",
73
  "lm_head"
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:836e22041658a79b3024c37b093b0fa02e7f8a3480c306c3491e3353d2e5bf00
3
- size 4400287762
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3f94cc1d9b92c63daae79a819a45fbe669bbf1afaba13435ab0f0de7aa47130f
3
+ size 4400287663
recipe.yaml CHANGED
@@ -5,6 +5,17 @@ test_stage:
5
  config_groups:
6
  group_0:
7
  weights: {num_bits: 8, type: int, symmetric: true, strategy: tensor}
8
- input_activations: {num_bits: 8, type: int, symmetric: true, strategy: tensor}
9
  output_activations: null
10
  targets: [Linear]
 
 
 
 
 
 
 
 
 
 
 
 
5
  config_groups:
6
  group_0:
7
  weights: {num_bits: 8, type: int, symmetric: true, strategy: tensor}
8
+ input_activations: {num_bits: 8, type: int, symmetric: false, strategy: tensor}
9
  output_activations: null
10
  targets: [Linear]
11
+ group_1:
12
+ weights: {num_bits: 8, type: int, symmetric: true, strategy: tensor}
13
+ input_activations: null
14
+ output_activations: null
15
+ targets: [Embedding]
16
+ SparseGPTModifier:
17
+ sparsity: 0.0
18
+ block_size: 128
19
+ sequential_update: false
20
+ quantize: true
21
+ targets: ['re:model.layers.\d+$']