sadkins65 commited on
Commit
f8e86b7
·
verified ·
1 Parent(s): 63f3a45

Upload folder using huggingface_hub

Browse files
Files changed (3) hide show
  1. config.json +2 -19
  2. model.safetensors +2 -2
  3. recipe.yaml +1 -6
config.json CHANGED
@@ -23,23 +23,6 @@
23
  "quantization_config": {
24
  "config_groups": {
25
  "group_0": {
26
- "input_activations": null,
27
- "output_activations": null,
28
- "targets": [
29
- "Embedding"
30
- ],
31
- "weights": {
32
- "block_structure": null,
33
- "group_size": null,
34
- "num_bits": 8,
35
- "observer": "minmax",
36
- "observer_kwargs": {},
37
- "strategy": "tensor",
38
- "symmetric": true,
39
- "type": "int"
40
- }
41
- },
42
- "group_1": {
43
  "input_activations": {
44
  "block_structure": null,
45
  "group_size": null,
@@ -47,7 +30,7 @@
47
  "observer": "minmax",
48
  "observer_kwargs": {},
49
  "strategy": "tensor",
50
- "symmetric": false,
51
  "type": "int"
52
  },
53
  "output_activations": null,
@@ -67,7 +50,7 @@
67
  }
68
  },
69
  "format": "fakequant",
70
- "global_compression_ratio": 1.4416825559554713,
71
  "ignore": [
72
  "model.layers.0.mlp.down_proj",
73
  "lm_head"
 
23
  "quantization_config": {
24
  "config_groups": {
25
  "group_0": {
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
26
  "input_activations": {
27
  "block_structure": null,
28
  "group_size": null,
 
30
  "observer": "minmax",
31
  "observer_kwargs": {},
32
  "strategy": "tensor",
33
+ "symmetric": true,
34
  "type": "int"
35
  },
36
  "output_activations": null,
 
50
  }
51
  },
52
  "format": "fakequant",
53
+ "global_compression_ratio": 1.4375595368095078,
54
  "ignore": [
55
  "model.layers.0.mlp.down_proj",
56
  "lm_head"
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3f94cc1d9b92c63daae79a819a45fbe669bbf1afaba13435ab0f0de7aa47130f
3
- size 4400287663
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5aae4cc5021049424ebba130bff524467cf1575b37914eb22011e67fd9a5e531
3
+ size 4400287458
recipe.yaml CHANGED
@@ -5,14 +5,9 @@ test_stage:
5
  config_groups:
6
  group_0:
7
  weights: {num_bits: 8, type: int, symmetric: true, strategy: tensor}
8
- input_activations: {num_bits: 8, type: int, symmetric: false, strategy: tensor}
9
  output_activations: null
10
  targets: [Linear]
11
- group_1:
12
- weights: {num_bits: 8, type: int, symmetric: true, strategy: tensor}
13
- input_activations: null
14
- output_activations: null
15
- targets: [Embedding]
16
  SparseGPTModifier:
17
  sparsity: 0.0
18
  block_size: 128
 
5
  config_groups:
6
  group_0:
7
  weights: {num_bits: 8, type: int, symmetric: true, strategy: tensor}
8
+ input_activations: {num_bits: 8, type: int, symmetric: true, strategy: tensor}
9
  output_activations: null
10
  targets: [Linear]
 
 
 
 
 
11
  SparseGPTModifier:
12
  sparsity: 0.0
13
  block_size: 128