Update quantization_benchmarks.csv
Browse files- quantization_benchmarks.csv +39 -146
quantization_benchmarks.csv
CHANGED
@@ -1,146 +1,39 @@
|
|
1 |
-
model,quant method,truthfulqa_mc2 acc ± stderr,arc:challenge acc ± stderr,hellaswag acc,winogrande acc,average,throughput (tok/s),peak process vram (GB),calibration/quantization time,throughput w/ torch.compile,peak process vram w/ torch.compile,througput w/ marlin,peak process vram w/ marlin
|
2 |
-
Llama 3.1 8B,baseline (bf16),0.5456 ± 0.0150,0.5623 ± 0.0145,0.5973 ± 0.0049,0.7372 ± 0.0124,0.6106,38.82,16.77302,,79.27,17.12954,,
|
3 |
-
Llama 3.1 8B,AWQ (4 bit),0.5281 ± 0.0150,0.5589 ± .0145,.5869 ± .0049,0.7285 ± 0.0125,0.6006,43.09,8.11808,~10 minutes,,,,
|
4 |
-
Llama 3.1 8B,GPTQModel (4 bit),,,,,,36.83,6.31872,~23 minutes,,,37.84,6318.72
|
5 |
-
Llama 3.1 8B,AutoGPTQ (4 bit),0.5361 ± 0.0149,0.5589 ± 0.0145,0.5758 ± 0.0049,0.7103 ± 0.0127,0.595275,43.71,6.49278,~30 minutes,,,,
|
6 |
-
Llama 3.1 8B,bnb (nf4),0.5446 ± 0.0149,0.5563 ± 0.0145,0.5769 ± 0.0049,0.7301 ± 0.0125,0.601975,24.35,6.44035,~1 minute,,,,
|
7 |
-
Llama 3.1 8B,optimum quanto (w4a16),0.5162 ± 0.0150,0.5427 ± 0.0146,0.5856 ± 0.0049,0.7419 ± 0.0123,0.5966,31.22,6.57667,~30 seconds,,,,
|
8 |
-
Llama 3.1 8B,torchao (int4wo),0.5166 ±0.0150,0.5418 ± 0.0146,0.5855 ± 0.0049,0.7395 ± 0.0123,0.59585,24.98,6.49907,~20 seconds,85.76,6.8493,,
|
9 |
-
Llama 3.1 8B,HQQ (4 bit),0.5485 ± 0.0150,0.5495 ± 0.0145,0.5859 ± 0.0049,0.7285 ± 0.0125,0.6031,34.44,6.71718,,,,,
|
10 |
-
Llama 3.1 8B,HIGGS (4 bit),0.5214 ± 0.0149,0.5486 ± 0.0145,0.5836 ± 0.0049,0.7182 ± 0.0126,0.59295,28.35,6.81994,~5 minutes,,,,
|
11 |
-
Llama 3.1 8B,bnb (llm.int8()),0.5446 ± 0.0150,0.5623 ± 0.0145,0.5949 ± 0.0049,0.7324 ± 0.0124,0.60855,20.75,9.70772,~20 seconds,,,,
|
12 |
-
Llama 3.1 8B,HQQ (8 bit),0.5441 ± 0.0150, 0.5666 ± 0.0145, 0.5979 ± 0.0049, 0.7380 ± 0.0124,0.61165,9.07,10.6011,~80 seconds,,,,
|
13 |
-
Llama 3.1 8B,optimum quanto (int8wo),0.5436 ± 0.0150, 0.5640 ± 0.0145,0.5992 ± 0.0049, 0.7372 ± 0.0124,0.611,15.59,9.81887,~20 seconds,16.01,10.07052,,
|
14 |
-
Llama 3.1 8B,torchao (int8wo),0.5449 ± 0.0150,0.5640 ± 0.0145,0.5975 ± 0.0049,0.7380 ± 0.0124,0.6111,5.98,13.07155,~30 seconds,43.79,13.66714,,
|
15 |
-
Llama 3.1 8B,fbgemm (fp8),0.5430 ± 0.0150,0.5580 ± 0.0145,0.5958 ± 0.0049,0.7411 ± 0.0123,0.609475,33.83,10.00551,~30 seconds,,,,
|
16 |
-
Llama 3.1 8B,compressed-tensors (fp8),0.5398 ± 0.0151,0.5589 ± 0.0145,0.5950 ± 0.0049,0.7356 ± 0.0124,0.607325,,,,,,,
|
17 |
-
Llama 3.1 8B,VPTQ (2 bit),0.4543 ± 0.0149, 0.4923 ± 0.0146, 0.5258 ± 0.0050,0.6930 ± 0.0130,0.54135,32.35,5.28902,~2 hours,31.48,5.28692,,
|
18 |
-
Llama 3.1 8B,AQLM + PV (2 bit),0.5036 ± 0.0148,0.5230 ± 0.0146,0.5628 ± 0.0050,0.6938 ± 0.0130,0.5708,22.28,4.84023,~1 day,27.27,4.85491,,
|
19 |
-
Llama 3.1 8B,GPTQModel (2 bit),,,,,,19.02,18.45284,~26 minutes,,,,
|
20 |
-
Llama 3.1 8B,AutoGPTQ (2 bit),0.5127 ± 0.0150,0.1988 ± 0.0117,0.2665 ± 0.0044,0.4799 ± 0.0140,0.364475,6.25,11.02473,~26 minutes,,,,
|
21 |
-
Llama 3.1 70B,baseline (bf16),0.6068 ± 0.0147,0.6732 ± 0.0137,0.6666 ± 0.0047,0.8248 ± 0.0107,0.69285,9.73,142.26869,,10.1,142.81395,,
|
22 |
-
Llama 3.1 70B,AWQ (4 bit),0.5706 ± 0.0150,0.6681 ± 0.0138,0.6598 ± 0.0047,0.8193 ± 0.0108,0.67945,15.74,43.75288,~1 hour,,,,
|
23 |
-
Llama 3.1 70B,GPTQModel (4 bit),,,,,,14.84,40.5757,,,,15.28,40.5757
|
24 |
-
Llama 3.1 70B,AutoGPTQ (4 bit),0.5937 ± 0.0147,0.6655 ± 0.0138,0.6568 ± 0.0047,0.8185 ± 0.0108,0.683625,0.46,42.40022,~2 hours,,,,
|
25 |
-
Llama 3.1 70B,bnb (nf4),0.5939 ± 0.0148,0.6724 ± 0.0137,0.6592 ± 0.0047,0.8098 ± 0.0110,0.683825,11.27,44.62949,~2 minutes,,,,
|
26 |
-
Llama 3.1 70B,optimum quanto (w4a16),0.4847 ± 0.0164,0.2082 ± 0.0119,0.2582 ± 0.0044,0.4878 ± 0.0140,0.359725,12.97,80.39013,~2 minutes,,,,
|
27 |
-
Llama 3.1 70B,torchao (int4wo),0.4847 ± 0.0164,0.2108 ± 0.0119, 0.2581 ± 0.0044, 0.4980 ± 0.0141,0.3629,10.56,41.6054,~2 minutes,18.95,42.26181,,
|
28 |
-
Llama 3.1 70B,HQQ (4 bit),0.5882 ± 0.0146,0.6706 ± 0.0137, 0.6597 ± 0.0047,0.8035 ± 0.0112,0.6805,13.92,44.50366,~10 minutes,,,,
|
29 |
-
Llama 3.1 70B,HIGGS (4 bit),0.4871 ± 0.0163,0.1971 ± 0.0116,0.2575 ± 0.0044,0.4893 ± 0.0140,0.35775,11.61,41.52571,~6 minutes,12.38,41.02868,,
|
30 |
-
Llama 3.1 70B,bnb (llm.int8()),0.5604 ± 0.0169,0.6544 ± 0.0139,0.6382 ± 0.0048,0.7940 ± 0.0114,0.66175,6.87,74.26428,~2 minutes,,,,
|
31 |
-
Llama 3.1 70B,HQQ (8 bit),0.6112 ± 0.0146,0.6732 ± 0.0137,0.6661 ± 0.0047,0.8327 ± 0.0105,0.6958,0.98,80.52435,~10 minutes,0.98,80.39013,,
|
32 |
-
Llama 3.1 70B,optimum quanto (int8wo),0.5591 ± 0.0150,0.6459 ± 0.0140,0.6413 ± 0.0048,0.7979 ± 0.0113,0.66105,1.79,74.21192,~2 minutes,1.8,74.21401,,
|
33 |
-
Llama 3.1 70B,torchao (int8wo),0.6094 ± 0.0146,0.6732 ± 0.0137,0.6659 ± 0.0047,0.8240 ± 0.0107,0.693125,0.65,89.85038,~2 minutes,0.65,89.84619,,
|
34 |
-
Llama 3.1 70B,fbgemm (fp8),0.6075 ± 0.0146,0.6732 ± 0.0137,0.6671 ± 0.0047,0.8216 ± 0.0108,0.69235,13.61,74.04624,~6 minutes,,,,
|
35 |
-
Llama 3.1 70B,compressed-tensors (fp8),0.6062 ± 0.0146,0.6741 ± 0.0137,0.6652 ± 0.0047,0.8216 ± 0.0108,0.691775,,,,,,,
|
36 |
-
Llama 3.1 70B,VPTQ (2 bit),0.5451 ± 0.0150, 0.6212 ± 0.0142,0.6073 ± 0.0049, 0.7901 ± 0.0114,0.640925,6.29,24.89949,~19 hours,6.18,24.89949,,
|
37 |
-
Llama 3.1 70B,AQLM + PV (2 bit),0.5706 ± 0.0150,0.6365 ± 0.0141,0.6401 ± 0.0048,0.8066 ± 0.0111,0.66345,6.75,23.12739,10-14 days,7.09,23607.64,,
|
38 |
-
Llama 3.1 70B,GPTQModel (2 bit),,,,,,,,,,,,
|
39 |
-
Llama 3.1 70B,AutoGPTQ (2 bit),0.4556 ± 0.0147,0.2807 ± 0.0131,0.3642 ± 0.0048,0.5470 ± 0.0140,0.411875,,,,,,,
|
40 |
-
,,,,,,,,,,,,,
|
41 |
-
,,,,,,,,,,,,,
|
42 |
-
,,,,,,,,,,,,,
|
43 |
-
,,,,,,,,,,,,,
|
44 |
-
,,,,,,,,,,,,,
|
45 |
-
,,,,,,,,,,,,,
|
46 |
-
,,,,,,,,,,,,,
|
47 |
-
,,,,,,,,,,,,,
|
48 |
-
,,,,,,,,,,,,,
|
49 |
-
,,,,,,,,,,,,,
|
50 |
-
,,,,,,,,,,,,,
|
51 |
-
,,,,,,,,,,,,,
|
52 |
-
,,,,,,,,,,,,,
|
53 |
-
,,,,,,,,,,,,,
|
54 |
-
,,,,,,,,,,,,,
|
55 |
-
,,,,,,,,,,,,,
|
56 |
-
,,,,,,,,,,,,,
|
57 |
-
,,,,,,,,,,,,,
|
58 |
-
,,,,,,,,,,,,,
|
59 |
-
,,,,,,,,,,,,,
|
60 |
-
,,,,,,,,,,,,,
|
61 |
-
,,,,,,,,,,,,,
|
62 |
-
,,,,,,,,,,,,,
|
63 |
-
,,,,,,,,,,,,,
|
64 |
-
,,,,,,,,,,,,,
|
65 |
-
,,,,,,,,,,,,,
|
66 |
-
,,,,,,,,,,,,,
|
67 |
-
,,,,,,,,,,,,,
|
68 |
-
,,,,,,,,,,,,,
|
69 |
-
,,,,,,,,,,,,,
|
70 |
-
,,,,,,,,,,,,,
|
71 |
-
,,,,,,,,,,,,,
|
72 |
-
,,,,,,,,,,,,,
|
73 |
-
,,,,,,,,,,,,,
|
74 |
-
,,,,,,,,,,,,,
|
75 |
-
,,,,,,,,,,,,,
|
76 |
-
,,,,,,,,,,,,,
|
77 |
-
,,,,,,,,,,,,,
|
78 |
-
,,,,,,,,,,,,,
|
79 |
-
,,,,,,,,,,,,,
|
80 |
-
,,,,,,,,,,,,,
|
81 |
-
,,,,,,,,,,,,,
|
82 |
-
,,,,,,,,,,,,,
|
83 |
-
,,,,,,,,,,,,,
|
84 |
-
,,,,,,,,,,,,,
|
85 |
-
,,,,,,,,,,,,,
|
86 |
-
,,,,,,,,,,,,,
|
87 |
-
,,,,,,,,,,,,,
|
88 |
-
,,,,,,,,,,,,,
|
89 |
-
,,,,,,,,,,,,,
|
90 |
-
,,,,,,,,,,,,,
|
91 |
-
,,,,,,,,,,,,,
|
92 |
-
,,,,,,,,,,,,,
|
93 |
-
,,,,,,,,,,,,,
|
94 |
-
,,,,,,,,,,,,,
|
95 |
-
,,,,,,,,,,,,,
|
96 |
-
,,,,,,,,,,,,,
|
97 |
-
,,,,,,,,,,,,,
|
98 |
-
,,,,,,,,,,,,,
|
99 |
-
,,,,,,,,,,,,,
|
100 |
-
,task1 acc ± stderr,,,,,,,,,,,,
|
101 |
-
,,,,,,,,,,,,,
|
102 |
-
,,,,,,,,,,,,,
|
103 |
-
,,,,,,,,,,,,,
|
104 |
-
,,,,,,,,,,,,,
|
105 |
-
,,,,,,,,,,,,,
|
106 |
-
,,,,,,,,,,,,,
|
107 |
-
,,,,,,,,,,,,,
|
108 |
-
,,,,,,,,,,,,,
|
109 |
-
,,,,,,,,,,,,,
|
110 |
-
,,,,,,,,,,,,,
|
111 |
-
,,,,,,,,,,,,,
|
112 |
-
,,,,,,,,,,,,,
|
113 |
-
,,,,,,,,,,,,,
|
114 |
-
,,,,,,,,,,,,,
|
115 |
-
,,,,,,,,,,,,,
|
116 |
-
,,,,,,,,,,,,,
|
117 |
-
,,,,,,,,,,,,,
|
118 |
-
,,,,,,,,,,,,,
|
119 |
-
,,,,,,,,,,,,,
|
120 |
-
,,,,,,,,,,,,,
|
121 |
-
,,,,,,,,,,,,,
|
122 |
-
,,,,,,,,,,,,,
|
123 |
-
,task1 acc ± stderr,,,,,,,,,,,,
|
124 |
-
,,,,,,,,,,,,,
|
125 |
-
,,,,,,,,,,,,,
|
126 |
-
,,,,,,,,,,,,,
|
127 |
-
,,,,,,,,,,,,,
|
128 |
-
,,,,,,,,,,,,,
|
129 |
-
,,,,,,,,,,,,,
|
130 |
-
,,,,,,,,,,,,,
|
131 |
-
,,,,,,,,,,,,,
|
132 |
-
,,,,,,,,,,,,,
|
133 |
-
,,,,,,,,,,,,,
|
134 |
-
,,,,,,,,,,,,,
|
135 |
-
,,,,,,,,,,,,,
|
136 |
-
,,,,,,,,,,,,,
|
137 |
-
,,,,,,,,,,,,,
|
138 |
-
,,,,,,,,,,,,,
|
139 |
-
,,,,,,,,,,,,,
|
140 |
-
,,,,,,,,,,,,,
|
141 |
-
,,,,,,,,,,,,,
|
142 |
-
,,,,,,,,,,,,,
|
143 |
-
,,,,,,,,,,,,,
|
144 |
-
,,,,,,,,,,,,,
|
145 |
-
,,,,,,,,,,,,,
|
146 |
-
,task1 acc ± stderr,,,,,,,,,,,,
|
|
|
1 |
+
model,quant method,truthfulqa_mc2 acc ± stderr,arc:challenge acc ± stderr,hellaswag acc,winogrande acc,average,throughput (tok/s),peak process vram (GB),calibration/quantization time,throughput w/ torch.compile,peak process vram w/ torch.compile,througput w/ marlin,peak process vram w/ marlin
|
2 |
+
Llama 3.1 8B,baseline (bf16),0.5456 ± 0.0150,0.5623 ± 0.0145,0.5973 ± 0.0049,0.7372 ± 0.0124,0.6106,38.82,16.77302,,79.27,17.12954,,
|
3 |
+
Llama 3.1 8B,AWQ (4 bit),0.5281 ± 0.0150,0.5589 ± .0145,.5869 ± .0049,0.7285 ± 0.0125,0.6006,43.09,8.11808,~10 minutes,,,,
|
4 |
+
Llama 3.1 8B,GPTQModel (4 bit),,,,,,36.83,6.31872,~23 minutes,,,37.84,6318.72
|
5 |
+
Llama 3.1 8B,AutoGPTQ (4 bit),0.5361 ± 0.0149,0.5589 ± 0.0145,0.5758 ± 0.0049,0.7103 ± 0.0127,0.595275,43.71,6.49278,~30 minutes,,,,
|
6 |
+
Llama 3.1 8B,bnb (nf4),0.5446 ± 0.0149,0.5563 ± 0.0145,0.5769 ± 0.0049,0.7301 ± 0.0125,0.601975,24.35,6.44035,~1 minute,,,,
|
7 |
+
Llama 3.1 8B,optimum quanto (w4a16),0.5162 ± 0.0150,0.5427 ± 0.0146,0.5856 ± 0.0049,0.7419 ± 0.0123,0.5966,31.22,6.57667,~30 seconds,,,,
|
8 |
+
Llama 3.1 8B,torchao (int4wo),0.5166 ±0.0150,0.5418 ± 0.0146,0.5855 ± 0.0049,0.7395 ± 0.0123,0.59585,24.98,6.49907,~20 seconds,85.76,6.8493,,
|
9 |
+
Llama 3.1 8B,HQQ (4 bit),0.5485 ± 0.0150,0.5495 ± 0.0145,0.5859 ± 0.0049,0.7285 ± 0.0125,0.6031,34.44,6.71718,,,,,
|
10 |
+
Llama 3.1 8B,HIGGS (4 bit),0.5214 ± 0.0149,0.5486 ± 0.0145,0.5836 ± 0.0049,0.7182 ± 0.0126,0.59295,28.35,6.81994,~5 minutes,,,,
|
11 |
+
Llama 3.1 8B,bnb (llm.int8()),0.5446 ± 0.0150,0.5623 ± 0.0145,0.5949 ± 0.0049,0.7324 ± 0.0124,0.60855,20.75,9.70772,~20 seconds,,,,
|
12 |
+
Llama 3.1 8B,HQQ (8 bit),0.5441 ± 0.0150, 0.5666 ± 0.0145, 0.5979 ± 0.0049, 0.7380 ± 0.0124,0.61165,9.07,10.6011,~80 seconds,,,,
|
13 |
+
Llama 3.1 8B,optimum quanto (int8wo),0.5436 ± 0.0150, 0.5640 ± 0.0145,0.5992 ± 0.0049, 0.7372 ± 0.0124,0.611,15.59,9.81887,~20 seconds,16.01,10.07052,,
|
14 |
+
Llama 3.1 8B,torchao (int8wo),0.5449 ± 0.0150,0.5640 ± 0.0145,0.5975 ± 0.0049,0.7380 ± 0.0124,0.6111,5.98,13.07155,~30 seconds,43.79,13.66714,,
|
15 |
+
Llama 3.1 8B,fbgemm (fp8),0.5430 ± 0.0150,0.5580 ± 0.0145,0.5958 ± 0.0049,0.7411 ± 0.0123,0.609475,33.83,10.00551,~30 seconds,,,,
|
16 |
+
Llama 3.1 8B,compressed-tensors (fp8),0.5398 ± 0.0151,0.5589 ± 0.0145,0.5950 ± 0.0049,0.7356 ± 0.0124,0.607325,,,,,,,
|
17 |
+
Llama 3.1 8B,VPTQ (2 bit),0.4543 ± 0.0149, 0.4923 ± 0.0146, 0.5258 ± 0.0050,0.6930 ± 0.0130,0.54135,32.35,5.28902,~2 hours,31.48,5.28692,,
|
18 |
+
Llama 3.1 8B,AQLM + PV (2 bit),0.5036 ± 0.0148,0.5230 ± 0.0146,0.5628 ± 0.0050,0.6938 ± 0.0130,0.5708,22.28,4.84023,~1 day,27.27,4.85491,,
|
19 |
+
Llama 3.1 8B,GPTQModel (2 bit),,,,,,19.02,18.45284,~26 minutes,,,,
|
20 |
+
Llama 3.1 8B,AutoGPTQ (2 bit),0.5127 ± 0.0150,0.1988 ± 0.0117,0.2665 ± 0.0044,0.4799 ± 0.0140,0.364475,6.25,11.02473,~26 minutes,,,,
|
21 |
+
Llama 3.1 70B,baseline (bf16),0.6068 ± 0.0147,0.6732 ± 0.0137,0.6666 ± 0.0047,0.8248 ± 0.0107,0.69285,9.73,142.26869,,10.1,142.81395,,
|
22 |
+
Llama 3.1 70B,AWQ (4 bit),0.5706 ± 0.0150,0.6681 ± 0.0138,0.6598 ± 0.0047,0.8193 ± 0.0108,0.67945,15.74,43.75288,~1 hour,,,,
|
23 |
+
Llama 3.1 70B,GPTQModel (4 bit),,,,,,14.84,40.5757,,,,15.28,40.5757
|
24 |
+
Llama 3.1 70B,AutoGPTQ (4 bit),0.5937 ± 0.0147,0.6655 ± 0.0138,0.6568 ± 0.0047,0.8185 ± 0.0108,0.683625,0.46,42.40022,~2 hours,,,,
|
25 |
+
Llama 3.1 70B,bnb (nf4),0.5939 ± 0.0148,0.6724 ± 0.0137,0.6592 ± 0.0047,0.8098 ± 0.0110,0.683825,11.27,44.62949,~2 minutes,,,,
|
26 |
+
Llama 3.1 70B,optimum quanto (w4a16),0.4847 ± 0.0164,0.2082 ± 0.0119,0.2582 ± 0.0044,0.4878 ± 0.0140,0.359725,12.97,80.39013,~2 minutes,,,,
|
27 |
+
Llama 3.1 70B,torchao (int4wo),0.4847 ± 0.0164,0.2108 ± 0.0119, 0.2581 ± 0.0044, 0.4980 ± 0.0141,0.3629,10.56,41.6054,~2 minutes,18.95,42.26181,,
|
28 |
+
Llama 3.1 70B,HQQ (4 bit),0.5882 ± 0.0146,0.6706 ± 0.0137, 0.6597 ± 0.0047,0.8035 ± 0.0112,0.6805,13.92,44.50366,~10 minutes,,,,
|
29 |
+
Llama 3.1 70B,HIGGS (4 bit),0.4871 ± 0.0163,0.1971 ± 0.0116,0.2575 ± 0.0044,0.4893 ± 0.0140,0.35775,11.61,41.52571,~6 minutes,12.38,41.02868,,
|
30 |
+
Llama 3.1 70B,bnb (llm.int8()),0.5604 ± 0.0169,0.6544 ± 0.0139,0.6382 ± 0.0048,0.7940 ± 0.0114,0.66175,6.87,74.26428,~2 minutes,,,,
|
31 |
+
Llama 3.1 70B,HQQ (8 bit),0.6112 ± 0.0146,0.6732 ± 0.0137,0.6661 ± 0.0047,0.8327 ± 0.0105,0.6958,0.98,80.52435,~10 minutes,0.98,80.39013,,
|
32 |
+
Llama 3.1 70B,optimum quanto (int8wo),0.5591 ± 0.0150,0.6459 ± 0.0140,0.6413 ± 0.0048,0.7979 ± 0.0113,0.66105,1.79,74.21192,~2 minutes,1.8,74.21401,,
|
33 |
+
Llama 3.1 70B,torchao (int8wo),0.6094 ± 0.0146,0.6732 ± 0.0137,0.6659 ± 0.0047,0.8240 ± 0.0107,0.693125,0.65,89.85038,~2 minutes,0.65,89.84619,,
|
34 |
+
Llama 3.1 70B,fbgemm (fp8),0.6075 ± 0.0146,0.6732 ± 0.0137,0.6671 ± 0.0047,0.8216 ± 0.0108,0.69235,13.61,74.04624,~6 minutes,,,,
|
35 |
+
Llama 3.1 70B,compressed-tensors (fp8),0.6062 ± 0.0146,0.6741 ± 0.0137,0.6652 ± 0.0047,0.8216 ± 0.0108,0.691775,,,,,,,
|
36 |
+
Llama 3.1 70B,VPTQ (2 bit),0.5451 ± 0.0150, 0.6212 ± 0.0142,0.6073 ± 0.0049, 0.7901 ± 0.0114,0.640925,6.29,24.89949,~19 hours,6.18,24.89949,,
|
37 |
+
Llama 3.1 70B,AQLM + PV (2 bit),0.5706 ± 0.0150,0.6365 ± 0.0141,0.6401 ± 0.0048,0.8066 ± 0.0111,0.66345,6.75,23.12739,10-14 days,7.09,23607.64,,
|
38 |
+
Llama 3.1 70B,GPTQModel (2 bit),,,,,,,,,,,,
|
39 |
+
Llama 3.1 70B,AutoGPTQ (2 bit),0.4556 ± 0.0147,0.2807 ± 0.0131,0.3642 ± 0.0048,0.5470 ± 0.0140,0.411875,,,,,,,
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|