Update README.md
Browse files
README.md
CHANGED
@@ -184,6 +184,9 @@ huggingface-cli login
|
|
184 |
and use a token with write access, from https://huggingface.co/settings/tokens
|
185 |
|
186 |
# Model Quality
|
|
|
|
|
|
|
187 |
| Benchmark | | |
|
188 |
|----------------------------------|----------------|---------------------------|
|
189 |
| | Phi-4-mini-ins | Phi-4-mini-ins-int4wo-hqq |
|
@@ -209,8 +212,6 @@ and use a token with write access, from https://huggingface.co/settings/tokens
|
|
209 |
<details>
|
210 |
<summary> Reproduce Model Quality Results </summary>
|
211 |
|
212 |
-
We rely on [lm-evaluation-harness](https://github.com/EleutherAI/lm-evaluation-harness) to evaluate the quality of the quantized model.
|
213 |
-
|
214 |
Need to install lm-eval from source:
|
215 |
https://github.com/EleutherAI/lm-evaluation-harness#install
|
216 |
|
|
|
184 |
and use a token with write access, from https://huggingface.co/settings/tokens
|
185 |
|
186 |
# Model Quality
|
187 |
+
|
188 |
+
We rely on [lm-evaluation-harness](https://github.com/EleutherAI/lm-evaluation-harness) to evaluate the quality of the quantized model.
|
189 |
+
|
190 |
| Benchmark | | |
|
191 |
|----------------------------------|----------------|---------------------------|
|
192 |
| | Phi-4-mini-ins | Phi-4-mini-ins-int4wo-hqq |
|
|
|
212 |
<details>
|
213 |
<summary> Reproduce Model Quality Results </summary>
|
214 |
|
|
|
|
|
215 |
Need to install lm-eval from source:
|
216 |
https://github.com/EleutherAI/lm-evaluation-harness#install
|
217 |
|