Add more detail on minimum GPU
Browse files
README.md
CHANGED
|
@@ -135,11 +135,11 @@ See sample configs in [configs](configs) folder or [examples](examples) for quic
|
|
| 135 |
|
| 136 |
- loading
|
| 137 |
```yaml
|
| 138 |
-
load_4bit: true
|
| 139 |
load_in_8bit: true
|
| 140 |
-
|
|
|
|
| 141 |
fp16: true
|
| 142 |
-
tf32: true
|
| 143 |
```
|
| 144 |
Note: Repo does not do 4-bit quantization.
|
| 145 |
|
|
@@ -183,11 +183,11 @@ gptq_model_v1: false # v1 or v2
|
|
| 183 |
load_in_8bit: true
|
| 184 |
|
| 185 |
# Use CUDA bf16
|
| 186 |
-
bf16: true # bool or 'full' for `bf16_full_eval
|
| 187 |
# Use CUDA fp16
|
| 188 |
fp16: true
|
| 189 |
# Use CUDA tf32
|
| 190 |
-
tf32: true
|
| 191 |
|
| 192 |
# a list of one or more datasets to finetune the model with
|
| 193 |
datasets:
|
|
@@ -286,7 +286,7 @@ weight_decay:
|
|
| 286 |
# whether to use xformers attention patch https://github.com/facebookresearch/xformers:
|
| 287 |
xformers_attention:
|
| 288 |
# whether to use flash attention patch https://github.com/HazyResearch/flash-attention:
|
| 289 |
-
flash_attention:
|
| 290 |
|
| 291 |
# resume from a specific checkpoint dir
|
| 292 |
resume_from_checkpoint:
|
|
|
|
| 135 |
|
| 136 |
- loading
|
| 137 |
```yaml
|
|
|
|
| 138 |
load_in_8bit: true
|
| 139 |
+
load_in_8bit: true
|
| 140 |
+
bf16: true # require >=ampere
|
| 141 |
fp16: true
|
| 142 |
+
tf32: true # require >=ampere
|
| 143 |
```
|
| 144 |
Note: Repo does not do 4-bit quantization.
|
| 145 |
|
|
|
|
| 183 |
load_in_8bit: true
|
| 184 |
|
| 185 |
# Use CUDA bf16
|
| 186 |
+
bf16: true # bool or 'full' for `bf16_full_eval`. require >=ampere
|
| 187 |
# Use CUDA fp16
|
| 188 |
fp16: true
|
| 189 |
# Use CUDA tf32
|
| 190 |
+
tf32: true # require >=ampere
|
| 191 |
|
| 192 |
# a list of one or more datasets to finetune the model with
|
| 193 |
datasets:
|
|
|
|
| 286 |
# whether to use xformers attention patch https://github.com/facebookresearch/xformers:
|
| 287 |
xformers_attention:
|
| 288 |
# whether to use flash attention patch https://github.com/HazyResearch/flash-attention:
|
| 289 |
+
flash_attention: # require a100 for llama
|
| 290 |
|
| 291 |
# resume from a specific checkpoint dir
|
| 292 |
resume_from_checkpoint:
|