Oleg Shulyakov commited on
Commit
dd433e4
·
1 Parent(s): 280ff7d

Change imatrix format to gguf

Browse files
Files changed (1) hide show
  1. app.py +2 -3
app.py CHANGED
@@ -152,7 +152,6 @@ class HuggingFaceModelProcessor:
152
  "-f", train_data_path,
153
  "-ngl", "99",
154
  "--output-frequency", "10",
155
- "--output-format", "dat",
156
  "-o", quant_config.imatrix_file,
157
  ]
158
 
@@ -417,7 +416,7 @@ llama-server --hf-repo "{processing_config.new_repo_id}" --hf-file "{processing_
417
  if quant_config.use_imatrix and os.path.isfile(quant_config.imatrix_file):
418
  try:
419
  print(f"Uploading imatrix.dat: {os.path.abspath(quant_config.imatrix_file)}")
420
- self._upload_file(processing_config, quant_config.imatrix_file, f"{processing_config.model_name}-imatrix.dat")
421
  except Exception as e:
422
  raise GGUFConverterError(f"Error uploading imatrix.dat: {e}")
423
 
@@ -783,7 +782,7 @@ class GGUFConverterUI:
783
  )
784
 
785
  quant_config.fp16_model = f"{outdir}/{model_name}-fp16.gguf"
786
- quant_config.imatrix_file = f"{outdir}/{model_name}-imatrix.dat"
787
  quant_config.quantized_gguf = f"{outdir}/{gguf_name}"
788
 
789
  processing_config = ModelProcessingConfig(
 
152
  "-f", train_data_path,
153
  "-ngl", "99",
154
  "--output-frequency", "10",
 
155
  "-o", quant_config.imatrix_file,
156
  ]
157
 
 
416
  if quant_config.use_imatrix and os.path.isfile(quant_config.imatrix_file):
417
  try:
418
  print(f"Uploading imatrix.dat: {os.path.abspath(quant_config.imatrix_file)}")
419
+ self._upload_file(processing_config, quant_config.imatrix_file, f"{processing_config.model_name}-imatrix.gguf")
420
  except Exception as e:
421
  raise GGUFConverterError(f"Error uploading imatrix.dat: {e}")
422
 
 
782
  )
783
 
784
  quant_config.fp16_model = f"{outdir}/{model_name}-fp16.gguf"
785
+ quant_config.imatrix_file = f"{outdir}/{model_name}-imatrix.gguf"
786
  quant_config.quantized_gguf = f"{outdir}/{gguf_name}"
787
 
788
  processing_config = ModelProcessingConfig(