https://github.com/ggerganov/llama.cpp/pull/6920 (#26)
Browse files- https://github.com/ggerganov/llama.cpp/pull/6920 (2d9ade45393570ab4a8ca7a7c257842e726df2d6)
Co-authored-by: Lewdiculous <Lewdiculous@users.noreply.huggingface.co>
- gguf-imat-llama-3.py +2 -2
gguf-imat-llama-3.py
CHANGED
|
@@ -101,7 +101,7 @@ def download_model_repo():
|
|
| 101 |
convert_model_to_gguf_f16(base_dir, model_dir, model_name, delete_model_dir, imatrix_file_name)
|
| 102 |
|
| 103 |
def convert_model_to_gguf_f16(base_dir, model_dir, model_name, delete_model_dir, imatrix_file_name):
|
| 104 |
-
convert_script = os.path.join(base_dir, "llama.cpp", "convert.py")
|
| 105 |
gguf_dir = os.path.join(base_dir, "models", f"{model_name}-GGUF")
|
| 106 |
gguf_model_path = os.path.join(gguf_dir, f"{model_name}-F16.gguf")
|
| 107 |
|
|
@@ -109,7 +109,7 @@ def convert_model_to_gguf_f16(base_dir, model_dir, model_name, delete_model_dir,
|
|
| 109 |
os.makedirs(gguf_dir)
|
| 110 |
|
| 111 |
if not os.path.exists(gguf_model_path):
|
| 112 |
-
subprocess.run(["python", convert_script, model_dir, "--outfile", gguf_model_path, "--outtype", "f16"
|
| 113 |
|
| 114 |
if delete_model_dir == 'yes' or delete_model_dir == 'y':
|
| 115 |
shutil.rmtree(model_dir)
|
|
|
|
| 101 |
convert_model_to_gguf_f16(base_dir, model_dir, model_name, delete_model_dir, imatrix_file_name)
|
| 102 |
|
| 103 |
def convert_model_to_gguf_f16(base_dir, model_dir, model_name, delete_model_dir, imatrix_file_name):
|
| 104 |
+
convert_script = os.path.join(base_dir, "llama.cpp", "convert-hf-to-gguf.py")
|
| 105 |
gguf_dir = os.path.join(base_dir, "models", f"{model_name}-GGUF")
|
| 106 |
gguf_model_path = os.path.join(gguf_dir, f"{model_name}-F16.gguf")
|
| 107 |
|
|
|
|
| 109 |
os.makedirs(gguf_dir)
|
| 110 |
|
| 111 |
if not os.path.exists(gguf_model_path):
|
| 112 |
+
subprocess.run(["python", convert_script, model_dir, "--outfile", gguf_model_path, "--outtype", "f16"])
|
| 113 |
|
| 114 |
if delete_model_dir == 'yes' or delete_model_dir == 'y':
|
| 115 |
shutil.rmtree(model_dir)
|