startswith("hf_"): logger.info("Huggingface token seems invalid") logger.info("Usage: python convert_hf_to_gguf_update.py <huggingface_token>") sys.exit(1) else: logger.info("Usage: python convert_hf_to_gguf_update.py <huggingface_token>") sys.exit(1) # TODO: add ...
# TODO: auto-update convert-hf-to-gguf.py with the generated function src_ifs = "" for model in models: @@ -224,11 +208,18 @@ def get_vocab_base_pre(self, tokenizer) -> str: return res """ print(src_func) # noqa: NP100 convert_py_pth = pathlib.Path("convert-hf-to-gguf...
Update convert-hf-to-gguf.py No worries! Looking at it now, I'm really not sure if the CI ran prior to the merge or not, so I don't know if there was a way you reasonably could have caught it. We may need to modify the CI triggers on Python-only changes -- I'm really no...
if 'NO_LOCAL_GGUF' not in os.environ: sys.path.insert(1, str(Path(__file__).parent / 'gguf-py')) import gguf # reuse model definitions from convert_hf_to_gguf.py from convert_hf_to_gguf import LazyTorchTensor, Model ...
Search or jump to... Search code, repositories, users, issues, pull requests... Provide feedback We read every piece of feedback, and take your input very seriously. Include my email address so I can be contacted Cancel Submit feedback Saved searches Use saved searches to filter your...
sys.path.insert(1,str(Path(__file__).parent/'gguf-py'/'gguf')) importgguf NUMPY_TYPE_TO_FTYPE:dict[str,int]={"float32":0,"float16":1} defwrite_file_header(fout:BinaryIO,params:dict[str,Any])->None: fout.write(b"ggla"[::-1])# magic (ggml lora) ...
Name and Version latest Operating systems Linux GGML backends CUDA Hardware A6000 Models https://huggingface.co/meta-llama/Llama-3.2-11B-Vision-Instruct Problem description & steps to reproduce when using convert_hf_to_gguf.py convert ll...
if 'NO_LOCAL_GGUF' not in os.environ: sys.path.insert(1, str(Path(__file__).parent / 'gguf-py')) import gguf logger = logging.getLogger("ggml-to-gguf") class GGMLFormat(IntEnum): GGML = 0 GGMF = 1 GGJT = 2 class GGMLFType(IntEnum): ALL_F32 = 0 MO...
2 changes: 1 addition & 1 deletion 2 convert_hf_to_gguf.py Original file line numberDiff line numberDiff line change @@ -2719,7 +2719,7 @@ def set_vocab(self): for line in lines: parts = line.split(' ') assert len(parts) >= 3 _, token, token_len = int(parts[0]), ast...
29 changes: 27 additions & 2 deletions 29 convert-hf-to-gguf.py Original file line numberDiff line numberDiff line change @@ -1416,8 +1416,32 @@ def set_vocab(self): self.gguf_writer.add_add_space_prefix(add_prefix)special_vocab = gguf.SpecialVocab(self.dir_model, n_vocab=len(...