llama_model_load: error loading model: error loading model vocabulary: unknown pre-tokenizer type: 'qwen2' llama_load_model_from_file: exception loading model terminate called after throwing an instance of 'std::runtime_error' what(): error loading model vocabulary: unknown pre-tokenizer type:...
tokenizer.ggml.model str = gpt2 llama_model_loader: - kv 12: tokenizer.ggml.pre str = jina-v2-de llama_model_loader: - kv 13: tokenizer.ggml.tokens arr[str,61056] = ["<s>", "<pad>", "</s>", "<unk>", "<m... llama_model_loader: - kv 14: tokenizer.ggml.token_type...
➜ ~ brew search ollama ==> Formulae ollama ==> Casks notesollama ollama ✔ ollamac ➜ ~ ollama run qwen2:7b Error: llama runner process has terminated: signal: abort trap error:error loading model vocabulary: unknown pre-tokenizer type: 'qwen2' Copy link smg1024 commented Jun ...
missing or unrecognized pre-tokenizer type, using: 'default' ollama | llm_load_vocab: special tokens cache size = 944 ollama | llm_load_vocab: token to piece cache size = 0.3151 MB ollama | llm_load_print_meta: format = GGUF V3 (latest) ollama | llm_load_print_meta: arch = phi...
tokenizer.o" PGEN = "Parser/pgen" PGENOBJS = "\ \" PGENSRCS = "\ \" PGOBJS = "\" PGO_PROF_GEN_FLAG = "-fprofile-generate" PGO_PROF_USE_FLAG = "-fprofile-use -fprofile-correction" PGSRCS = "\" PLATDIR = "plat-linux2" PLATMACDIRS = "plat-mac plat-mac/Carbon plat-mac/...
torchrun --nproc_per_node 1 example_text_infilling.py \ --ckpt_dir CodeLlama-7b/ \ --tokenizer_path CodeLlama-7b/tokenizer.model \ --max_seq_len 192 --max_batch_size 4 Pretrained infilling models are: the Code Llama models CodeLlama-7b and CodeLlama-13b and the Code Llama - Instr...