llama_model_loader: - type f16: 170 tensors time=2024-05-20T16:44:58.427+08:00 level=INFO source=server.go:540 msg="waiting for server to become available" status="llm server loading model" llama_model_load: error loading model: error loading model vocabulary: unknown pre-tokenizer type:...
.file_type u32 = 1 llama_model_loader: - kv 9: jina-bert-v2.attention.causal bool = false llama_model_loader: - kv 10: jina-bert-v2.pooling_type u32 = 1 llama_model_loader: - kv 11: tokenizer.ggml.model str = gpt2 llama_model_loader: - kv 12: tokenizer.ggml.pre str =...
➜ ~ brew search ollama ==> Formulae ollama ==> Casks notesollama ollama ✔ ollamac ➜ ~ ollama run qwen2:7b Error: llama runner process has terminated: signal: abort trap error:error loading model vocabulary: unknown pre-tokenizer type: 'qwen2' Copy link smg1024 commented Jun ...
- type f32: 147 tensors ollama | llama_model_loader: - type q4_0: 97 tensors ollama | llama_model_loader: - type q6_K: 1 tensors ollama | llm_load_vocab: missing or unrecognized pre-tokenizer type, using: 'default' ollama | llm_load_vocab: special tokens cache size = 944 o...
HAVE_RL_PRE_INPUT_HOOK = "1" HAVE_RL_RESIZE_TERMINAL = "1" HAVE_ROUND = "1" HAVE_SELECT = "1" HAVE_SEM_GETVALUE = "1" HAVE_SEM_OPEN = "1" HAVE_SEM_TIMEDWAIT = "1" HAVE_SEM_UNLINK = "1" HAVE_SETEGID = "1" HAVE_SETEUID = "1" HAVE_SETGID = "1" HAVE_SETGROUPS = ...
Steps to reproduce - !pip install --quiet transformers==4.5.0 !pip install --quiet pytorch-lightning==1.2.7 from transformers import ( AdamW, T5ForConditionalGeneration, T5TokenizerFast as T5Tokenizer ) Throws error - ---...