'disable_exllamav2':shared.args.disable_exllamav2, } logger.info(f"The AutoGPTQ params are:{params}") Expand Down 2 changes: 2 additions & 0 deletions2modules/loaders.py Original file line numberDiff line numberDiff line change Expand Up@@ -25,6 +25,7 @@ ...
Original file line numberDiff line numberDiff line change @@ -119,6 +119,7 @@ def create_ui(): shared.gradio['use_double_quant'] = gr.Checkbox(label="use_double_quant", value=shared.args.use_double_quant) shared.gradio['use_flash_attention_2'] = gr.Checkbox(label="use_flash_atten...
I am unable to use the --rwkv-cuda-on option, as I get this error: File "E:\oobabooga\text-generation-webui\server.py", line 241, in <module> shared.model, shared.tokenizer = load_model(shared.model_name) File "E:\oobabooga\text-generation-webui\modules\models.py", line 90, ...
Original file line numberDiff line numberDiff line change Expand Up@@ -126,6 +126,10 @@ def huggingface_loader(model_name): 'torch_dtype':torch.bfloat16ifshared.args.bf16elsetorch.float16, 'use_safetensors':Trueifshared.args.force_safetensorselseNone ...
: Traceback (most recent call last): File "D:\AI\oobabooga_windows\installer_files\env\lib\threading.py", line 1016, in _bootstrap_inner self.run() File "D:\AI\oobabooga_windows\installer_files\env\lib\threading.py", line 953, in run self._target(*self._args, **self._kwargs) ...
Sorry super dumb but do I pass this to start-webui.sh? Like sh start-webui.sh --gpu-memory 10 5? Ah, that should work, but if not, edit the file and add this at the end ofcall python server.py --auto-devices --cai-chat ...
pip uninstall llama-cpp-python CMAKE_ARGS="-DLLAMA_CLBLAST=ON" FORCE_CMAKE=1 pip install --no-cache-dir llama-cpp-python Step 4 In terminal (not .\cmd_linux.sh), cd into the text-generation-webui directory if you're not still there. Go to conda' clinfo -l output and note the...
return super().to(*args, **kwargs) File "C:\AI\oobabooga_windows\installer_files\env\lib\site-packages\torch\nn\modules\module.py", line 1145, in to return self._apply(convert) File "C:\AI\oobabooga_windows\installer_files\env\lib\site-packages\torch\nn\modules\module.py", line 797...
CalledProcessError(retcode, process.args, subprocess.CalledProcessError: Command'['ninja','-v']'returned non-zeroexitstatus 1. The above exception was the direct cause of the following exception: Traceback (most recent call last): File"E:\AI\oobabooga\text-gen-webui-2\text-generation-webui...
line 929, in <module> create_interface() File "/home/roman/oobabooga_linux/text-generation-webui/server.py", line 854, in create_interface shared.gradio['interface'].launch(prevent_thread_lock=True, share=shared.args.share, server_name=shared.args.listen_host or '0.0.0.0', server_port=...