To make your build sharable and capable of working on other devices, you must useLLAMA_PORTABLE=1 After all binaries are built, you can run the python script with the commandkoboldcpp.py [ggml_model.gguf] [port] Compiling on Windows ...
After all binaries are built, you can run the python script with the command koboldcpp.py --model [ggml_model.gguf] (and add --gpulayers (number of layer) if you wish to offload layers to GPU).Compiling on Android (Termux Installation)Install...
After all binaries are built, you can run the python script with the commandkoboldcpp.py --model [ggml_model.gguf](and add--gpulayers (number of layer)if you wish to offload layers to GPU). Compiling on Android (Termux Installation) ...