To start, Ollama doesn’tofficiallyrun on Windows. With enough hacking you could get a Python environment going and figure it out. But we don’t have to because we can use one of my favorite features, WSL orWindows Subsystem for Linux. If you need to install WSL, here’s how you do...
You can now run ollama and it will run on the remote in your Colab (so long as that's stays up and running). e.g. run this on your local machine and it will look as if it's running locally but it's really running in your Colab and the results are being served to wherever yo...
docker run -d -p 3000:8080 -e OLLAMA_BASE_URL=https://example.com -v open-webui:/app/backend/data --name open-webui --restart always ghcr.io/open-webui/open-webui:main To run Open WebUI with Nvidia GPU support, use this command: ...
from_dockerhub( "nvidia/cuda:11.7.0-devel-ubuntu20.04", setup_dockerfile_commands=[ "RUN apt-get update", "RUN apt-get install -y python3 python3-pip python-is-python3", ], ) # it also reaches the same output with debian slim image if we were to swap out the base image # Image...