Finally, in a separate shell, run a model: ./ollama run llama2 REST API Ollama has a REST API for running and managing models. Generate a response curl http://localhost:11434/api/generate -d '{ "model": "llama2", "prompt":"Why is the sky blue?" }' Chat with a model curl...
Finally, in a separate shell, run a model: ./ollama run llama2 REST API Ollama has a REST API for running and managing models. Generate a response curl http://localhost:11434/api/generate -d '{ "model": "llama2", "prompt":"Why is the sky blue?" }' Chat with a model curl...
Chat with a model curl http://localhost:11434/api/chat -d '{ "model": "mistral", "messages": [ { "role": "user", "content": "why is the sky blue?" } ] }' See theAPI documentationfor all endpoints. Community Integrations
curl http://localhost:11434/api/chat -d '{ "model": "mistral", "messages": [ { "role": "user", "content": "why is the sky blue?" } ] }' See theAPI documentationfor all endpoints. Community Integrations HTML UI Chatbot UI ...
Or send a chat message (coming in 0.1.14): curl http://localhost:11434/api/chat -d '{ "model": "mistral", "messages": [ { "role": "user", "content": "why is the sky blue?" } ] }' See theAPI documentationfor all endpoints. ...
curl http://localhost:11434/api/chat -d '{ "model": "mistral", "messages": [ { "role": "user", "content": "why is the sky blue?" } ] }' See theAPI documentationfor all endpoints. Community Integrations Pacman Releases No releases published ...