📝 Edit page
➕ Add page
CLI
Download models
$ ollama pull llama3.1
Start interactive CLI
See Model library in the GitHub docs for models and CLI commands to download.
e.g.
$ ollama run llama3.1
Run a once-off request
Pass as an argument e.g.
$ ollama run llama3.1 'Tell me a joke'
Read a text file:
$ ollama run --nowordwrap llama3.1 "Summarize this file in one sentence: $(cat file.txt)"
Read an image file such as with bakllava or llava image models:
$ ollama run --nowordwrap bakllava "What's in this image? ./photo.jpg"
Access via HTTP
When a model is requested, it will be loaded and used.
Then you can find it under running models here:
$ ollama ps
Stop
See conversation on GitHub here. There is no builtin way to stop ollama. Use your system manager or use pkill.
On macOS, find the icon as running apps and click Quit.
If you started with ollam serve, then just stop that process to stop Ollama.