Ollama – Local AI Model Runner
Run, manage, and switch between a wide range of open‑source LLMs directly on your local machine. Ollama provides fast, offline inference with a simple CLI and API, ensuring your data never leaves the device. Ideal for developers, researchers, and anyone who wants powerful AI without cloud dependencies.
You are about to open
Do you wish to proceed?
Thank you for your report. Information you provided will help us investigate further.
There was an error while sending your report. Please try again later.
Snaps are applications packaged with all their dependencies to run on all popular Linux distributions from a single build. They update automatically and roll back gracefully.
Snaps are discoverable and installable from the Snap Store, an app store with an audience of millions.
On a Raspberry Pi running the latest version of Raspbian snap can be installed directly from the command line:
sudo apt update
sudo apt install snapd
You will also need to reboot your device:
sudo reboot
After this, install the snapd snap in order to get the latest snapd:
sudo snap install snapd
To install ollama, simply use the following command:
sudo snap install ollama
Browse and find snaps from the convenience of your desktop using the snap store snap.
Interested to find out more about snaps? Want to publish your own application? Visit snapcraft.io now.