Toggle menu
Toggle preferences menu
Toggle personal menu
Not logged in
Your IP address will be publicly visible if you make any edits.
This is the latest revision of this page; it has no approved revision.

Ollama is a runtime for running AI models locally or on a personal server.

Installation

Official Install Script

curl -fsSL https://ollama.com/install.sh | sh

Arch Linux

No GPU

sudo pacman -S ollama 

ROCm (AMD)

sudo pacman -S ollama-rocm

CUDA (Nvidia)

sudo pacman -S ollama-cuda

Using Models

When picking models, ensure there is enough memory to fit the model you are using. Models small enough to fit in the GPUs memory will only use the GPU while models too big to fit in the GPU will use a mix of GPU and CPU. To check GPU and CPU memory usage run:

ollama ps

From Terminal

Gemma 3 will download if not already on the system and then run.

ollama run gemma3

Gemma 3 by default uses the 4B version of the model but different versions can be specified.

ollama run gemma3:12b

Ollama uses q4_K_M quantisation by default but different quantisation level can be specified.

ollama run gemma3:12b-it-q8_0

From a Graphical User Interface

Alpaca

Alpaca is officially distributed as a Flatpak.

flatpak install flathub com.jeffser.Alpaca

Alpaca can be configured to use Ollama externally including remote hosts but can also install Ollama internally as a Flatpak.

flatpak install flathub com.jeffser.Alpaca.Plugins.Ollama

The Flatpak Ollama plugin can support AMD GPUs using another plugin.

flatpal install flathub com.jeffser.Alpaca.Plugins.AMD

Visual Studio Code

Ollama can be integrated into Visual Studio Code using Continue.

Officially Supported GPUs

Consult this table for official GPU support.

Use Unsupported AMD GPUs on ROCm

To use ROCm on GPUs not officially supported by Ollama we can extend the Systemd service file with the following command:

sudo systemctl edit ollama.service

This will open up a file in the systems text editor where you can extend the service. Below are examples of what to add to the file:

Example: Radeon RX 6000 series

[Service]
Environment="HSA_OVERRIDE_GFX_VERSION=10.3.0"

Example: Radeon RX 5000 series

[Service]

Environment="HSA_OVERRIDE_GFX_VERSION=10.1.0"

The Ollama service will need to be restarted before the GPU can be used.

sudo systemctl restart ollama.service