Ollama on Raspberry Pi
As part of getting into SLM for small AI devices, we are going to look at inference speed on the Raspberry Pi. Mostly six models are of interest:
- tinyllama
- tinydolphin
- phi3
- smallthinker
- granite3.1-moe
- llama3.2:1b
Each model has a small number of parameters, to make sure we get usable speeds for inference.
We will get speed figures on simple inference for each of them.
In a previous post, we installed Arch Linux on the raspberry pi (make sure to have a look at Running Arch on Pi ).