Ollama: Difference between revisions

Klinger (talk | contribs)
Add troubleshooting instructions for AMD GPUs
Line 18: Line 18:
ollama run mistral
ollama run mistral
</syntaxhighlight>For ruther models see [https://ollama.ai/library Ollama library].
</syntaxhighlight>For ruther models see [https://ollama.ai/library Ollama library].
== Troubleshooting ==
=== AMD GPU with open source driver ===
In certain cases ollama might not allow your system to use GPU acceleration if it cannot be sure your GPU/driver is compatible.
However you can attempt to force-enable the usage of your GPU by overriding the LLVM target. <ref>https://github.com/ollama/ollama/blob/main/docs/gpu.md#overrides</ref>
You can get the version for your GPU from the logs or like so:
<syntaxhighlight lang="bash">
$ nix-shell -p "rocmPackages.rocminfo" --run "rocminfo" | grep "gfx"
Name:                    gfx1031
</syntaxhighlight>
In this example the LLVM target is "gfx1031", that is, version "10.3.1", you can then override that value for ollama:
<syntaxhighlight lang="nix">
services.ollama = {
  enable = true;
  acceleration = "rocm";
};
systemd.services.ollama.environment = {
  HSA_OVERRIDE_GFX_VERSION = "10.3.1";
};
</syntaxhighlight>
If there are still errors, you can attempt to set a similar value that is listed [https://github.com/ollama/ollama/blob/main/docs/gpu.md#overrides here].


[[Category:Server]]
[[Category:Server]]