Ollama

From Freephile Wiki
Revision as of 10:53, 19 June 2025 by Greg Rundlett (talk | contribs) (link to Open-Webui)

Ollama is a tool that allows users to run large language models (LLMs) directly on their own computers, making powerful AI technology accessible without relying on cloud services. It provides a user-friendly way to manage, deploy, and integrate LLMs, offering greater control, privacy, and customization compared to traditional cloud-based solutions.

Ollama was funded by Jared Friedman out of Y Combinator (YC). Founders Jeffrey Morgan and Michael Chiang wanted an easier way to run LLMs than having to do it in the cloud. In fact, they were previously founders of a startup project named Kitematic which was the early UI for Docker. Acquired by Docker, it was the precursor to Docker Desktop.

Ollama will enable you to get up and running with Llama 3.3, DeepSeek-R1, Phi-4, Gemma 3, Mistral Small 3.1 and other large language models.

Installing it

I had some problems getting off the ground with Ollama. Some details are in Ollama/install

Docs

The docs tell you how you can customize and update or uninstall the environment.

Looking at the logs with journalctl -e -u ollama told me what my new generated public key is, but also that it could not load a compatible GPU so I spent time fixing that.

Start with the README for an intro.

Interface

Although you can instantly begin using a model from the command line with something like

ollama run gemma3 [1] there are many User Interfaces or front-ends that can be coupled with Ollama such as Open-Webui.== References ==

  1. This will download and run a 4B parameter model.