3.0 KiB
title, homepage, tagline
| title | homepage | tagline |
|---|---|---|
| ollama | https://github.com/jmorganca/ollama | ollama is a golang LLM server built for ease of use. |
To update or switch versions, run webi ollama@stable (or @v0.1.5, etc).
Cheat Sheet
ollamais an LLM serving platform written in golang. It makes LLMs built on Llama standards easy to run with an API.
To get started quickly with the open source LLM Mistral-7b as an example is two commands.
- Open TWO Terminals
- In the first, start the
ollamaserverOLLAMA_ORIGINS='*' OLLAMA_HOST=localhost:11434 ollama serve - In the second, run the
ollamaCLI (using the Mistral-7b model)ollama pull mistral ollama run mistral
Table of Contents
- Files
- ChatGPT-style Web UI
- System Notes
- Models to Try
- As a Network API
Files
These are the files / directories that are created and/or modified with this install:
~/.config/envman/PATH.env
~/.local/bin/ollama
~/.ollama/models/
How to Use a ChatGPT-style Web Interface
There are many Ollama UIs to choose from, but ollama-webui is
easy to start with (and can be built as a static page):
- Install
nodewebi node@lts source ~/.config/envman/PATH.env - Clone and enter the
ollama-webuirepogit clone https://github.com/ollama-webui/ollama-webui.git ./ollama-webui/ pushd ./ollama-webui/ - Install and start the project
cp -RPp ./example.env ./.env npm clean-install npm run dev - Enjoy!
Note: Be sure to run ollama with CORS enabled:
OLLAMA_ORIGINS='*' OLLAMA_HOST=localhost:11434 ollama serve
System Notes
You'll need a fairly modern computer. An Apple M1 Air works great.
- 8GB+ RAM
- 4GB+ Storage
- Models range between 3GB and 30GB+
(they can take a while to download, and several seconds to initialize)
How to Downloads Other Models
See the list at https://ollama.ai/library.
For example, we could try sqlcoder, or orca-mini (because it's small):
ollama pull sqlcoder
ollama run sqlcoder
ollama pull orca-mini
ollama run orca-mini
How to Use as an API on a Network
If you'd like ollama to be accessible beyond localhost (127.0.0.1):
- set the host to
0.0.0.0, which makes it accessible to ALL networks - you may wish to limit origins
# fully open to all
OLLAMA_ORIGINS='*' OLLAMA_HOST=0.0.0.0:11435 ollama serve
# restrict browsers (not APIs) to requests from https://example.com
OLLAMA_ORIGINS='https://example.com' OLLAMA_HOST=0.0.0.0:11435 ollama serve
See also:

