Bez popisu

Matt Williams e734b62d4d had the instruction and input reversed před 1 rokem
api 8b1e791820 allow specifying zero values in modelfile před 1 rokem
app 6fbb380076 hide dock icon if window closes před 1 rokem
cmd bd6d741d87 tell users to check the server error logs před 1 rokem
docs da2459d519 Update README.md (#265) před 1 rokem
examples 42bdcee473 update to nous-hermes modelfile před 1 rokem
format 5bea29f610 add new list command (#97) před 1 rokem
library e734b62d4d had the instruction and input reversed před 1 rokem
llama c5bcf32823 update llama.cpp před 1 rokem
parser 9c7f30d31c use max scan token size to hold large objects před 1 rokem
progressbar e4d7f3e287 vendor in progress bar and change to bytes instead of bibytes (#130) před 1 rokem
scripts da52f5bfdd run `npm install` on build před 1 rokem
server e3fb1fd3f1 server: compare options correctly před 1 rokem
web 6fff59c637 Update discord invite link před 1 rokem
.dockerignore 6292f4b64c update `Dockerfile` před 1 rokem
.gitignore 67b6f8ba86 add `ggml-metal.metal` to `.gitignore` před 1 rokem
.prettierrc.json 8685a5ad18 move .prettierrc.json to root před 1 rokem
Dockerfile 7c71c10d4f fix compilation issue in Dockerfile, remove from `README.md` until ready před 1 rokem
LICENSE df5fdd6647 `proto` -> `ollama` před 1 rokem
README.md df8048fecd Update README.md před 1 rokem
go.mod 1c5a8770ee read runner parameter options from map před 1 rokem
go.sum 1c5a8770ee read runner parameter options from map před 1 rokem
main.go 1775647f76 continue conversation před 1 rokem

README.md

logo

Ollama

Discord

Note: Ollama is in early preview. Please report any issues you find.

Run, create, and share large language models (LLMs).

Download

  • Download for macOS on Apple Silicon (Intel coming soon)
  • Download for Windows and Linux (coming soon)
  • Build from source

Quickstart

To run and chat with Llama 2, the new model by Meta:

ollama run llama2

Model library

ollama includes a library of open-source models:

Model Parameters Size Download
Llama2 7B 3.8GB ollama pull llama2
Llama2 Uncensored 7B 3.8GB ollama pull llama2-uncensored
Llama2 13B 13B 7.3GB ollama pull llama2:13b
Orca Mini 3B 1.9GB ollama pull orca
Vicuna 7B 3.8GB ollama pull vicuna
Nous-Hermes 13B 7.3GB ollama pull nous-hermes
Wizard Vicuna Uncensored 13B 7.3GB ollama pull wizard-vicuna

Note: You should have at least 8 GB of RAM to run the 3B models, 16 GB to run the 7B models, and 32 GB to run the 13B models.

Examples

Run a model

ollama run llama2
>>> hi
Hello! How can I help you today?

Create a custom model

Pull a base model:

ollama pull llama2

To update a model to the latest version, run ollama pull llama2 again. The model will be updated (if necessary).

Create a Modelfile:

FROM llama2

# set the temperature to 1 [higher is more creative, lower is more coherent]
PARAMETER temperature 1

# set the system prompt
SYSTEM """
You are Mario from Super Mario Bros. Answer as Mario, the assistant, only.
"""

Next, create and run the model:

ollama create mario -f ./Modelfile
ollama run mario
>>> hi
Hello! It's your friend Mario.

For more examples, see the examples directory.

For more information on creating a Modelfile, see the Modelfile documentation.

Pull a model from the registry

ollama pull orca

Listing local models

ollama list

Model packages

Overview

Ollama bundles model weights, configuration, and data into a single package, defined by a Modelfile.

logo

Building

go build .

To run it start the server:

./ollama serve &

Finally, run a model!

./ollama run llama2

REST API

POST /api/generate

Generate text from a model.

curl -X POST http://localhost:11434/api/generate -d '{"model": "llama2", "prompt":"Why is the sky blue?"}'

POST /api/create

Create a model from a Modelfile.

curl -X POST http://localhost:11434/api/create -d '{"name": "my-model", "path": "/path/to/modelfile"}'

Projects built with Ollama

  • Continue - embeds Ollama inside Visual Studio Code. The extension lets you highlight code to add to the prompt, ask questions in the sidebar, and generate code inline.
  • Discord AI Bot - interact with Ollama as a chatbot on Discord.
  • Raycast Ollama - Raycast extension to use Ollama for local llama inference on Raycast.
  • Simple HTML UI for Ollama