Nessuna descrizione

Jeffrey Morgan 1178fd2cbb build with cmake 1 anno fa
api f723bf0879 ignore nil map values 1 anno fa
app 3d79b414d3 app: package `ggml-metal.metal` from correct directory 1 anno fa
cmd 0ebec07569 Merge pull request #345 from jmorganca/exit-non-zero 1 anno fa
docs 519f4d98ef add embed docs for modelfile 1 anno fa
examples 08265515b3 Merge pull request #303 from jmorganca/matt/dockerit 1 anno fa
format 9770e3b325 Generate private/public keypair for use w/ auth (#324) 1 anno fa
llm a894cc792d model and file type as strings 1 anno fa
parser 6517bcc53c Merge pull request #290 from jmorganca/add-adapter-layers 1 anno fa
progressbar e4d7f3e287 vendor in progress bar and change to bytes instead of bibytes (#130) 1 anno fa
scripts 639288bf2b make `ollama` binary executable on build 1 anno fa
server 086449b6c7 fmt 1 anno fa
vector a6f6d18f83 embed text document in modelfile 1 anno fa
.dockerignore 6292f4b64c update `Dockerfile` 1 anno fa
.gitignore 1178fd2cbb build with cmake 1 anno fa
.prettierrc.json 8685a5ad18 move .prettierrc.json to root 1 anno fa
CMakeLists.txt 1178fd2cbb build with cmake 1 anno fa
Dockerfile 7c71c10d4f fix compilation issue in Dockerfile, remove from `README.md` until ready 1 anno fa
LICENSE df5fdd6647 `proto` -> `ollama` 1 anno fa
README.md 8ca50f24f3 fix nous-hermes model file size listing in readme (#367) 1 anno fa
deps.sh 1178fd2cbb build with cmake 1 anno fa
go.mod d791df75dd check memory requirements before loading 1 anno fa
go.sum d791df75dd check memory requirements before loading 1 anno fa
main.go 76b85bc0e9 set non-zero error code on error 1 anno fa

README.md

logo

Ollama

Discord

Run, create, and share large language models (LLMs).

Note: Ollama is in early preview. Please report any issues you find.

Download

Quickstart

To run and chat with Llama 2, the new model by Meta:

ollama run llama2

Model library

Ollama supports a list of open-source models available on ollama.ai/library

Here are some example open-source models that can be downloaded:

Model Parameters Size Download
Llama2 7B 3.8GB ollama pull llama2
Llama2 13B 13B 7.3GB ollama pull llama2:13b
Llama2 70B 70B 39GB ollama pull llama2:70b
Llama2 Uncensored 7B 3.8GB ollama pull llama2-uncensored
Orca Mini 3B 1.9GB ollama pull orca-mini
Vicuna 7B 3.8GB ollama pull vicuna
Nous-Hermes 7B 3.8GB ollama pull nous-hermes
Nous-Hermes 13B 13B 7.3GB ollama pull nous-hermes:13b
Wizard Vicuna Uncensored 13B 7.3GB ollama pull wizard-vicuna

Note: You should have at least 8 GB of RAM to run the 3B models, 16 GB to run the 7B models, and 32 GB to run the 13B models.

Examples

Run a model

ollama run llama2
>>> hi
Hello! How can I help you today?

For multiline input, you can wrap text with """:

>>> """Hello,
... world!
... """
I'm a basic program that prints the famous "Hello, world!" message to the console.

Create a custom model

Pull a base model:

ollama pull llama2

To update a model to the latest version, run ollama pull llama2 again. The model will be updated (if necessary).

Create a Modelfile:

FROM llama2

# set the temperature to 1 [higher is more creative, lower is more coherent]
PARAMETER temperature 1

# set the system prompt
SYSTEM """
You are Mario from Super Mario Bros. Answer as Mario, the assistant, only.
"""

Next, create and run the model:

ollama create mario -f ./Modelfile
ollama run mario
>>> hi
Hello! It's your friend Mario.

For more examples, see the examples directory. For more information on creating a Modelfile, see the Modelfile documentation.

Pull a model from the registry

ollama pull orca

Listing local models

ollama list

Model packages

Overview

Ollama bundles model weights, configuration, and data into a single package, defined by a Modelfile.

logo

Building

go build .

To run it start the server:

./ollama serve &

Finally, run a model!

./ollama run llama2

REST API

See the API documentation for all endpoints.

Ollama has an API for running and managing models. For example to generate text from a model:

curl -X POST http://localhost:11434/api/generate -d '{
  "model": "llama2",
  "prompt":"Why is the sky blue?"
}'

Tools using Ollama