Keine Beschreibung

Patrick Devine 5c26b81a2f skip files in the list if we can't get the correct model path vor 1 Jahr
api 5bea29f610 add new list command (#97) vor 1 Jahr
app a9cc270b4d icon update vor 1 Jahr
cmd 885f67a471 Merge pull request #92 from jmorganca/create-model-spinner vor 1 Jahr
docs 9310ee3967 First stab at a modelfile doc vor 1 Jahr
examples 3d9498dc95 Some simple modelfile examples vor 1 Jahr
format 5bea29f610 add new list command (#97) vor 1 Jahr
llama 40c9dc0a31 fix multibyte responses vor 1 Jahr
parser 1f45f7bb52 convert commands to uppercase in parser vor 1 Jahr
scripts 4dd296e155 build app in publish script vor 1 Jahr
server 5c26b81a2f skip files in the list if we can't get the correct model path vor 1 Jahr
web 787d965331 web: disable signup button while submitting vor 1 Jahr
.dockerignore 6292f4b64c update `Dockerfile` vor 1 Jahr
.gitignore 7c71c10d4f fix compilation issue in Dockerfile, remove from `README.md` until ready vor 1 Jahr
.prettierrc.json 8685a5ad18 move .prettierrc.json to root vor 1 Jahr
Dockerfile 7c71c10d4f fix compilation issue in Dockerfile, remove from `README.md` until ready vor 1 Jahr
LICENSE df5fdd6647 `proto` -> `ollama` vor 1 Jahr
README.md 529ff9ab6d Add note to `README.md` about Apple Silicon support vor 1 Jahr
ggml-metal.metal e64ef69e34 look for ggml-metal in the same directory as the binary vor 1 Jahr
go.mod 5bea29f610 add new list command (#97) vor 1 Jahr
go.sum 5bea29f610 add new list command (#97) vor 1 Jahr
main.go 1775647f76 continue conversation vor 1 Jahr
models.json 5028de2901 update vicuna model vor 1 Jahr

README.md

ollama

Ollama

Run large language models with llama.cpp.

Note: certain models that can be run with Ollama are intended for research and/or non-commercial use only.

Features

  • Download and run popular large language models
  • Switch between multiple models on the fly
  • Hardware acceleration where available (Metal, CUDA)
  • Fast inference server written in Go, powered by llama.cpp
  • REST API to use with your application (python, typescript SDKs coming soon)

Install

  • Download for macOS with Apple Silicon (Intel coming soon)
  • Download for Windows (coming soon)

You can also build the binary from source.

Quickstart

Run a fast and simple model.

ollama run orca

Example models

💬 Chat

Have a conversation.

ollama run vicuna "Why is the sky blue?"

🗺️ Instructions

Get a helping hand.

ollama run orca "Write an email to my boss."

🔎 Ask questions about documents

Send the contents of a document and ask questions about it.

ollama run nous-hermes "$(cat input.txt)", please summarize this story

📖 Storytelling

Venture into the unknown.

ollama run nous-hermes "Once upon a time"

Advanced usage

Run a local model

ollama run ~/Downloads/vicuna-7b-v1.3.ggmlv3.q4_1.bin

Building

go build .

To run it start the server:

./ollama server &

Finally, run a model!

./ollama run ~/Downloads/vicuna-7b-v1.3.ggmlv3.q4_1.bin

API Reference

POST /api/pull

Download a model

curl -X POST http://localhost:11343/api/pull -d '{"model": "orca"}'

POST /api/generate

Complete a prompt

curl -X POST http://localhost:11434/api/generate -d '{"model": "orca", "prompt": "hello!"}'