Sen descrición

Jeffrey Morgan 268e362fa7 fix binding build hai 1 ano
api f533f85d44 pr feedback hai 1 ano
app 840f87430a remove double `heartbeat` hai 1 ano
cmd 07a4c1e3fb take all args as one prompt hai 1 ano
docs 1358e27b77 add publish script hai 1 ano
examples 993cb9fad6 examples: add basic python example hai 1 ano
llama 268e362fa7 fix binding build hai 1 ano
scripts 39e946f256 fix download url hai 1 ano
server 74e92d1258 add basic `/` route for server hai 1 ano
web 20dae6b38f add `vercel.json` to silence PR comments hai 1 ano
.dockerignore 6292f4b64c update `Dockerfile` hai 1 ano
.gitignore 9fe018675f use `Makefile` for dependency building instead of `go generate` hai 1 ano
.prettierrc.json 8685a5ad18 move .prettierrc.json to root hai 1 ano
Dockerfile ea809df196 update `Dockerfile` to use OLLAMA_HOST hai 1 ano
LICENSE df5fdd6647 `proto` -> `ollama` hai 1 ano
Makefile 1358e27b77 add publish script hai 1 ano
README.md 07a4c1e3fb take all args as one prompt hai 1 ano
go.mod c4b9e84945 progress hai 1 ano
go.sum c4b9e84945 progress hai 1 ano
main.go 6093a88c1a add llama.cpp go bindings hai 1 ano
models.json 962d351281 remove replit example which does not run currently hai 1 ano

README.md

ollama

Ollama

Run large language models with llama.cpp.

Note: certain models that can be run with Ollama are intended for research and/or non-commercial use only.

Features

  • Download and run popular large language models
  • Switch between multiple models on the fly
  • Hardware acceleration where available (Metal, CUDA)
  • Fast inference server written in Go, powered by llama.cpp
  • REST API to use with your application (python, typescript SDKs coming soon)

Install

  • Download for macOS
  • Download for Windows (coming soon)
  • Docker: docker run -p 11434:11434 ollama/ollama

You can also build the binary from source.

Quickstart

Run a fast and simple model.

ollama run orca

Example models

💬 Chat

Have a conversation.

ollama run vicuna "Why is the sky blue?"

🗺️ Instructions

Get a helping hand.

ollama run orca "Write an email to my boss."

🔎 Ask questions about documents

Send the contents of a document and ask questions about it.

ollama run nous-hermes "$(cat input.txt)", please summarize this story

📖 Storytelling

Venture into the unknown.

ollama run nous-hermes "Once upon a time"

Advanced usage

Run a local model

ollama run ~/Downloads/vicuna-7b-v1.3.ggmlv3.q4_1.bin

Building

make

To run it start the server:

./ollama server &

Finally, run a model!

./ollama run ~/Downloads/vicuna-7b-v1.3.ggmlv3.q4_1.bin

API Reference

POST /api/pull

Download a model

curl -X POST http://localhost:11343/api/pull -d '{"model": "orca"}'

POST /api/generate

Complete a prompt

curl -X POST http://localhost:11434/api/generate -d '{"model": "orca", "prompt": "hello!", "stream": true}'