|
1 год назад | |
---|---|---|
api | 1 год назад | |
app | 1 год назад | |
cmd | 1 год назад | |
docs | 1 год назад | |
llama | 1 год назад | |
server | 1 год назад | |
web | 1 год назад | |
.dockerignore | 1 год назад | |
.gitignore | 1 год назад | |
.prettierrc.json | 1 год назад | |
Dockerfile | 1 год назад | |
LICENSE | 1 год назад | |
Makefile | 1 год назад | |
README.md | 1 год назад | |
go.mod | 1 год назад | |
go.sum | 1 год назад | |
main.go | 1 год назад | |
models.json | 1 год назад |
Run large language models with llama.cpp
.
Note: certain models that can be run with this project are intended for research and/or non-commercial use only.
docker run -p 11434:11434 ollama/ollama
You can also build the binary from source.
Run the model that started it all.
ollama run llama
Have a conversation.
ollama run vicuna "Why is the sky blue?"
Ask questions. Get answers.
ollama run orca "Write an email to my boss."
Sometimes you just need a little help writing code.
ollama run replit "Give me react code to render a button"
Venture into the unknown.
ollama run nous-hermes "Once upon a time"
ollama run ~/Downloads/vicuna-7b-v1.3.ggmlv3.q4_1.bin
make
To run it start the server:
./ollama server &
Finally, run a model!
./ollama run ~/Downloads/vicuna-7b-v1.3.ggmlv3.q4_1.bin
POST /api/pull
Download a model
curl -X POST http://localhost:11343/api/pull -d '{"model": "orca"}'
POST /api/generate
Complete a prompt
curl -X POST http://localhost:11434/api/generate -d '{"model": "orca", "prompt": "hello!", "stream": true}'