Bruce MacDonald
|
f221637053
first pass at linux gpu support (#454)
|
1 year ago |
Bruce MacDonald
|
09dd2aeff9
GGUF support (#441)
|
1 year ago |
Bruce MacDonald
|
42998d797d
subprocess llama.cpp server (#401)
|
1 year ago |
Quinn Slack
|
f4432e1dba
treat stop as stop sequences, not exact tokens (#442)
|
1 year ago |
Michael Yang
|
5ca05c2e88
fix ModelType()
|
1 year ago |
Michael Yang
|
a894cc792d
model and file type as strings
|
1 year ago |
Bruce MacDonald
|
4b2d366c37
Update llama.go
|
1 year ago |
Bruce MacDonald
|
56fd4e4ef2
log embedding eval timing
|
1 year ago |
Jeffrey Morgan
|
22885aeaee
update `llama.cpp` to `f64d44a`
|
1 year ago |
Michael Yang
|
6de5d032e1
implement loading ggml lora adapters through the modelfile
|
1 year ago |
Michael Yang
|
fccf8d179f
partial decode ggml bin for more info
|
1 year ago |