.. |
ext_server
|
fcf4d60eee
llm: add back check for empty token cache
|
1 年之前 |
generate
|
8a65717f55
Do not build AVX runners on ARM64
|
1 年之前 |
llama.cpp @ 952d03dbea
|
e33d5c2dbc
update llama.cpp commit to `952d03d`
|
1 年之前 |
patches
|
85801317d1
Fix clip log import
|
1 年之前 |
ggla.go
|
8b2c10061c
refactor tensor query
|
1 年之前 |
ggml.go
|
435cc866a3
fix: mixtral graph
|
1 年之前 |
gguf.go
|
14476d48cc
fixes for gguf (#3863)
|
1 年之前 |
llm.go
|
86e67fc4a9
Add import declaration for windows,arm64 to llm.go
|
1 年之前 |
llm_darwin_amd64.go
|
58d95cc9bd
Switch back to subprocessing for llama.cpp
|
1 年之前 |
llm_darwin_arm64.go
|
58d95cc9bd
Switch back to subprocessing for llama.cpp
|
1 年之前 |
llm_linux.go
|
58d95cc9bd
Switch back to subprocessing for llama.cpp
|
1 年之前 |
llm_windows.go
|
058f6cd2cc
Move nested payloads to installer and zip file on windows
|
1 年之前 |
memory.go
|
f0c454ab57
gpu: add 512MiB to darwin minimum, metal doesn't have partial offloading overhead (#4068)
|
1 年之前 |
payload.go
|
058f6cd2cc
Move nested payloads to installer and zip file on windows
|
1 年之前 |
server.go
|
321d57e1a0
Removing go routine calling .wait from load.
|
1 年之前 |
status.go
|
58d95cc9bd
Switch back to subprocessing for llama.cpp
|
1 年之前 |