Blake Mizerany acbffa59e9 llm: suppress large allocations for GGUF arrays 10 mesi fa
..
ext_server 9d91e5e587 remove confusing log message 10 mesi fa
generate 96624aa412 Merge pull request #5072 from dhiltgen/windows_path 10 mesi fa
llama.cpp @ 7c26775adb 152fc202f5 llm: update llama.cpp commit to `7c26775` (#4896) 10 mesi fa
patches 152fc202f5 llm: update llama.cpp commit to `7c26775` (#4896) 10 mesi fa
filetype.go d6f692ad1a Add support for IQ1_S, IQ3_S, IQ2_S, IQ4_XS. IQ4_NL (#4322) 11 mesi fa
ggla.go 171eb040fc simplify safetensors reading 11 mesi fa
ggml.go acbffa59e9 llm: suppress large allocations for GGUF arrays 10 mesi fa
gguf.go acbffa59e9 llm: suppress large allocations for GGUF arrays 10 mesi fa
llm.go 829ff87bd1 revert tokenize ffi (#4761) 11 mesi fa
llm_darwin_amd64.go 58d95cc9bd Switch back to subprocessing for llama.cpp 1 anno fa
llm_darwin_arm64.go 58d95cc9bd Switch back to subprocessing for llama.cpp 1 anno fa
llm_linux.go 58d95cc9bd Switch back to subprocessing for llama.cpp 1 anno fa
llm_windows.go 058f6cd2cc Move nested payloads to installer and zip file on windows 1 anno fa
memory.go 8e0641a9bf handle asymmetric embedding KVs 10 mesi fa
memory_test.go 6f351bf586 review comments and coverage 10 mesi fa
payload.go b2799f111b Move libraries out of users path 10 mesi fa
server.go 5bf5aeec01 Refine mmap default logic on linux 10 mesi fa
status.go 58d95cc9bd Switch back to subprocessing for llama.cpp 1 anno fa