Blake Mizerany acbffa59e9 llm: suppress large allocations for GGUF arrays há 11 meses atrás
..
ext_server 9d91e5e587 remove confusing log message há 11 meses atrás
generate 96624aa412 Merge pull request #5072 from dhiltgen/windows_path há 11 meses atrás
llama.cpp @ 7c26775adb 152fc202f5 llm: update llama.cpp commit to `7c26775` (#4896) há 11 meses atrás
patches 152fc202f5 llm: update llama.cpp commit to `7c26775` (#4896) há 11 meses atrás
filetype.go d6f692ad1a Add support for IQ1_S, IQ3_S, IQ2_S, IQ4_XS. IQ4_NL (#4322) há 1 ano atrás
ggla.go 171eb040fc simplify safetensors reading há 1 ano atrás
ggml.go acbffa59e9 llm: suppress large allocations for GGUF arrays há 11 meses atrás
gguf.go acbffa59e9 llm: suppress large allocations for GGUF arrays há 11 meses atrás
llm.go 829ff87bd1 revert tokenize ffi (#4761) há 11 meses atrás
llm_darwin_amd64.go 58d95cc9bd Switch back to subprocessing for llama.cpp há 1 ano atrás
llm_darwin_arm64.go 58d95cc9bd Switch back to subprocessing for llama.cpp há 1 ano atrás
llm_linux.go 58d95cc9bd Switch back to subprocessing for llama.cpp há 1 ano atrás
llm_windows.go 058f6cd2cc Move nested payloads to installer and zip file on windows há 1 ano atrás
memory.go 8e0641a9bf handle asymmetric embedding KVs há 11 meses atrás
memory_test.go 6f351bf586 review comments and coverage há 11 meses atrás
payload.go b2799f111b Move libraries out of users path há 11 meses atrás
server.go 5bf5aeec01 Refine mmap default logic on linux há 11 meses atrás
status.go 58d95cc9bd Switch back to subprocessing for llama.cpp há 1 ano atrás