ext_server_windows.go 540 B

123456789101112131415
  1. package llm
  2. import (
  3. "fmt"
  4. "github.com/jmorganca/ollama/api"
  5. )
  6. func newDefaultExtServer(model string, adapters, projectors []string, opts api.Options) (extServer, error) {
  7. // On windows we always load the llama.cpp libraries dynamically to avoid startup DLL dependencies
  8. // This ensures we can update the PATH at runtime to get everything loaded
  9. // This should never happen as we'll always try to load one or more cpu dynamic libaries before hitting default
  10. return nil, fmt.Errorf("no available default llm library on windows")
  11. }