ext_server.go 559 B

1234567891011121314151617
  1. //go:build !darwin
  2. package llm
  3. import (
  4. "fmt"
  5. "github.com/jmorganca/ollama/api"
  6. )
  7. func newDefaultExtServer(model string, adapters, projectors []string, opts api.Options) (extServer, error) {
  8. // On windows and linux we always load the llama.cpp libraries dynamically to avoid startup DLL dependencies
  9. // This ensures we can update the PATH at runtime to get everything loaded
  10. // This should never happen as we'll always try to load one or more cpu dynamic libaries before hitting default
  11. return nil, fmt.Errorf("no available default llm library")
  12. }