|
@@ -320,6 +320,10 @@ func NewLlamaServer(gpus discover.GpuInfoList, model string, f *ggml.GGML, adapt
|
|
return nil, fmt.Errorf("unable to lookup executable path: %w", err)
|
|
return nil, fmt.Errorf("unable to lookup executable path: %w", err)
|
|
}
|
|
}
|
|
|
|
|
|
|
|
+ if eval, err := filepath.EvalSymlinks(exe); err == nil {
|
|
|
|
+ exe = eval
|
|
|
|
+ }
|
|
|
|
+
|
|
// TODO - once fully switched to the Go runner, load the model here for tokenize/detokenize cgo access
|
|
// TODO - once fully switched to the Go runner, load the model here for tokenize/detokenize cgo access
|
|
s := &llmServer{
|
|
s := &llmServer{
|
|
port: port,
|
|
port: port,
|