Bläddra i källkod

llm: do not error if LibOllamaPath does not exist (#8801)

Jeffrey Morgan 2 månader sedan
förälder
incheckning
50566113ac
1 ändrade filer med 3 tillägg och 14 borttagningar
  1. 3 14
      llm/server.go

+ 3 - 14
llm/server.go

@@ -89,7 +89,6 @@ func LoadModel(model string, maxArraySize int) (*GGML, error) {
 // NewLlamaServer will run a server for the given GPUs
 // The gpu list must be a single family.
 func NewLlamaServer(gpus discover.GpuInfoList, model string, ggml *GGML, adapters, projectors []string, opts api.Options, numParallel int) (LlamaServer, error) {
-	var err error
 	var systemTotalMemory uint64
 	var systemFreeMemory uint64
 	var systemSwapFreeMemory uint64
@@ -233,19 +232,9 @@ func NewLlamaServer(gpus discover.GpuInfoList, model string, ggml *GGML, adapter
 		params = append(params, "--multiuser-cache")
 	}
 
-	// get available libraries
-	if err != nil {
-		return nil, fmt.Errorf("could not get libollama dir: %w", err)
-	}
-
-	entries, err := os.ReadDir(discover.LibOllamaPath)
-	if err != nil {
-		return nil, fmt.Errorf("could not read libollama dir: %w", err)
-	}
-
 	libs := make(map[string]string)
-	for _, entry := range entries {
-		if entry.IsDir() {
+	if entries, err := os.ReadDir(discover.LibOllamaPath); err == nil {
+		for _, entry := range entries {
 			libs[entry.Name()] = filepath.Join(discover.LibOllamaPath, entry.Name())
 		}
 	}
@@ -285,7 +274,7 @@ func NewLlamaServer(gpus discover.GpuInfoList, model string, ggml *GGML, adapter
 			}
 		}
 		if port == 0 {
-			slog.Debug("ResolveTCPAddr failed ", "error", err)
+			slog.Debug("ResolveTCPAddr failed, using random port")
 			port = rand.Intn(65535-49152) + 49152 // get a random port in the ephemeral range
 		}
 		finalParams := []string{"runner"}