Browse Source

no gpu for 70B model

Michael Yang 1 year ago
parent
commit
74a5f7e698
1 changed files with 5 additions and 0 deletions
  1. 5 0
      llama/llama.go

+ 5 - 0
llama/llama.go

@@ -128,6 +128,11 @@ func New(model string, opts api.Options) (*LLM, error) {
 
 
 	C.llama_backend_init(C.bool(llm.UseNUMA))
 	C.llama_backend_init(C.bool(llm.UseNUMA))
 
 
+	// TODO: GQA == 8 suggests 70B model which doesn't support metal
+	if llm.NumGQA == 8 {
+		llm.NumGPU = 0
+	}
+
 	params := C.llama_context_default_params()
 	params := C.llama_context_default_params()
 	params.seed = C.uint(llm.Seed)
 	params.seed = C.uint(llm.Seed)
 	params.n_ctx = C.int(llm.NumCtx)
 	params.n_ctx = C.int(llm.NumCtx)