Explorar el Código

move prompt templates out of python bindings

Jeffrey Morgan hace 1 año
padre
commit
9164981d72

+ 4 - 4
cmd/cmd.go

@@ -59,7 +59,7 @@ func NewCLI() *cobra.Command {
 
 	rootCmd := &cobra.Command{
 		Use:   "ollama",
-		Short: "Run any large language model on any machine.",
+		Short: "Large language model runner",
 		CompletionOptions: cobra.CompletionOptions{
 			DisableDefaultCmd: true,
 		},
@@ -88,14 +88,13 @@ func NewCLI() *cobra.Command {
 		},
 	}
 
-/*
 	runCmd := &cobra.Command{
 		Use: "run",
 		Short: "Run a model and submit prompts.",
-		RunE: func(cmd *cobra.Command. args []string) error {
+		RunE: func(cmd *cobra.Command,args []string) error {
+			return nil
 		},
 	}
-*/
 
 	serveCmd := &cobra.Command{
 		Use:     "serve",
@@ -132,6 +131,7 @@ func NewCLI() *cobra.Command {
 	rootCmd.AddCommand(
 		modelsCmd,
 		serveCmd,
+		runCmd,
 	)
 
 	return rootCmd

+ 1 - 0
llama/CMakeLists.txt

@@ -23,5 +23,6 @@ target_include_directories(binding PRIVATE ${llama_cpp_SOURCE_DIR})
 target_include_directories(binding PRIVATE ${llama_cpp_SOURCE_DIR}/examples)
 target_link_libraries(binding llama ggml_static)
 
+configure_file(${llama_cpp_SOURCE_DIR}/ggml-metal.metal ${CMAKE_CURRENT_BINARY_DIR}/ggml-metal.metal COPYONLY)
 configure_file(${llama_cpp_BINARY_DIR}/libllama.a ${CMAKE_CURRENT_BINARY_DIR}/libllama.a COPYONLY)
 configure_file(${llama_cpp_BINARY_DIR}/libggml_static.a ${CMAKE_CURRENT_BINARY_DIR}/libggml_static.a COPYONLY)

+ 6 - 7
server/routes.go

@@ -36,6 +36,8 @@ func Serve(ln net.Listener) error {
 	})
 
 	r.POST("/api/generate", func(c *gin.Context) {
+		// TODO: set prompt from template
+
 		var req api.GenerateRequest
 		if err := c.ShouldBindJSON(&req); err != nil {
 			c.JSON(http.StatusBadRequest, gin.H{"message": err.Error()})
@@ -64,13 +66,10 @@ func Serve(ln net.Listener) error {
 			return true
 		})
 
-/*
-                embeds, err := l.Embeddings(text)
-                if err != nil {
-                        fmt.Printf("Embeddings: error %s \n", err.Error())
-                }
-*/
-		
+		// embeds, err := l.Embeddings(text)
+		// if err != nil {
+		//         fmt.Printf("Embeddings: error %s \n", err.Error())
+		// }
 	})
 
 	log.Printf("Listening on %s", ln.Addr())

+ 0 - 0
python/ollama/templates/alpaca.prompt → templates/alpaca.prompt


+ 0 - 0
python/ollama/templates/falcon.prompt → templates/falcon.prompt


+ 0 - 0
python/ollama/templates/gpt4.prompt → templates/gpt4.prompt


+ 0 - 0
python/ollama/templates/hermes.prompt → templates/hermes.prompt


+ 0 - 0
python/ollama/templates/mpt.prompt → templates/mpt.prompt


+ 0 - 0
python/ollama/templates/oasst.prompt → templates/oasst.prompt


+ 0 - 0
python/ollama/templates/orca.prompt → templates/orca.prompt


+ 0 - 0
python/ollama/templates/qlora.prompt → templates/qlora.prompt


+ 0 - 0
python/ollama/templates/tulu.prompt → templates/tulu.prompt


+ 0 - 0
python/ollama/templates/ultralm.prompt → templates/ultralm.prompt


+ 0 - 0
python/ollama/templates/vicuna.prompt → templates/vicuna.prompt


+ 0 - 0
python/ollama/templates/wizardcoder.prompt → templates/wizardcoder.prompt


+ 0 - 0
python/ollama/templates/wizardlm.prompt → templates/wizardlm.prompt