|
@@ -1,8 +1,17 @@
|
|
|
|
+From 544a2d2e646d39e878d87dfbb3398a356bc560ab Mon Sep 17 00:00:00 2001
|
|
|
|
+From: Michael Yang <mxyng@pm.me>
|
|
|
|
+Date: Thu, 23 May 2024 11:18:45 -0700
|
|
|
|
+Subject: [PATCH] throw exception on load errors
|
|
|
|
+
|
|
|
|
+---
|
|
|
|
+ llama.cpp | 25 ++++++++++++++++---------
|
|
|
|
+ 1 file changed, 16 insertions(+), 9 deletions(-)
|
|
|
|
+
|
|
diff --git a/llama.cpp b/llama.cpp
|
|
diff --git a/llama.cpp b/llama.cpp
|
|
-index 4225f955..7b762f86 100644
|
|
|
|
|
|
+index 15c66077..8ba90b6a 100644
|
|
--- a/llama.cpp
|
|
--- a/llama.cpp
|
|
+++ b/llama.cpp
|
|
+++ b/llama.cpp
|
|
-@@ -4756,7 +4756,7 @@ static int llama_model_load(const std::string & fname, llama_model & model, llam
|
|
|
|
|
|
+@@ -6346,7 +6346,7 @@ static int llama_model_load(const std::string & fname, llama_model & model, llam
|
|
}
|
|
}
|
|
} catch (const std::exception & err) {
|
|
} catch (const std::exception & err) {
|
|
LLAMA_LOG_ERROR("%s: error loading model: %s\n", __func__, err.what());
|
|
LLAMA_LOG_ERROR("%s: error loading model: %s\n", __func__, err.what());
|
|
@@ -11,10 +20,10 @@ index 4225f955..7b762f86 100644
|
|
}
|
|
}
|
|
|
|
|
|
return 0;
|
|
return 0;
|
|
-@@ -12102,16 +12102,22 @@ struct llama_model * llama_load_model_from_file(
|
|
|
|
- };
|
|
|
|
|
|
+@@ -15600,16 +15600,23 @@ struct llama_model * llama_load_model_from_file(
|
|
|
|
+ }
|
|
|
|
+ model->rpc_servers.push_back(servers);
|
|
}
|
|
}
|
|
-
|
|
|
|
- int status = llama_model_load(path_model, *model, params);
|
|
- int status = llama_model_load(path_model, *model, params);
|
|
- GGML_ASSERT(status <= 0);
|
|
- GGML_ASSERT(status <= 0);
|
|
- if (status < 0) {
|
|
- if (status < 0) {
|
|
@@ -22,6 +31,7 @@ index 4225f955..7b762f86 100644
|
|
- LLAMA_LOG_ERROR("%s: failed to load model\n", __func__);
|
|
- LLAMA_LOG_ERROR("%s: failed to load model\n", __func__);
|
|
- } else if (status == -2) {
|
|
- } else if (status == -2) {
|
|
- LLAMA_LOG_INFO("%s: cancelled model load\n", __func__);
|
|
- LLAMA_LOG_INFO("%s: cancelled model load\n", __func__);
|
|
|
|
++
|
|
+ try {
|
|
+ try {
|
|
+ int status = llama_model_load(path_model, *model, params);
|
|
+ int status = llama_model_load(path_model, *model, params);
|
|
+ GGML_ASSERT(status <= 0);
|
|
+ GGML_ASSERT(status <= 0);
|
|
@@ -42,3 +52,6 @@ index 4225f955..7b762f86 100644
|
|
}
|
|
}
|
|
|
|
|
|
return model;
|
|
return model;
|
|
|
|
+--
|
|
|
|
+2.45.1
|
|
|
|
+
|