|
@@ -1,5 +1,5 @@
|
|
|
/**
|
|
|
- * llama.cpp - git ee459f40f65810a810151b24eba5b8bd174ceffe
|
|
|
+ * llama.cpp - git ee459f40f65810a810151b24eba5b8bd174ceffe - do not edit this file
|
|
|
*
|
|
|
* MIT License
|
|
|
*
|
|
@@ -377,8 +377,8 @@ static struct ggml_metal_context * ggml_metal_init(int n_cb) {
|
|
|
#if GGML_METAL_EMBED_LIBRARY
|
|
|
GGML_METAL_LOG_INFO("%s: using embedded metal library\n", __func__);
|
|
|
|
|
|
- extern const char* ggml_metallib_start;
|
|
|
- extern const char* ggml_metallib_end;
|
|
|
+ extern const char ggml_metallib_start[];
|
|
|
+ extern const char ggml_metallib_end[];
|
|
|
|
|
|
NSString * src = [[NSString alloc] initWithBytes:ggml_metallib_start length:(ggml_metallib_end-ggml_metallib_start) encoding:NSUTF8StringEncoding];
|
|
|
#else
|