|
@@ -1,5 +1,5 @@
|
|
|
/**
|
|
|
- * llama.cpp - git 7c529cede6e84054e77a3eceab31c53de7b2f55b
|
|
|
+ * llama.cpp - git d91f3f0c55663719ea03b76311e8c36ed55eb0e2
|
|
|
*
|
|
|
* MIT License
|
|
|
*
|
|
@@ -3689,7 +3689,7 @@ size_t llama_copy_state_data(struct llama_context * ctx, uint8_t * dst) {
|
|
|
const auto & kv_self = ctx->kv_self;
|
|
|
const auto & hparams = ctx->model.hparams;
|
|
|
const int n_layer = hparams.n_layer;
|
|
|
- const int n_embd = hparams.n_embd;
|
|
|
+ const int n_embd = hparams.n_embd_gqa();
|
|
|
const int n_ctx = hparams.n_ctx;
|
|
|
|
|
|
const size_t kv_size = kv_self.buf.size;
|
|
@@ -3792,7 +3792,7 @@ size_t llama_set_state_data(struct llama_context * ctx, uint8_t * src) {
|
|
|
const auto & kv_self = ctx->kv_self;
|
|
|
const auto & hparams = ctx->model.hparams;
|
|
|
const int n_layer = hparams.n_layer;
|
|
|
- const int n_embd = hparams.n_embd;
|
|
|
+ const int n_embd = hparams.n_embd_gqa();
|
|
|
const int n_ctx = hparams.n_ctx;
|
|
|
|
|
|
size_t kv_size;
|