|
@@ -1206,7 +1206,6 @@ struct llama_server_context
|
|
|
res.result_json = json
|
|
|
{
|
|
|
{"embedding", std::vector<float>(n_embd, 0.0f)},
|
|
|
- {"truncated", slot.truncated}
|
|
|
};
|
|
|
}
|
|
|
else
|
|
@@ -1224,7 +1223,6 @@ struct llama_server_context
|
|
|
res.result_json = json
|
|
|
{
|
|
|
{"embedding", std::vector<float>(n_embd, 0.0f)},
|
|
|
- {"truncated", slot.truncated}
|
|
|
};
|
|
|
continue;
|
|
|
}
|
|
@@ -1233,7 +1231,6 @@ struct llama_server_context
|
|
|
res.result_json = json
|
|
|
{
|
|
|
{"embedding", std::vector<float>(embd, embd + n_embd)},
|
|
|
- {"truncated", slot.truncated}
|
|
|
};
|
|
|
}
|
|
|
}
|
|
@@ -3063,7 +3060,6 @@ int main(int argc, char **argv) {
|
|
|
if (!json_value(data, "stream", false)) {
|
|
|
std::string completion_text;
|
|
|
task_result result = llama.queue_results.recv(task_id);
|
|
|
- LOG_INFO("completion", {{"result", result.result_json}});
|
|
|
if (!result.error && result.stop) {
|
|
|
res.set_content(result.result_json.dump(-1, ' ', false, json::error_handler_t::replace), "application/json; charset=utf-8");
|
|
|
}
|
|
@@ -3079,7 +3075,6 @@ int main(int argc, char **argv) {
|
|
|
while (true)
|
|
|
{
|
|
|
task_result result = llama.queue_results.recv(task_id);
|
|
|
- LOG_INFO("completion", {{"result", result.result_json}});
|
|
|
if (!result.error) {
|
|
|
const std::string str =
|
|
|
"data: " +
|