|
@@ -1186,8 +1186,6 @@ struct llama_server_context
|
|
|
{"model", params.model_alias},
|
|
|
{"tokens_predicted", slot.n_decoded},
|
|
|
{"tokens_evaluated", slot.n_prompt_tokens},
|
|
|
- {"generation_settings", get_formated_generation(slot)},
|
|
|
- {"prompt", slot.prompt},
|
|
|
{"truncated", slot.truncated},
|
|
|
{"stopped_eos", slot.stopped_eos},
|
|
|
{"stopped_word", slot.stopped_word},
|