Ver código fonte

Remove mapping of max_completion_tokens

1) max_completion_tokens is being looked for in openai_payload, but would be located in openai_payload['options'], so is never found. (This applies to the prior two commits as well).
2) max_completion_tokens is not sent from the frontend, only max_tokens. It does not appear in AdvancedParams.svelte.
2b) Openai.py does use max_completion_tokens, but for o1,o3 models and converts it from max_tokens.
ferret99gt 2 meses atrás
pai
commit
aea8977d05
1 arquivos alterados com 1 adições e 3 exclusões
  1. 1 3
      backend/open_webui/utils/payload.py

+ 1 - 3
backend/open_webui/utils/payload.py

@@ -185,9 +185,7 @@ def convert_payload_openai_to_ollama(openai_payload: dict) -> dict:
             ollama_options[param] = openai_payload[param]
             ollama_options[param] = openai_payload[param]
 
 
     # Mapping OpenAI's `max_tokens` -> Ollama's `num_predict`
     # Mapping OpenAI's `max_tokens` -> Ollama's `num_predict`
-    if "max_completion_tokens" in openai_payload:
-        ollama_options["num_predict"] = openai_payload["max_completion_tokens"]
-    elif "max_tokens" in openai_payload:
+    if "max_tokens" in openai_payload:
         ollama_options["num_predict"] = openai_payload["max_tokens"]
         ollama_options["num_predict"] = openai_payload["max_tokens"]
 
 
     # Add options to payload if any have been set
     # Add options to payload if any have been set