|
@@ -206,7 +206,7 @@ async def generate_chat_completion(
|
|
except Exception as e:
|
|
except Exception as e:
|
|
raise e
|
|
raise e
|
|
|
|
|
|
- if model["owned_by"] == "arena":
|
|
|
|
|
|
+ if model.get("owned_by") == "arena":
|
|
model_ids = model.get("info", {}).get("meta", {}).get("model_ids")
|
|
model_ids = model.get("info", {}).get("meta", {}).get("model_ids")
|
|
filter_mode = model.get("info", {}).get("meta", {}).get("filter_mode")
|
|
filter_mode = model.get("info", {}).get("meta", {}).get("filter_mode")
|
|
if model_ids and filter_mode == "exclude":
|
|
if model_ids and filter_mode == "exclude":
|
|
@@ -259,7 +259,7 @@ async def generate_chat_completion(
|
|
return await generate_function_chat_completion(
|
|
return await generate_function_chat_completion(
|
|
request, form_data, user=user, models=models
|
|
request, form_data, user=user, models=models
|
|
)
|
|
)
|
|
- if model["owned_by"] == "ollama":
|
|
|
|
|
|
+ if model.get("owned_by") == "ollama":
|
|
# Using /ollama/api/chat endpoint
|
|
# Using /ollama/api/chat endpoint
|
|
form_data = convert_payload_openai_to_ollama(form_data)
|
|
form_data = convert_payload_openai_to_ollama(form_data)
|
|
response = await generate_ollama_chat_completion(
|
|
response = await generate_ollama_chat_completion(
|