chat.py 9.7 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316
  1. import time
  2. import logging
  3. import sys
  4. from aiocache import cached
  5. from typing import Any, Optional
  6. import random
  7. import json
  8. import inspect
  9. from fastapi import Request
  10. from starlette.responses import Response, StreamingResponse
  11. from open_webui.models.users import UserModel
  12. from open_webui.socket.main import (
  13. get_event_call,
  14. get_event_emitter,
  15. )
  16. from open_webui.functions import generate_function_chat_completion
  17. from open_webui.routers.openai import (
  18. generate_chat_completion as generate_openai_chat_completion,
  19. )
  20. from open_webui.routers.ollama import (
  21. generate_chat_completion as generate_ollama_chat_completion,
  22. )
  23. from open_webui.routers.pipelines import (
  24. process_pipeline_inlet_filter,
  25. process_pipeline_outlet_filter,
  26. )
  27. from open_webui.models.functions import Functions
  28. from open_webui.models.models import Models
  29. from open_webui.utils.plugin import load_function_module_by_id
  30. from open_webui.utils.models import get_all_models, check_model_access
  31. from open_webui.utils.payload import convert_payload_openai_to_ollama
  32. from open_webui.utils.response import (
  33. convert_response_ollama_to_openai,
  34. convert_streaming_response_ollama_to_openai,
  35. )
  36. from open_webui.utils.filter import (
  37. get_sorted_filter_ids,
  38. process_filter_functions,
  39. )
  40. from open_webui.env import SRC_LOG_LEVELS, GLOBAL_LOG_LEVEL, BYPASS_MODEL_ACCESS_CONTROL
  41. logging.basicConfig(stream=sys.stdout, level=GLOBAL_LOG_LEVEL)
  42. log = logging.getLogger(__name__)
  43. log.setLevel(SRC_LOG_LEVELS["MAIN"])
  44. async def generate_chat_completion(
  45. request: Request,
  46. form_data: dict,
  47. user: Any,
  48. bypass_filter: bool = False,
  49. ):
  50. if BYPASS_MODEL_ACCESS_CONTROL:
  51. bypass_filter = True
  52. models = request.app.state.MODELS
  53. model_id = form_data["model"]
  54. if model_id not in models:
  55. raise Exception("Model not found")
  56. # Process the form_data through the pipeline
  57. try:
  58. form_data = process_pipeline_inlet_filter(request, form_data, user, models)
  59. except Exception as e:
  60. raise e
  61. model = models[model_id]
  62. # Check if user has access to the model
  63. if not bypass_filter and user.role == "user":
  64. try:
  65. check_model_access(user, model)
  66. except Exception as e:
  67. raise e
  68. if model["owned_by"] == "arena":
  69. model_ids = model.get("info", {}).get("meta", {}).get("model_ids")
  70. filter_mode = model.get("info", {}).get("meta", {}).get("filter_mode")
  71. if model_ids and filter_mode == "exclude":
  72. model_ids = [
  73. model["id"]
  74. for model in list(request.app.state.MODELS.values())
  75. if model.get("owned_by") != "arena" and model["id"] not in model_ids
  76. ]
  77. selected_model_id = None
  78. if isinstance(model_ids, list) and model_ids:
  79. selected_model_id = random.choice(model_ids)
  80. else:
  81. model_ids = [
  82. model["id"]
  83. for model in list(request.app.state.MODELS.values())
  84. if model.get("owned_by") != "arena"
  85. ]
  86. selected_model_id = random.choice(model_ids)
  87. form_data["model"] = selected_model_id
  88. if form_data.get("stream") == True:
  89. async def stream_wrapper(stream):
  90. yield f"data: {json.dumps({'selected_model_id': selected_model_id})}\n\n"
  91. async for chunk in stream:
  92. yield chunk
  93. response = await generate_chat_completion(
  94. request, form_data, user, bypass_filter=True
  95. )
  96. return StreamingResponse(
  97. stream_wrapper(response.body_iterator),
  98. media_type="text/event-stream",
  99. background=response.background,
  100. )
  101. else:
  102. return {
  103. **(
  104. await generate_chat_completion(
  105. request, form_data, user, bypass_filter=True
  106. )
  107. ),
  108. "selected_model_id": selected_model_id,
  109. }
  110. if model.get("pipe"):
  111. # Below does not require bypass_filter because this is the only route the uses this function and it is already bypassing the filter
  112. return await generate_function_chat_completion(
  113. request, form_data, user=user, models=models
  114. )
  115. if model["owned_by"] == "ollama":
  116. # Using /ollama/api/chat endpoint
  117. form_data = convert_payload_openai_to_ollama(form_data)
  118. response = await generate_ollama_chat_completion(
  119. request=request, form_data=form_data, user=user, bypass_filter=bypass_filter
  120. )
  121. if form_data.get("stream"):
  122. response.headers["content-type"] = "text/event-stream"
  123. return StreamingResponse(
  124. convert_streaming_response_ollama_to_openai(response),
  125. headers=dict(response.headers),
  126. background=response.background,
  127. )
  128. else:
  129. return convert_response_ollama_to_openai(response)
  130. else:
  131. return await generate_openai_chat_completion(
  132. request=request, form_data=form_data, user=user, bypass_filter=bypass_filter
  133. )
  134. chat_completion = generate_chat_completion
  135. async def chat_completed(request: Request, form_data: dict, user: Any):
  136. if not request.app.state.MODELS:
  137. await get_all_models(request)
  138. models = request.app.state.MODELS
  139. data = form_data
  140. model_id = data["model"]
  141. if model_id not in models:
  142. raise Exception("Model not found")
  143. model = models[model_id]
  144. try:
  145. data = process_pipeline_outlet_filter(request, data, user, models)
  146. except Exception as e:
  147. return Exception(f"Error: {e}")
  148. metadata = {
  149. "chat_id": data["chat_id"],
  150. "message_id": data["id"],
  151. "session_id": data["session_id"],
  152. "user_id": user.id,
  153. }
  154. extra_params = {
  155. "__event_emitter__": get_event_emitter(metadata),
  156. "__event_call__": get_event_call(metadata),
  157. "__user__": {
  158. "id": user.id,
  159. "email": user.email,
  160. "name": user.name,
  161. "role": user.role,
  162. },
  163. "__metadata__": metadata,
  164. "__request__": request,
  165. "__model__": model,
  166. }
  167. try:
  168. result, _ = await process_filter_functions(
  169. request=request,
  170. filter_ids=get_sorted_filter_ids(model),
  171. filter_type="outlet",
  172. form_data=data,
  173. extra_params=extra_params,
  174. )
  175. return result
  176. except Exception as e:
  177. return Exception(f"Error: {e}")
  178. async def chat_action(request: Request, action_id: str, form_data: dict, user: Any):
  179. if "." in action_id:
  180. action_id, sub_action_id = action_id.split(".")
  181. else:
  182. sub_action_id = None
  183. action = Functions.get_function_by_id(action_id)
  184. if not action:
  185. raise Exception(f"Action not found: {action_id}")
  186. if not request.app.state.MODELS:
  187. await get_all_models(request)
  188. models = request.app.state.MODELS
  189. data = form_data
  190. model_id = data["model"]
  191. if model_id not in models:
  192. raise Exception("Model not found")
  193. model = models[model_id]
  194. __event_emitter__ = get_event_emitter(
  195. {
  196. "chat_id": data["chat_id"],
  197. "message_id": data["id"],
  198. "session_id": data["session_id"],
  199. "user_id": user.id,
  200. }
  201. )
  202. __event_call__ = get_event_call(
  203. {
  204. "chat_id": data["chat_id"],
  205. "message_id": data["id"],
  206. "session_id": data["session_id"],
  207. "user_id": user.id,
  208. }
  209. )
  210. if action_id in request.app.state.FUNCTIONS:
  211. function_module = request.app.state.FUNCTIONS[action_id]
  212. else:
  213. function_module, _, _ = load_function_module_by_id(action_id)
  214. request.app.state.FUNCTIONS[action_id] = function_module
  215. if hasattr(function_module, "valves") and hasattr(function_module, "Valves"):
  216. valves = Functions.get_function_valves_by_id(action_id)
  217. function_module.valves = function_module.Valves(**(valves if valves else {}))
  218. if hasattr(function_module, "action"):
  219. try:
  220. action = function_module.action
  221. # Get the signature of the function
  222. sig = inspect.signature(action)
  223. params = {"body": data}
  224. # Extra parameters to be passed to the function
  225. extra_params = {
  226. "__model__": model,
  227. "__id__": sub_action_id if sub_action_id is not None else action_id,
  228. "__event_emitter__": __event_emitter__,
  229. "__event_call__": __event_call__,
  230. "__request__": request,
  231. }
  232. # Add extra params in contained in function signature
  233. for key, value in extra_params.items():
  234. if key in sig.parameters:
  235. params[key] = value
  236. if "__user__" in sig.parameters:
  237. __user__ = {
  238. "id": user.id,
  239. "email": user.email,
  240. "name": user.name,
  241. "role": user.role,
  242. }
  243. try:
  244. if hasattr(function_module, "UserValves"):
  245. __user__["valves"] = function_module.UserValves(
  246. **Functions.get_user_valves_by_id_and_user_id(
  247. action_id, user.id
  248. )
  249. )
  250. except Exception as e:
  251. print(e)
  252. params = {**params, "__user__": __user__}
  253. if inspect.iscoroutinefunction(action):
  254. data = await action(**params)
  255. else:
  256. data = action(**params)
  257. except Exception as e:
  258. return Exception(f"Error: {e}")
  259. return data