Browse Source

chore: format

Timothy J. Baek 10 months ago
parent
commit
e8fc522eba
2 changed files with 21 additions and 14 deletions
  1. 11 4
      backend/apps/rag/main.py
  2. 10 10
      backend/apps/rag/search/serply.py

+ 11 - 4
backend/apps/rag/main.py

@@ -831,7 +831,9 @@ def search_web(engine: str, query: str) -> list[SearchResult]:
 @app.post("/web/search")
 def store_web_search(form_data: SearchForm, user=Depends(get_current_user)):
     try:
-        logging.info(f"trying to web search with {app.state.config.RAG_WEB_SEARCH_ENGINE, form_data.query}")
+        logging.info(
+            f"trying to web search with {app.state.config.RAG_WEB_SEARCH_ENGINE, form_data.query}"
+        )
         web_results = search_web(
             app.state.config.RAG_WEB_SEARCH_ENGINE, form_data.query
         )
@@ -1241,8 +1243,10 @@ def reset(user=Depends(get_admin_user)) -> bool:
 
     return True
 
+
 class SafeWebBaseLoader(WebBaseLoader):
     """WebBaseLoader with enhanced error handling for URLs."""
+
     def lazy_load(self) -> Iterator[Document]:
         """Lazy load text from the url(s) in web_path with error handling."""
         for path in self.web_paths:
@@ -1255,15 +1259,18 @@ class SafeWebBaseLoader(WebBaseLoader):
                 if title := soup.find("title"):
                     metadata["title"] = title.get_text()
                 if description := soup.find("meta", attrs={"name": "description"}):
-                    metadata["description"] = description.get("content", "No description found.")
+                    metadata["description"] = description.get(
+                        "content", "No description found."
+                    )
                 if html := soup.find("html"):
                     metadata["language"] = html.get("lang", "No language found.")
-                
+
                 yield Document(page_content=text, metadata=metadata)
             except Exception as e:
                 # Log the error and continue with the next URL
                 log.error(f"Error loading {path}: {e}")
-                
+
+
 if ENV == "dev":
 
     @app.get("/ef")

+ 10 - 10
backend/apps/rag/search/serply.py

@@ -12,14 +12,14 @@ log.setLevel(SRC_LOG_LEVELS["RAG"])
 
 
 def search_serply(
-        api_key: str,
-        query: str,
-        count: int,
-        hl: str = "us",
-        limit: int = 10,
-        device_type: str = "desktop",
-        proxy_location: str = "US"
-    ) -> list[SearchResult]:
+    api_key: str,
+    query: str,
+    count: int,
+    hl: str = "us",
+    limit: int = 10,
+    device_type: str = "desktop",
+    proxy_location: str = "US",
+) -> list[SearchResult]:
     """Search using serper.dev's API and return the results as a list of SearchResult objects.
 
     Args:
@@ -37,7 +37,7 @@ def search_serply(
         "language": "en",
         "num": limit,
         "gl": proxy_location.upper(),
-        "hl": hl.lower()
+        "hl": hl.lower(),
     }
 
     url = f"{url}{urlencode(query_payload)}"
@@ -45,7 +45,7 @@ def search_serply(
         "X-API-KEY": api_key,
         "X-User-Agent": device_type,
         "User-Agent": "open-webui",
-        "X-Proxy-Location": proxy_location
+        "X-Proxy-Location": proxy_location,
     }
 
     response = requests.request("GET", url, headers=headers)