123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130 |
- from fastapi import APIRouter, UploadFile, File, BackgroundTasks
- from fastapi import Depends, HTTPException, status
- from starlette.responses import StreamingResponse
- from pydantic import BaseModel
- from utils.misc import calculate_sha256
- import requests
- import os
- import asyncio
- import json
- from config import OLLAMA_API_BASE_URL
- router = APIRouter()
- class UploadBlobForm(BaseModel):
- filename: str
- from urllib.parse import urlparse
- def parse_huggingface_url(hf_url):
- # Parse the URL
- parsed_url = urlparse(hf_url)
- # Get the path and split it into components
- path_components = parsed_url.path.split("/")
- # Extract the desired output
- user_repo = "/".join(path_components[1:3])
- model_file = path_components[-1]
- return [user_repo, model_file]
- def download_file_stream(url, file_path, chunk_size=1024 * 1024):
- done = False
- if os.path.exists(file_path):
- current_size = os.path.getsize(file_path)
- else:
- current_size = 0
- headers = {"Range": f"bytes={current_size}-"} if current_size > 0 else {}
- with requests.get(url, headers=headers, stream=True) as response:
- total_size = int(response.headers.get("content-length", 0)) + current_size
- with open(file_path, "ab") as file:
- for data in response.iter_content(chunk_size=chunk_size):
- current_size += len(data)
- file.write(data)
- done = current_size == total_size
- progress = round((current_size / total_size) * 100, 2)
- yield f'data: {{"progress": {progress}, "current": {current_size}, "total": {total_size}}}\n\n'
- @router.get("/download")
- async def download(
- url: str = "https://huggingface.co/TheBloke/stablelm-zephyr-3b-GGUF/resolve/main/stablelm-zephyr-3b.Q2_K.gguf",
- ):
- user_repo, model_file = parse_huggingface_url(url)
- os.makedirs("./uploads", exist_ok=True)
- file_path = os.path.join("./uploads", f"{model_file}")
- return StreamingResponse(
- download_file_stream(url, file_path), media_type="text/event-stream"
- )
- @router.post("/upload")
- async def upload(file: UploadFile = File(...)):
- os.makedirs("./uploads", exist_ok=True)
- file_path = os.path.join("./uploads", file.filename)
- async def file_write_stream():
- total = 0
- total_size = file.size
- chunk_size = 1024 * 1024
- done = False
- try:
- with open(file_path, "wb+") as f:
- while True:
- chunk = file.file.read(chunk_size)
- if not chunk:
- break
- f.write(chunk)
- total += len(chunk)
- done = total_size == total
- res = {
- "total": total_size,
- "uploaded": total,
- }
- yield f"data: {json.dumps(res)}\n\n"
- if done:
- f.seek(0)
- hashed = calculate_sha256(f)
- f.seek(0)
- url = f"{OLLAMA_API_BASE_URL}/blobs/sha256:{hashed}"
- response = requests.post(url, data=f)
- if response.ok:
- res = {
- "done": done,
- "blob": f"sha256:{hashed}",
- }
- os.remove(file_path)
- yield f"data: {json.dumps(res)}\n\n"
- else:
- raise "Ollama: Could not create blob, Please try again."
- except Exception as e:
- res = {"error": str(e)}
- yield f"data: {json.dumps(res)}\n\n"
- return StreamingResponse(file_write_stream(), media_type="text/event-stream")
|