Browse Source

Merge branch 'main' into dev

Timothy Jaeryang Baek 1 năm trước cách đây
mục cha
commit
f5fdc203d7
1 tập tin đã thay đổi với 50 bổ sung11 xóa
  1. 50 11
      README.md

+ 50 - 11
README.md

@@ -115,43 +115,53 @@ docker build -t ollama-webui .
 docker run -d -p 3000:8080 -e OLLAMA_API_BASE_URL=https://example.com/api --name ollama-webui --restart always ollama-webui
 ```
 
-## How to Build for Static Deployment
+## How to Install Without Docker
 
-1. Clone & Enter the project
+While we strongly recommend using our convenient Docker container installation for optimal support, we understand that some situations may require a non-Docker setup, especially for development purposes. Please note that non-Docker installations are not officially supported, and you might need to troubleshoot on your own.
+
+### Project Components
+
+The Ollama Web UI consists of two primary components: the frontend and the backend (which serves as a reverse proxy, handling static frontend files, and additional features). Both need to be running concurrently for the development environment using `npm run dev`. Alternatively, you can set the `PUBLIC_API_BASE_URL` during the build process to have the frontend connect directly to your Ollama instance or build the frontend as static files and serve them with the backend.
+
+### Prerequisites
+
+1. **Clone and Enter the Project:**
 
    ```sh
    git clone https://github.com/ollama-webui/ollama-webui.git
-   pushd ./ollama-webui/
+   cd ollama-webui/
    ```
 
-2. Create and edit `.env`
+2. **Create and Edit `.env`:**
 
    ```sh
    cp -RPp example.env .env
    ```
 
-3. Install node dependencies
+### Building Ollama Web UI Frontend
+
+1. **Install Node Dependencies:**
 
    ```sh
-   npm i
+   npm install
    ```
 
-4. Run in dev mode, or build the site for deployment
+2. **Run in Dev Mode or Build for Deployment:**
 
-   - Test in Dev mode:
+   - Dev Mode (requires the backend to be running simultaneously):
 
      ```sh
      npm run dev
      ```
 
-   - Build for Deploy:
+   - Build for Deployment:
 
      ```sh
-     #`PUBLIC_API_BASE_URL` will overwrite the value in `.env`
+     # `PUBLIC_API_BASE_URL` overwrites the value in `.env`
      PUBLIC_API_BASE_URL='https://example.com/api' npm run build
      ```
 
-5. Test the build with `caddy` (or the server of your choice)
+3. **Test the Build with `Caddy` (or your preferred server):**
 
    ```sh
    curl https://webi.sh/caddy | sh
@@ -160,6 +170,35 @@ docker run -d -p 3000:8080 -e OLLAMA_API_BASE_URL=https://example.com/api --name
    caddy run --envfile .env --config ./Caddyfile.localhost
    ```
 
+### Running Ollama Web UI Backend
+
+If you wish to run the backend for deployment, ensure that the frontend is built so that the backend can serve the frontend files along with the API route.
+
+#### Setup Instructions
+
+1. **Install Python Requirements:**
+
+   ```sh
+   cd ./backend
+   pip install -r requirements.txt
+   ```
+
+2. **Run Python Backend:**
+
+   - Dev Mode with Hot Reloading:
+
+     ```sh
+     sh dev.sh
+     ```
+
+   - Deployment:
+
+     ```sh
+     sh start.sh
+     ```
+
+Now, you should have the Ollama Web UI up and running at [http://localhost:8080/](http://localhost:8080/). Feel free to explore the features and functionalities of Ollama! If you encounter any issues, please refer to the instructions above or reach out to the community for assistance.
+
 ## Troubleshooting
 
 See [TROUBLESHOOTING.md](/TROUBLESHOOTING.md) for information on how to troubleshoot and/or join our [Ollama Web UI Discord community](https://discord.gg/5rJgQTnV4s).