Browse Source

update linux.md

Michael Yang 1 year ago
parent
commit
92119de9d8
2 changed files with 53 additions and 40 deletions
  1. 53 39
      docs/linux.md
  2. 0 1
      scripts/install.sh

+ 53 - 39
docs/linux.md

@@ -1,45 +1,25 @@
-# Installing Ollama on Linux
+# Ollama on Linux
 
-> Note: A one line installer for Ollama is available by running:
->
-> ```bash
-> curl https://ollama.ai/install.sh | sh
-> ```
-
-## Download the `ollama` binary
-
-Ollama is distributed as a self-contained binary. Download it to a directory in your PATH:
-
-```bash
-sudo curl -L https://ollama.ai/download/ollama-linux-amd64 -o /usr/bin/ollama
-sudo chmod +x /usr/bin/ollama
-```
-
-## Start Ollama
-
-Start Ollama by running `ollama serve`:
-
-```bash
-ollama serve
-```
-
-Once Ollama is running, run a model in another terminal session:
+## Install
 
+Install Ollama running this one-liner:
+>
 ```bash
-ollama run llama2
+curl https://ollama.ai/install.sh | sh
 ```
 
-## Install CUDA drivers (optional – for Nvidia GPUs)
+## Manual install
 
-[Download and install](https://developer.nvidia.com/cuda-downloads) CUDA.
+### Download the `ollama` binary
 
-Verify that the drivers are installed by running the following command, which should print details about your GPU:
+Ollama is distributed as a self-contained binary. Download it to a directory in your PATH:
 
 ```bash
-nvidia-smi
+sudo curl -L https://ollama.ai/download/ollama-linux-amd64 -o /usr/bin/ollama
+sudo chmod +x /usr/bin/ollama
 ```
 
-## Adding Ollama as a startup service (optional)
+### Adding Ollama as a startup service (recommended)
 
 Create a user for Ollama:
 
@@ -60,7 +40,6 @@ User=ollama
 Group=ollama
 Restart=always
 RestartSec=3
-Environment="HOME=/usr/share/ollama"
 
 [Install]
 WantedBy=default.target
@@ -73,7 +52,40 @@ sudo systemctl daemon-reload
 sudo systemctl enable ollama
 ```
 
-### Viewing logs
+### Install CUDA drivers (optional – for Nvidia GPUs)
+
+[Download and install](https://developer.nvidia.com/cuda-downloads) CUDA.
+
+Verify that the drivers are installed by running the following command, which should print details about your GPU:
+
+```bash
+nvidia-smi
+```
+
+### Start Ollama
+
+Start Ollama using `systemd`:
+
+```bash
+sudo systemctl start ollama
+```
+
+## Update
+
+Update ollama by running the install script again:
+
+```bash
+curl https://ollama.ai/install.sh | sh
+```
+
+Or by downloading the ollama binary:
+
+```bash
+sudo curl -L https://ollama.ai/download/ollama-linux-amd64 -o /usr/bin/ollama
+sudo chmod +x /usr/bin/ollama
+```
+
+## Viewing logs
 
 To view logs of Ollama running as a startup service, run:
 
@@ -84,19 +96,21 @@ journalctl -u ollama
 ## Uninstall
 
 Remove the ollama service:
+
 ```bash
-systemctl stop ollama
-systemctl disable ollama
-rm /etc/systemd/system/ollama.service
+sudo systemctl stop ollama
+sudo systemctl disable ollama
+sudo rm /etc/systemd/system/ollama.service
 ```
 
 Remove the ollama binary from your bin directory (either `/usr/local/bin`, `/usr/bin`, or `/bin`):
+
 ```bash
-rm /usr/local/bin/ollama
+sudo rm $(which ollama)
 ```
 
 Remove the downloaded models and Ollama service user:
 ```bash
-rm /usr/share/ollama
-userdel ollama
+sudo rm -r /usr/share/ollama
+sudo userdel ollama
 ```

+ 0 - 1
scripts/install.sh

@@ -89,7 +89,6 @@ User=ollama
 Group=ollama
 Restart=always
 RestartSec=3
-Environment="HOME=/usr/share/ollama"
 Environment="PATH=$PATH"
 
 [Install]