123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185 |
- import os
- import sys
- from argparse import ArgumentParser, HelpFormatter, PARSER
- from yaspin import yaspin
- from ollama import model, engine
- from ollama.cmd import server
- class CustomHelpFormatter(HelpFormatter):
- """
- This class is used to customize the way the argparse help text is displayed.
- We specifically override the _format_action method to exclude the line that
- shows all the subparser command options in the help text. This line is typically
- in the form "{serve,models,pull,run}".
- """
- def _format_action(self, action):
- # get the original help text
- parts = super()._format_action(action)
- if action.nargs == PARSER:
- # remove the unwanted first line
- parts = "\n".join(parts.split("\n")[1:])
- return parts
- def main():
- parser = ArgumentParser(
- description='Ollama: Run any large language model on any machine.',
- formatter_class=CustomHelpFormatter,
- )
- # create models home if it doesn't exist
- os.makedirs(model.MODELS_CACHE_PATH, exist_ok=True)
- subparsers = parser.add_subparsers(
- title='commands',
- )
- list_parser = subparsers.add_parser(
- "models",
- description="List all available models stored locally.",
- help="List all available models stored locally.",
- )
- list_parser.set_defaults(fn=list_models)
- search_parser = subparsers.add_parser(
- "search",
- description="Search for compatible models that Ollama can run.",
- help="Search for compatible models that Ollama can run. Usage: search [model]",
- )
- search_parser.add_argument(
- "query",
- nargs="?",
- help="Optional name of the model to search for.",
- )
- search_parser.set_defaults(fn=search)
- pull_parser = subparsers.add_parser(
- "pull",
- description="Download a specified model from a remote source.",
- help="Download a specified model from a remote source. Usage: pull [model]",
- )
- pull_parser.add_argument("model", help="Name of the model to download.")
- pull_parser.set_defaults(fn=pull)
- run_parser = subparsers.add_parser(
- "run",
- description="Run a model and submit prompts.",
- help="Run a model and submit prompts. Usage: run [model] [prompt]",
- )
- run_parser.add_argument("model", help="Name of the model to run.")
- run_parser.add_argument(
- "prompt",
- nargs="?",
- help="Optional prompt for the model, interactive mode enabled when not specified.",
- )
- run_parser.set_defaults(fn=run)
- server.set_parser(
- subparsers.add_parser(
- "serve",
- description="Start a persistent server to interact with models via the API.",
- help="Start a persistent server to interact with models via the API.",
- )
- )
- args = parser.parse_args()
- args = vars(args)
- try:
- fn = args.pop("fn")
- fn(**args)
- except KeyboardInterrupt:
- pass
- except KeyError:
- parser.print_help()
- except Exception as e:
- print(e)
- def list_models(*args, **kwargs):
- for m in model.models(*args, **kwargs):
- print(m)
- def generate(*args, **kwargs):
- if prompt := kwargs.get("prompt"):
- print(">>>", prompt, flush=True)
- generate_oneshot(*args, **kwargs)
- return
- if sys.stdin.isatty():
- return generate_interactive(*args, **kwargs)
- return generate_batch(*args, **kwargs)
- def generate_oneshot(*args, **kwargs):
- print(flush=True)
- spinner = yaspin()
- spinner.start()
- spinner_running = True
- try:
- for output in engine.generate(model_name=kwargs.pop('model'), *args, **kwargs):
- choices = output.get("choices", [])
- if len(choices) > 0:
- if spinner_running:
- spinner.stop()
- spinner_running = False
- print("\r", end="") # move cursor back to beginning of line again
- print(choices[0].get("text", ""), end="", flush=True)
- except Exception:
- spinner.stop()
- raise
- # end with a new line
- print(flush=True)
- print(flush=True)
- def generate_interactive(*args, **kwargs):
- while True:
- print(">>> ", end="", flush=True)
- line = next(sys.stdin)
- if not line:
- return
- kwargs.update({"prompt": line})
- generate_oneshot(*args, **kwargs)
- def generate_batch(*args, **kwargs):
- for line in sys.stdin:
- print(">>> ", line, end="", flush=True)
- kwargs.update({"prompt": line})
- generate_oneshot(*args, **kwargs)
- def search(*args, **kwargs):
- try:
- model_names = model.search_directory(*args, **kwargs)
- if len(model_names) == 0:
- print("No models found.")
- return
- elif len(model_names) == 1:
- print(f"Found {len(model_names)} available model:")
- else:
- print(f"Found {len(model_names)} available models:")
- for model_name in model_names:
- print(model_name.lower())
- except Exception as e:
- print("Failed to fetch available models, check your network connection")
- def pull(*args, **kwargs):
- model.pull(model_name=kwargs.pop('model'), *args, **kwargs)
- def run(*args, **kwargs):
- name = model.pull(model_name=kwargs.pop('model'), *args, **kwargs)
- kwargs.update({"model": name})
- print(f"Running {name}...")
- generate(*args, **kwargs)
|