7 releases (stable)

1.0.5 Dec 1, 2023
0.1.0 Nov 30, 2023

#81 in Machine learning

MIT/Apache

9KB
109 lines

Ask-Ollama

Ask-Ollama is a command-line tool that allows users to interact with the Ollama LLM models directly from the terminal. This tool provides a simple and intuitive way to ask questions and receive responses from Ollama models.

Features

  • Interactive CLI: Easily ask questions and get responses.
  • Model Selection: Choose different models for varied responses.

Installation

To install Ask-Ollama, you need to have Rust and Cargo installed on your system. If you haven't already installed Rust, you can do so by following the instructions here.

Once Rust is installed, you can install Ask-Ollama using Cargo:

cargo install ask-ollama

Usage

After installation, you can start using Ask-Ollama by running:

ask [OPTIONS] [PROMPT]

Options

  • --model=[MODEL]: Specify the model to use (default is 'mistral').
  • --version: Display the version of the installed Ask-Ollama tool.
  • [PROMPT]: The question or prompt to send to Ollama. Quotation marks are not required.

Examples

Asking a question using the default model:

ask "What is the capital of France?"

or

ask What is the capital of France?

Specifying a different model:

ask --model=gale "Explain the theory of relativity"

Find all available models from Ollama here.

Checking the version:

ask --version

Seeing the help info:

ask --help

Contributing

Contributions to Ask-Ollama are welcome! If you have suggestions for improvements or encounter any issues, please feel free to open an issue or submit a pull request on our GitHub repository.

License

Ask-Ollama is licensed under the MIT License.

Dependencies

~1.7–6.5MB
~50K SLoC