#repl #language-model #nlp #gpt #chatgpt #command-line-tool #localai

app aichat

All-in-one chat and copilot CLI that integrates 10+ AI platforms

15 breaking releases

0.16.0 Apr 11, 2024
0.14.0 Mar 7, 2024
0.12.0 Dec 26, 2023
0.11.0 Nov 29, 2023
0.8.0 Mar 21, 2023

#1 in #language-model

Download history 69/week @ 2024-01-02 31/week @ 2024-01-09 20/week @ 2024-01-16 14/week @ 2024-01-23 12/week @ 2024-01-30 29/week @ 2024-02-06 27/week @ 2024-02-13 189/week @ 2024-02-20 120/week @ 2024-02-27 297/week @ 2024-03-05 143/week @ 2024-03-12 66/week @ 2024-03-19 44/week @ 2024-03-26 226/week @ 2024-04-02 360/week @ 2024-04-09 155/week @ 2024-04-16

792 downloads per month

MIT/Apache and LGPL-3.0-or-later

3MB
6.5K SLoC

AIChat

CI Crates Discord

All-in-one chat and copilot CLI that integrates 10+ AI platforms.

Command Mode:

command mode

Chat REPL mode:

chat-repl mode

Features

Integrated platforms

  • OpenAI: GPT3.5/GPT4 (paid, vision)
  • Azure-OpenAI (paid)
  • OpenAI-Compatible platforms
  • Gemini: Gemini-1.0/Gemini-1.5 (free, vision)
  • VertexAI (paid, vision)
  • Claude: Claude3 (vision, paid)
  • Mistral (paid)
  • Cohere (paid)
  • Ollama (free, local)
  • Ernie (paid)
  • Qianwen (paid, vision)
  • Moonshot (paid)

Install

Use a package management tool

For Rust programmer

cargo install aichat

For macOS Homebrew or a Linuxbrew user

brew install aichat

For Windows Scoop user

scoop install aichat

For Android Termux user

pkg install aichat

Binaries for macOS, Linux, and Windows

Download it from GitHub Releases, unzip, and add aichat to your $PATH.

Config

On first launch, aichat will guide you through the configuration.

> No config file, create a new one? Yes
> AI Platform: openai
> API Key: <your_api_key_here>

Feel free to adjust the configuration according to your needs.

model: openai:gpt-3.5-turbo      # LLM model
temperature: 1.0                 # LLM temperature
save: true                       # Whether to save the message
save_session: null               # Whether to save the session, if null, asking
highlight: true                  # Set false to turn highlight
light_theme: false               # Whether to use a light theme
wrap: no                         # Specify the text-wrapping mode (no, auto, <max-width>)
wrap_code: false                 # Whether wrap code block
ctrlc_exit: false                # Whether to exit REPL when Ctrl+C is pressed
auto_copy: false                 # Automatically copy the last output to the clipboard
keybindings: emacs               # REPL keybindings. values: emacs, vi
prelude: ''                      # Set a default role or session (role:<name>, session:<name>)
compress_threshold: 1000         # Compress session if tokens exceed this value (valid when >=1000)

clients:
  - type: openai
    api_key: sk-xxx

  - type: openai-compatible
    name: localai
    api_base: http://127.0.0.1:8080/v1
    models:
      - name: llama2
        max_input_tokens: 8192

Please review the config.example.yaml to see all available configuration options.

There are some configurations that can be set through environment variables, see Environment Variables.

Command

Usage: aichat [OPTIONS] [TEXT]...

Arguments:
  [TEXT]...  Input text

Options:
  -m, --model <MODEL>        Choose a LLM model
  -r, --role <ROLE>          Choose a role
  -s, --session [<SESSION>]  Create or reuse a session
  -e, --execute              Execute commands using natural language
  -c, --code                 Generate only code
  -f, --file <FILE>          Attach files to the message
  -H, --no-highlight         Disable syntax highlighting
  -S, --no-stream            No stream output
  -w, --wrap <WRAP>          Specify the text-wrapping mode (no, auto, <max-width>)
      --light-theme          Use light theme
      --dry-run              Run in dry run mode
      --info                 Print related information
      --list-models          List all available models
      --list-roles           List all available roles
      --list-sessions        List all available sessions
  -h, --help                 Print help
  -V, --version              Print version

Here are some practical examples:

aichat                                          # Start in REPL mode

aichat -e install nvim                          # Execute
aichat -c fibonacci in js                       # Code

aichat -s                                       # REPL + New session
aichat -s session1                              # REPL + New/Reuse 'session1'

aichat --info                                   # System info
aichat -r role1 --info                          # Role info
aichat -s session1 --info                       # Session info

cat data.toml | aichat -c to json > data.json   # Pipe stdio/stdout

aichat -f data.toml -c to json > data.json      # Attach files

aichat -f a.png -f b.png diff images            # Attach images

Shell commands

Simply input what you want to do in natural language, and aichat will prompt and run the command that achieves your intent.

aichat -e <text>...

Aichat is aware of OS and $SHELL you are using, it will provide shell command for specific system you have. For instance, if you ask aichat to update your system, it will return a command based on your OS. Here's an example using macOS:

aichat -e update my system
# sudo softwareupdate -i -a
# ? [e]xecute, [d]escribe, [a]bort:  (e)  

The same prompt, when used on Ubuntu, will generate a different suggestion:

 aichat -e update my system
# sudo apt update && sudo apt upgrade -y
# ? [e]xecute, [d]escribe, [a]bort:  (e)  

We can still use pipes to pass input to aichat and generate shell commands:

aichat -e POST localhost with < data.json
# curl -X POST -H "Content-Type: application/json" -d '{"a": 1, "b": 2}' localhost
# ? [e]xecute, [d]escribe, [a]bort:  (e)  

We can also pipe the output of aichat which will disable interactive mode.

aichat -e find all json files in current folder | pbcopy

Shell integration

This is a very handy feature, which allows you to use aichat shell completions directly in your terminal, without the need to type aichat with prompt and arguments. This feature puts aichat completions directly into terminal buffer (input line), allowing for immediate editing of suggested commands.

aichat-integration

To install shell integration, go to ./scripts/shell-integration to download the script and source the script in rc file. After that restart your shell. You can invoke the completion with alt+e hotkey.

Generating code

By using the --code or -c parameter, you can specifically request pure code output, for instance:

aichat --code a echo server in node.js
const net = require('net');

const server = net.createServer(socket => {
  socket.on('data', data => {
    socket.write(data);
  });

  socket.on('end', () => {
    console.log('Client disconnected');
  });
});

server.listen(3000, () => {
  console.log('Server running on port 3000');
});

Since it is valid js code, we can redirect the output to a file:

aichat --code a echo server in node.js > echo-server.js 
node echo-server.js

The -c/--code option ensures the extraction of code from Markdown.

Chat REPL

Aichat has a powerful Chat REPL.

The REPL supports:

  • Tab autocompletion
  • Custom REPL Prompt
  • Emacs/Vi keybinding
  • Edit/paste multi-line text
  • Open an editor to modify the current prompt
  • History
  • Undo support

.help - print help message

> .help
.help                    Print this help message
.info                    Print system info
.model                   Switch LLM model
.role                    Use a role
.info role               Show the role info
.exit role               Leave current role
.session                 Start a context-aware chat session
.info session            Show the session info
.save session            Save the session to the file
.clear messages          Clear messages in the session
.exit session            End the current session
.file                    Attach files to the message and then submit it
.set                     Modify the configuration parameters
.copy                    Copy the last reply to the clipboard
.exit                    Exit the REPL

Type ::: to begin multi-line editing, type ::: to end it.
Press Ctrl+O to open an editor to modify the current prompt.
Press Ctrl+C to abort readline, Ctrl+D to exit the REPL

.info - view information

> .info
model               openai:gpt-3.5-turbo
temperature         -
dry_run             false
save                true
save_session        true
highlight           true
light_theme         false
wrap                no
wrap_code           false
auto_copy           false
keybindings         emacs
prelude             -
compress_threshold  1000
config_file         /home/alice/.config/aichat/config.yaml
roles_file          /home/alice/.config/aichat/roles.yaml
messages_file       /home/alice/.config/aichat/messages.md
sessions_dir        /home/alice/.config/aichat/sessions

.model - choose a model

> .model openai:gpt-4
> .model ollama:llama2

You can easily enter model name using autocomplete.

.role - let the AI play a role

Select a role:

> .role emoji

Send message with the role:

emoji> hello
👋

Leave current role:

emoji> .exit role

> hello
Hello there! How can I assist you today?

Show role info:

emoji> .info role
name: emoji
prompt: I want you to translate the sentences I write into emojis. I will write the sentence, and you will express it with emojis. I just want you to express it with emojis. I don't want you to reply with anything but emoji. When I need to tell you something in English, I will do it by wrapping it in curly brackets like {like this}.
temperature: null

Temporarily use a role to send a message.

> ::: .role emoji
hello world
:::
👋🌍

> 

.session - context-aware conversation

By default, aichat behaves in a one-off request/response manner.

You should run aichat with -s/--session or use the .session command to start a session.

> .session

temp) 1 to 5, odd only                                                                    0
1, 3, 5

temp) to 7                                                                        19(0.46%)
1, 3, 5, 7

temp) .exit session                                                               42(1.03%)
? Save session? (y/N)  

The prompt on the right side is about the current usage of tokens and the proportion of tokens used, compared to the maximum number of tokens allowed by the model.

.file - attach files to the message

Usage: .file <file>... [-- text...]

.file message.txt
.file config.yaml -- convert to toml
.file a.jpg b.jpg -- What’s in these images?
.file https://ibb.co/a.png https://ibb.co/b.png -- what is the difference?

Only the current model that supports vision can process images submitted through .file command.

.set - modify the configuration temporarily

.set temperature 1.2
.set compress_threshold 1000
.set dry_run true
.set highlight false
.set save false
.set save_session true
.set auto_copy true

Roles

We can define a batch of roles in roles.yaml.

Retrieve the location of roles.yaml through the REPL .info command or CLI --info option.

For example, we can define a role:

- name: shell
  prompt: >
    I want you to act as a Linux shell expert.
    I want you to answer only with bash code.
    Do not provide explanations.

Let LLM answer questions in the role of a Linux shell expert.

> .role shell

shell>  extract encrypted zipfile app.zip to /tmp/app
mkdir /tmp/app
unzip -P PASSWORD app.zip -d /tmp/app

For more details about roles, please visit Role Guide.

License

Copyright (c) 2023-2024 aichat-developers.

Aichat is made available under the terms of either the MIT License or the Apache License 2.0, at your option.

See the LICENSE-APACHE and LICENSE-MIT files for license details.

Dependencies

~26–43MB
~757K SLoC