2 releases

Uses new Rust 2024

new 0.2.3 Mar 10, 2025
0.2.2 Mar 10, 2025

#148 in Machine learning

MIT license

120KB
120 lines

llmterm offers suggestions based on your shell usage

llmterm screenshot

cuda

The kalosm crate is used to interface with the LLM. Cuda support is enabled by default. To choose other inference methods, edit the Cargo.toml file and rebuild.

to build

cargo build --release

to run

cargo run --release

to install

cargo install llmterm

models

cargo run --release -- --model llama_3_1_8b_chat
  • llama_3_1_8b_chat
  • mistral_7b_instruct_2
  • phi_3_5_mini_4k_instruct

to exit

Use Ctrl-C, or type exit or quit.

todo

easy on ramp

  • command line switch for different shells
  • command line switch to suggest only a command
  • check if llm response is empty, if so, pop the last activity off the buffer and try again
  • allow loading of all local models supported by kalosm by name

more challenging

Dependencies

~81–120MB
~2.5M SLoC