8 releases
0.3.4 | Oct 10, 2024 |
---|---|
0.3.3 | Aug 21, 2024 |
0.2.2 | Feb 28, 2024 |
0.1.0 | Dec 16, 2023 |
#907 in Machine learning
546 downloads per month
Used in 4 crates
(via kalosm-language)
395KB
9K
SLoC
RLlama
RLlama is a Rust implementation of the quantized Llama 7B language model.
Llama 7B is a very small but performant language model that can be easily run on your local machine.
This library uses Candle to run Llama.
Usage
use kalosm_llama::prelude::*;
#[tokio::main]
async fn main() {
let mut model = Llama::new().await.unwrap();
let prompt = "The capital of France is ";
let mut result = model.stream_text(prompt).await.unwrap();
print!("{prompt}");
while let Some(token) = result.next().await {
print!("{token}");
}
}
Dependencies
~34–55MB
~1M SLoC