Crates.io | llm-cli |
lib.rs | llm-cli |
version | 0.1.1 |
source | src |
created_at | 2023-05-07 15:17:35.810673 |
updated_at | 2023-05-08 18:45:35.943941 |
description | A CLI for running inference on supported Large Language Models. Powered by the `llm` library. |
homepage | |
repository | https://github.com/rustformers/llm |
max_upload_size | |
id | 859227 |
size | 74,323 |
Image by @darthdeus, using Stable Diffusion
llm
is a Rust ecosystem of libraries for running inference on large language
models, inspired by llama.cpp.
The primary crate is the llm
crate, which wraps llm-base
and supported model
crates.
On top of llm
, there is a CLI application, llm-cli
, which provides a
convenient interface for running inference on supported models. This CLI
is available from the latest GitHub release.
It is powered by the ggml
tensor library,
and aims to bring the robustness and ease of use of Rust to the world of large
language models.
Currently, the following models are supported:
Make sure you have a Rust 1.65.0 or above and C toolchain1 set up.
llm
is a Rust library that re-exports llm-base
and the model crates (e.g.
bloom
, gpt2
llama
).
llm-cli
(binary name llm
) is a basic application that provides a CLI
interface to the library.
NOTE: For best results, make sure to build and run in release mode. Debug builds are going to be very slow.
cargo
Run
cargo install --git https://github.com/rustformers/llm llm-cli
to install llm
to your Cargo bin
directory, which rustup
is likely to have
added to your PATH
.
The CLI application can then be run through llm
.
Clone the repository and then build it with
git clone --recurse-submodules git@github.com:rustformers/llm.git
cargo build --release
The resulting binary will be at target/release/llm[.exe]
.
It can also be run directly through Cargo, using
cargo run --release -- <ARGS>
This is useful for development.
GGML files are easy to acquire. For a list of models that have been tested, see the known-good models.
Certain older GGML formats are not supported by this project, but the goal is to maintain feature parity with the upstream GGML project. For problems relating to loading models, or requesting support for supported GGML model types, please open an Issue.
Hugging Face 🤗 is a leader in open-source machine learning and hosts hundreds of GGML models. Search for GGML models on Hugging Face 🤗.
This Reddit community maintains a wiki related to GGML models, including well organized lists of links for acquiring GGML models (mostly from Hugging Face 🤗).
Currently, the only legal source to get the original weights is this repository.
After acquiring the weights, it is necessary to convert them into a format that is compatible with ggml. To achieve this, follow the steps outlined below:
Warning
To run the Python scripts, a Python version of 3.9 or 3.10 is required. 3.11 is unsupported at the time of writing.
# Convert the model to f16 ggml format
python3 scripts/convert-pth-to-ggml.py /path/to/your/models/7B/ 1
# Quantize the model to 4-bit ggml format
cargo run --release llama quantize /path/to/your/models/7B/ggml-model-f16.bin /path/to/your/models/7B/ggml-model-q4_0.bin q4_0
In future, we hope to provide a more streamlined way of converting models.
Note
The llama.cpp repository has additional information on how to obtain and run specific models.
For example, try the following prompt:
llm llama infer -m <path>/ggml-model-q4_0.bin -p "Tell me how cool the Rust programming language is:"
Some additional things to try:
Use --help
to see a list of available options.
If you have the alpaca-lora weights,
try repl
mode!
llm llama repl -m <path>/ggml-alpaca-7b-q4.bin -f examples/alpaca_prompt.txt
Sessions can be loaded (--load-session
) or saved (--save-session
) to file.
To automatically load and save the same session, use --persist-session
. This
can be used to cache prompts to reduce load time, too:
# To build (This will take some time, go grab some coffee):
docker build -t llm .
# To run with prompt:
docker run --rm --name llm -it -v ${PWD}/data:/data -v ${PWD}/examples:/examples llm llama infer -m data/gpt4all-lora-quantized-ggml.bin -p "Tell me how cool the Rust programming language is:"
# To run with prompt file and repl (will wait for user input):
docker run --rm --name llm -it -v ${PWD}/data:/data -v ${PWD}/examples:/examples llm llama repl -m data/gpt4all-lora-quantized-ggml.bin -f examples/alpaca_prompt.txt
It was not my choice. Ferris appeared to me in my dreams and asked me to rewrite this in the name of the Holy crab.
Come on! I don't want to get into a flame war. You know how it goes, something something memory something something cargo is nice, don't make me say it, everybody knows this already.
Sheesh! Okaaay. After seeing the huge potential for llama.cpp, the first
thing I did was to see how hard would it be to turn it into a library to embed
in my projects. I started digging into the code, and realized the heavy lifting
is done by ggml
(a C library, easy to bind to Rust) and the whole project was
just around ~2k lines of C++ code (not so easy to bind). After a couple of
(failed) attempts to build an HTTP server into the tool, I realized I'd be much
more productive if I just ported the code to Rust, where I'm more comfortable.
Haha. Of course not. I just like collecting imaginary internet points, in the form of little stars, that people seem to give to me whenever I embark on pointless quests for rewriting X thing, but in Rust.
llama.cpp
?This is a reimplementation of llama.cpp
that does not share any code with it
outside of ggml
. This was done for a variety of reasons:
llama.cpp
requires a C++ compiler, which can cause problems for
cross-compilation to more esoteric platforms. An example of such a platform is
WebAssembly, which can require a non-standard compiler SDK.ggml
an optional backend (see
this issue).In general, we hope to build a solution for model inferencing that is as easy to use and deploy as any other Rust crate.
llm
?llm
.A modern-ish C toolchain is required to compile ggml
. A C++ toolchain
should not be necessary. ↩