| Crates.io | llama_link |
| lib.rs | llama_link |
| version | 0.2.1 |
| created_at | 2024-12-13 07:12:24.037418+00 |
| updated_at | 2024-12-28 17:31:32.305042+00 |
| description | A llama.cpp server interface |
| homepage | |
| repository | https://github.com/mcmah309/llama_link |
| max_upload_size | |
| id | 1481951 |
| size | 88,801 |
read ./llama.cpp/docs/build.md to build
Run. e.g. ./build/bin/llama-server -m ./models/7B/ggml-model-f16.gguf --prompt "Once pick an action" --json-schema '{}'
Options:
Use the package https://search.nixos.org/options?channel=unstable&from=0&size=50&sort=relevance&type=packages&query=llama-cpp
Use the flake at ./llama.cpp/flake.nix. E.g.
{
description = "My CUDA-enabled llama.cpp development environment";
inputs = {
nixpkgs.url = "github:NixOS/nixpkgs/nixos-unstable";
flake-parts.url = "github:hercules-ci/flake-parts";
llama-cpp.url = "github:ggerganov/llama.cpp";
};
outputs = { self, nixpkgs, flake-parts, llama-cpp }@inputs:
flake-parts.lib.mkFlake { inherit inputs; } {
systems = [ "x86_64-linux" "aarch64-linux" ];
perSystem = { config, self', inputs', pkgs, system, ... }: {
devShells.default = pkgs.mkShell {
buildInputs = [
llama-cpp.packages.${system}.cuda
pkgs.cudatoolkit
pkgs.gcc
pkgs.cmake
];
shellHook = ''
export CUDA_PATH=${pkgs.cudatoolkit}
export LD_LIBRARY_PATH=${pkgs.cudatoolkit}/lib:$LD_LIBRARY_PATH
'';
};
};
};
}