| Crates.io | infernum |
| lib.rs | infernum |
| version | 0.1.0 |
| created_at | 2025-12-03 07:38:38.900124+00 |
| updated_at | 2025-12-03 07:38:38.900124+00 |
| description | Infernum CLI - From the depths, intelligence rises |
| homepage | |
| repository | https://github.com/Daemoniorum-LLC/infernum-framework |
| max_upload_size | |
| id | 1963522 |
| size | 192,481 |
"From the depths, intelligence rises"
Blazingly fast local LLM inference CLI for running large language models on your machine.
# Install
cargo install infernum
# Set your model
infernum config set-model TinyLlama/TinyLlama-1.1B-Chat-v1.0
# Start chatting
infernum chat
localhost:8080# Chat interface
infernum chat [--model MODEL] [--system PROMPT]
# Start API server
infernum server [--port PORT]
# Download a model
infernum pull meta-llama/Llama-3.2-3B-Instruct
# List available models
infernum list
# System diagnostics
infernum doctor
# Configuration
infernum config set-model MODEL
infernum config get-model
# Install with CUDA support
cargo install infernum --features cuda
# Install with Metal support (macOS)
cargo install infernum --features metal
Infernum is built from specialized components:
Full documentation available at infernum.daemoniorum.com
Licensed under either of Apache License, Version 2.0 or MIT license at your option.