| Crates.io | catgrad-llm |
| lib.rs | catgrad-llm |
| version | 0.1.1 |
| created_at | 2025-07-10 13:04:39.817165+00 |
| updated_at | 2025-07-10 13:04:39.817165+00 |
| description | Tools for LLMs built with catgrad |
| homepage | |
| repository | https://github.com/hellas-ai/catgrad |
| max_upload_size | |
| id | 1746387 |
| size | 145,518 |
LLMs in Catgrad.
The intent is to have three main self-contained modules:
models allow building LLM graphs: the core tensor network defining the modelrun (WIP) is for running LLM graphs given a 'package' of additional information like weights and tokenizers.
serve are abstract interfaces for serving LLMs as token iteratorsThis is not reflected by the current state of the code. What's missing:
runFuture changes: split run into generic "tensor runtime" and "llm code" where:
- tensor runtime: (run graph with arrays only; state/cache aware, tokenization-unaware, runnable on remote host)
- LLM code: tokenization, chat templates, etc.