| Crates.io | rusty-page-indexer |
| lib.rs | rusty-page-indexer |
| version | 0.5.5 |
| created_at | 2026-01-23 23:03:36.038949+00 |
| updated_at | 2026-01-24 08:14:08.658011+00 |
| description | A high-performance, reasoning-based RAG indexer in Rust following the PageIndex pattern. |
| homepage | https://github.com/Algiras/rusty-pageindex |
| repository | https://github.com/Algiras/rusty-pageindex |
| max_upload_size | |
| id | 2065622 |
| size | 997,048 |

RustyPageIndex is a high-performance Rust implementation of the PageIndex pattern. It transforms complex documents into hierarchical "Table-of-Contents" (TOC) trees for vectorless, reasoning-based RAG.
This project is inspired by VectifyAI/PageIndex but has diverged significantly with multi-repo support, parallel processing, and a unified tree architecture.
pdf-extract and pulldown-cmark for fast document processing| Feature | Original PageIndex | RustyPageIndex |
|---|---|---|
| Language | Python | Rust |
| Indexing | Per-file indices | Unified folder tree |
| Multi-repo | Not supported | Full support with list/clean |
| Parallelism | Sequential | Rayon parallel processing |
| Storage | Cloud-based (MCP) | Local filesystem |
| Tree Structure | Flat sections | Folder โ File โ Section hierarchy |
| Headerless Markdown | Empty tree | Auto-creates "Document" node |
One-liner Install (Unix/macOS):
curl -fsSL https://raw.githubusercontent.com/Algiras/rusty-pageindex/main/install.sh | bash
One-liner Install (Windows PowerShell):
irm https://raw.githubusercontent.com/Algiras/rusty-pageindex/main/install.ps1 | iex
Via Cargo:
cargo install rusty-page-indexer
npx skills add https://github.com/Algiras/rusty-pageindex --skill rusty-page-indexer
# For OpenAI
rusty-page-indexer auth --api-key "your-key-here"
# For Ollama (local LLM)
rusty-page-indexer auth --api-key "ollama" --api-base "http://localhost:11434/v1" --model "llama3.2"
# Index a repository
rusty-page-indexer index ./my-project
# Index with LLM-generated summaries
rusty-page-indexer index ./my-project --enrich
# Force re-index (ignores cache)
rusty-page-indexer index ./my-project --force
# Preview what would be indexed
rusty-page-indexer index ./my-project --dry-run
# Index multiple repos
rusty-page-indexer index ./repo-a
rusty-page-indexer index ./repo-b
# List all indexed repositories
rusty-page-indexer list
# Example output:
# ๐ Indexed Repositories
# โโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ
# ๐ repo-a (125.3 KB)
# /Users/you/projects/repo-a
# ๐ repo-b (89.7 KB)
# /Users/you/projects/repo-b
# โโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ
# Total: 2 indices
# Search across ALL indexed repositories
rusty-page-indexer query "how does authentication work"
# Search within a specific repository
rusty-page-indexer query "kafka messaging" --path repo-a
# Remove a specific index
rusty-page-indexer clean repo-a
# Remove all indices
rusty-page-indexer clean --all
rusty-page-indexer info
| Model | Cost | Speed | Notes |
|---|---|---|---|
gpt-4o |
$$$ | Fast | Best accuracy, recommended for complex queries |
gpt-4o-mini |
$ | Very Fast | Great balance of cost and quality โญ |
gpt-4.1-mini |
$ | Very Fast | Latest mini variant |
gpt-4-turbo |
$$ | Fast | Good for detailed reasoning |
gpt-3.5-turbo |
ยข | Very Fast | Budget option, decent accuracy |
# Configure for OpenAI
rusty-page-indexer auth --api-key "sk-..." --model "gpt-4o-mini"
# Override model per query
rusty-page-indexer query "question" --model gpt-4o
| Model | Size | Works | Notes |
|---|---|---|---|
gemma3:1b |
1B | โ | Minimum recommended for local use |
llama3.2:latest |
3B | โ | Good balance of speed and accuracy โญ |
qwen2.5:7b |
7B | โ | Reliable, slightly conservative |
llama3.1:latest |
8B | โ | Excellent accuracy |
mistral:7b |
7B | โ | Fast and capable |
phi3:mini |
3.8B | โ | Microsoft's compact model |
qwen2.5:0.5b |
0.5B | โ | Too small, unreliable responses |
tinyllama:1.1b |
1.1B | โ | Doesn't follow output format |
# Configure for Ollama
rusty-page-indexer auth --api-key "ollama" --api-base "http://localhost:11434/v1" --model "llama3.2"
# Make sure Ollama is running
ollama serve
Works with any OpenAI-compatible endpoint:
# Azure OpenAI
rusty-page-indexer auth --api-key "your-key" --api-base "https://your-resource.openai.azure.com/v1" --model "gpt-4"
# Together AI
rusty-page-indexer auth --api-key "your-key" --api-base "https://api.together.xyz/v1" --model "meta-llama/Llama-3-70b-chat-hf"
# Groq
rusty-page-indexer auth --api-key "your-key" --api-base "https://api.groq.com/openai/v1" --model "llama3-70b-8192"
Recommendation: Use gpt-4o-mini for remote or llama3.2 for local. Add --enrich during indexing for better search quality.
--path)The unified tree structure allows the LLM to see file names within folders, making navigation more accurate than flat file lists.
~/.rusty-page-indexer/
โโโ config.toml # API credentials and settings
โโโ manifest.json # Index registry with all repos
โโโ indices/
โโโ {hash-a}.json # Unified tree for repo-a
โโโ {hash-b}.json # Unified tree for repo-b
.md)#, ##, ###) into hierarchical tree.pdf)pdf-extractMIT