| Crates.io | token_trekker_rs |
| lib.rs | token_trekker_rs |
| version | 0.1.3 |
| created_at | 2023-03-22 00:35:31.696294+00 |
| updated_at | 2023-03-22 00:42:44.47864+00 |
| description | A fun and efficient Rust library to count tokens in text files using different tokenizers. |
| homepage | |
| repository | |
| max_upload_size | |
| id | 816617 |
| size | 53,911 |
token_trekker_rs is a command-line tool for counting the total number of tokens in all files within a directory or matching a glob pattern, using various tokenizers.
To install token_trekker_rs from crates.io, run:
cargo install token_trekker_rs
To build token_trekker_rs from the source code, first clone the repository:
git clone https://github.com/1rgs/token_trekker_rs.git
cd token_trekker_rs
Then build the project using cargo:
cargo build --release
The compiled binary will be available at ./target/release/token-trekker.
To count tokens in a directory or for files matching a glob pattern, run the following command:
token-trekker --path <path_or_glob_pattern> <tokenizer>
Replace <path_or_glob_pattern> with the path to the directory or the glob pattern of the files to process, and
For example:
token_trekker_rs --path "path/to/files/*.txt" p50k-base