# TiniestSegmenter A port of [TinySegmenter](http://chasen.org/~taku/software/TinySegmenter/) written in pure, safe rust with no dependencies. You can find bindings for both [Rust](https://github.com/jwnz/tiniestsegmenter/tree/master/tiniestsegmenter) and [Python](https://github.com/jwnz/tiniestsegmenter/tree/master/bindings/python/). TinySegmenter is an n-gram word tokenizer for Japanese text originally built by [Taku Kudo](http://chasen.org/~taku/) (2008). Usage Add the crate to your project: `cargo add tiniestsegmenter`. ```Rust use tiniestsegmenter as ts; fn main() { let tokens: Vec<&str> = ts::tokenize("ジャガイモが好きです。"); } ```