Crates.io | tensorrt-rs |
lib.rs | tensorrt-rs |
version | 0.3.0 |
source | src |
created_at | 2020-09-10 06:03:34.633585 |
updated_at | 2020-09-26 04:37:08.685401 |
description | Rust library for using Nvidia's TensorRT deep learning acceleration library |
homepage | |
repository | https://github.com/mstallmo/tensorrt-rs |
max_upload_size | |
id | 286922 |
size | 50,408 |
:warning: This crate currently only supports Linux :warning:
Rust library for creating and executing TensorRT engines.
This library depends on tensorrt-sys for Rust bindings to the underlying C++ TensorRT library. See the tensorrt-sys README for information on prerequisite dependencies.
This crate is still very much in early stage devleopment. Support for TensorRT functionality is only the basic of what is needed to read a model file in .uff format, parse that file into a TensorRT engine and execute that engine, and seralize a binary version of that engine to disk. Currently TensorRT plugins are not supported so there may be issues when trying to use a model that requires plugin support to operate.
Upcoming features are not constrained to those listed above. Any feature requests are welcome and appreciated!
See the examples directory for a basic example on loading a UFF model, creating an engine, and performing inference on an image using the engine. The sample uses the UFF MNIST model provided in the samples directory when installing TensorRT on Linux.
Contributions are always welcome! Feel free to reach out with any questions or improvements and I will try to get back to you as soon as possible!