pipeless-ai

Crates.iopipeless-ai
lib.rspipeless-ai
version1.11.0
sourcesrc
created_at2023-11-11 15:07:47.347346
updated_at2024-02-23 12:27:55.917562
descriptionAn open-source computer vision framework to build and deploy applications in minutes
homepagehttps://www.pipeless.ai/
repositoryhttps://github.com/pipeless-ai/pipeless/
max_upload_size
id1032127
size307,147
Miguel Ángel Cabrera Miñagorri (miguelaeh)

documentation

https://www.pipeless.ai/docs

README

Pipeless

Easily create, deploy and run computer vision applications.



Loading video...



Pipeless is an open-source framework that takes care of everything you need to develop and deploy computer vision applications in just minutes. That includes code parallelization, multimedia pipelines, memory management, model inference, multi-stream management, and more. Pipeless allows you to ship applications that work in real-time in minutes instead of weeks/months.

Pipeless is inspired by modern serverless technologies. You provide some functions and Pipeless takes care of executing them for new video frames and everything involved.

With Pipeless you create self-contained boxes that we call "stages". Each stage is a micro pipeline that performs a specific task. Then, you can combine stages dynamically per stream, allowing you to process each stream with a different pipeline without changing your code and without restarting the program. To create a stage you simply provide a pre-process function, a model and a post-process function.

You can load industry-standard models, such as YOLO, or custom models in one of the supported inference runtimes just by providing a URL. Pipeless ships some of the most popular inference runtimes, such as the ONNX Runtime, allowing you to run inference with high performance on CPU or GPU out-of-the-box.

You can deploy your Pipeless and your applications to edge and IoT devices or to the cloud. There are several tools for the deployment, including container images.

The following is a non-exhaustive set of relevant features that Pipeless includes:

  • Multi-stream support: process several streams at the same time.
  • Dynamic stream configuration: add, edit, and remove streams on the fly via a CLI or REST API (more adapters to come).
  • Multi-language support: you can Write your hooks in several languages, including Python.
  • Dynamic processing steps: you can add any number of steps to your stream processing, and even modify those steps dynamically on a per-stream basis.
  • Built-in restart policies: Forget about dealing with connection errors, cameras that fail, etc. You can easily specify restart policies per stream that handle those situations automatially.
  • Highly parallelized: do not worry about multi-threading and/or multi-processing, Pipeless takes care of that for you.
  • Several inference runtimes supported: Provide a model and select one of the supported inference runtimes to run it out-of-the-box in CPU or GPUs. We support CUDA, TensorRT, OpenVINO, CoreML, and more to come.
  • Well-defined project structure and highly reusable code: Pipeless uses the file system structure to load processing stages and hooks, helping you organize the code in highly reusable boxes. Each stage is a directory, each hook is defined on its own file.

Get started now!

Join our community and contribute to making the lives of computer vision developers easier!

Requirements ☝️

  • Python. Pre-built binaries are linked to Python 3.10 in Linux amd64, 3.8 in Linux arm64, and 3.12 in macOS. If you have a different Python version, provide the --build flag to the install script to build from source so Pipeless links to your installed Python version (or update your version and use a pre-built binary, which is simpler).
  • Gstreamer 1.20.3. Verify with gst-launch-1.0 --gst-version. Installation instructions here

Installation 🛠️

curl https://raw.githubusercontent.com/pipeless-ai/pipeless/main/install.sh | bash

Find more information and installation options here.

Using docker

Instead of installing locally, you can alternatively use docker and save the time of installing dependencies:

docker run miguelaeh/pipeless --help

To use it with CUDA:

docker run miguelaeh/pipeless:latest-cuda --help

To use with TensorRT use:

docker run miguelaeh/pipeless:latest-tensorrt --help

Find the whole container documentation here.

Getting Started 🚀

Init a project:

pipeless init my_project --template scaffold
cd my_project

Start Pipeless:

pipeless start --stages-dir .

Provide a stream:

pipeless add stream --input-uri "https://pipeless-public.s3.eu-west-3.amazonaws.com/cats.mp4" --output-uri "screen" --frame-path "my-stage"

The code generated is an empty template that scafold a stage so it will do nothing. Please go to the examples to complete that stage.

You can also use the interactive shell to create the project:

Loading video...

Check the complete getting started guide or plunge into the complete documentation.

Examples 🌟

You can find some examples under the examples directory. Just copy those folders inside your project and play with them.

Find here the whole list of examples and step by step guides.

Benchmark 📈

We deployed Pipeless to several different devices so you can have a general idea of its performance. Find the results at the benchmark section of the docs.

Notable Changes

Notable changes indicate important changes between versions. Please check the whole list of notable changes.

Contributing 🤝

Thanks for your interest in contributing! Contributions are welcome and encouraged. While we're working on creating detailed contributing guidelines, here are a few general steps to get started:

  1. Fork this repository.
  2. Create a new branch: git checkout -b feature-branch.
  3. Make your changes and commit them: git commit -m 'Add new feature'.
  4. Push your changes to your fork: git push origin feature-branch.
  5. Open a GitHub pull request describing your changes.

We appreciate your help in making this project better!

Please note that for major changes or new features, it's a good idea to discuss them in an issue first so we can coordinate efforts.

License 📄

This project is licensed under the Apache License 2.0.

Apache License 2.0 Summary

The Apache License 2.0 is a permissive open-source license that allows you to use, modify, and distribute this software for personal or commercial purposes. It comes with certain obligations, including providing attribution to the original authors and including the original license text in your distributions.

For the full license text, please refer to the Apache License 2.0.

Commit count: 565

cargo fmt