![example workflow](https://github.com/crambl/binary_file_multiplier/actions/workflows/github-ci.yml/badge.svg) ![rust-badge](https://img.shields.io/github/license/crambl/binary_file_multiplier.svg) ![Crates.io](https://img.shields.io/crates/d/binmult) ![issues-badge](https://img.shields.io/github/issues/crambl/binary_file_multiplier.svg) ![pr-badge](https://img.shields.io/github/issues-pr/crambl/binary_file_multiplier.svg) ![stars-badge](https://img.shields.io/github/stars/crambl/binary_file_multiplier.svg) ![forks-bade](https://img.shields.io/github/forks/crambl/binary_file_multiplier.svg) ![GitHub commit activity (branch)](https://img.shields.io/github/commit-activity/m/crambl/binary_file_multiplier) # binmult - Binary File Multiplier Copy and append content from raw data files without mutating any contents ## Purpose >Stop me from banging my head against the wall that is my ineffecient shell scripts. Efficient and easy way of duplicating the contents of a small binary file, and writing that duplicated content to a new file. ## Why? Good for certain CI pipelines where regression testing for performance is required, but downloading large external files is impractical. This is certainly possible in a simple shell script, but making a 240 byte file into a 1 GB file is painfully slow in bash, and I couldn't figure out how to make it faster while keeping the contents intact. ## Quickstart See help ```shell $ binmult --help ``` Duplicate a file up to the closest (rounded down) multiple of 100 MiB ```shell # Any size input duplicated up to 100 MiB (rounded down) $ binmult input.raw --size 100 -o out.raw ``` ## Joke ![Jokes Card](https://readme-jokes.vercel.app/api)