Crates.io | shepherd |
lib.rs | shepherd |
version | 0.1.5 |
source | src |
created_at | 2019-11-27 09:35:00.035849 |
updated_at | 2022-09-25 11:22:01.90284 |
description | A distributed video encoder that splits files into chunks to encode them on multiple machines in parallel. |
homepage | |
repository | https://github.com/martindisch/shepherd |
max_upload_size | |
id | 184779 |
size | 81,635 |
A distributed video encoder that splits files into chunks to encode them on multiple machines in parallel.
Using Cargo, you can do
$ cargo install shepherd
or just clone the repository and compile the binary with
$ git clone https://github.com/martindisch/shepherd
$ cd shepherd
$ cargo build --release
There's also a direct download for the latest x86-64 ELF binary.
The prerequisites are one or more (you'll want more) computers—which we'll
refer to as hosts—with ffmpeg
installed and configured such that you can
SSH into them directly. This means you'll have to ssh-copy-id
your public
key to them. I only tested it on Linux, but if you manage to set up
ffmpeg
and SSH, it might work on macOS or Windows directly or with little
modification.
The usage is pretty straightforward:
USAGE:
shepherd [FLAGS] [OPTIONS] <IN> <OUT> --clients <hostnames> [FFMPEG OPTIONS]...
FLAGS:
-h, --help Prints help information
-k, --keep Don't clean up temporary files
-V, --version Prints version information
OPTIONS:
-c, --clients <hostnames> Comma-separated list of encoding hosts
-l, --length <seconds> The length of video chunks in seconds
-t, --tmp <path> The path to the local temporary directory
ARGS:
<IN> The original video file
<OUT> The output video file
<FFMPEG OPTIONS>... Options/flags for ffmpeg encoding of chunks. The
chunks are video only, so don't pass in anything
concerning audio. Input/output file names are added
by the application, so there is no need for that
either. This is the last positional argument and
needs to be preceeded by double hypens (--) as in:
shepherd -c c1,c2 in.mp4 out.mp4 -- -c:v libx264
-crf 26 -preset veryslow -profile:v high -level 4.2
-pix_fmt yuv420p
This is also the default that is used if no options
are provided.
So if we have three machines c1, c2 and c3, we could do
$ shepherd -c c1,c2,c3 -l 30 source_file.mp4 output_file.mp4
to have it split the video in roughly 30 second chunks and encode them in
parallel. By default it encodes in H.264 with a CRF value of 26 and the
veryslow
preset. If you want to supply your own ffmpeg
options for more
control over the codec, you can do so by adding them to the end of the
invocation:
$ shepherd -c c1,c2 input.mkv output.mp4 -- -c:v libx264 -crf 40
ffmpeg
.Thanks to the work stealing method of distribution, having some hosts that are significantly slower than others does not delay the overall operation. In the worst case, the slowest machine is the last to start encoding a chunk and remains the only working encoder for the duration it takes to encode this one chunk. This window can easily be reduced by using smaller chunks.
As with all things parallel, Amdahl's law rears its ugly head and you don't just get twice the speed with twice the processing power. With this approach, you pay for having to split the video into chunks before you begin, transferring them to the encoders and the results back, and reassembling them. Although I should clarify that transferring the chunks to the encoders only causes a noticeable delay until every encoder has its first chunk, the subsequent ones can be sent while the encoders are working so they don't waste time waiting for that. And returning and assembling the encoded chunks doesn't carry too big of a penalty, since we're dealing with much more compressed data then.
To get a better understanding of the tradeoffs, I did some testing with a
couple of computers I had access to. They were my main, pretty capable
desktop, two older ones and a laptop. To figure out how capable each of
them is so we can compare the actual to the expected speedup, I let each of
them encode a relatively short clip of slightly less than 4 minutes taken
from the real video I want to encode, using the same settings I'd use for
the real job. And if you're wondering why encoding takes so long, it's
because I'm using the veryslow
preset for maximum efficiency, even though
it's definitely not worth the huge increase in encoding time. But it's a
nice simulation for how it would look if we were using an even more
demanding codec like AV1.
machine | duration (s) | power |
---|---|---|
desktop | 1373 | 1.000 |
old1 | 2571 | 0.53 |
old2 | 3292 | 0.42 |
laptop | 5572 | 0.25 |
total | - | 2.20 |
By giving my desktop the "power" level 1, we can determine how powerful the others are at this encoding task, based on how long it takes them in comparison. By adding the three other, less capable machines to the mix, we slightly more than double the theoretical encoding capability of our system.
I determined these power levels on a short clip, because encoding the full video would have taken very long on the less capable ones, especially the laptop. But I still needed to encode the full thing on at least one of them to make the comparison to the distributed encoding. I did that on my desktop since it's the fastest one, and to additionally verify that the power levels hold up for the full video, I bit the bullet and did the same on the second most powerful machine.
machine | duration (s) | power |
---|---|---|
desktop | 9356 | 1.00 |
old1 | 17690 | 0.53 |
Now we have the baseline we want to beat with parallel encoding, as well as confirmation that the power levels are valid for the full video. Let's see how much of the theoretical, but unreachable 2.2x speedup we can get.
Encoding the video in parallel took 5283 seconds, so 56.5% of the time using my fastest computer, or a 1.77x speedup. We committed about twice the computing power and we're not too far off that two times speedup. It's making use of the additionally available resources with an 80% efficiency in this case. I also tried to encode the short clip in parallel, which was very fast, but had a somewhat disappointing speedup of only 1.32x. I suspect that we get better results with longer videos, since encoding a chunk always takes longer than creating and transferring it (otherwise distributing wouldn't make sense at all). The longer the video then, the larger the ratio of encoding (which we can parallelize) in the total amount of time the process takes, and the more effective doing so becomes.
I've also looked at how the work is distributed over the nodes, depending on their processing power. At the end of a parallel encode, it's possible to determine how many chunks have been encoded by any given host.
host | chunks | power |
---|---|---|
desktop | 73 | 1.00 |
old1 | 39 | 0.53 |
old2 | 31 | 0.42 |
laptop | 19 | 0.26 |
Inferring the processing power from the number of chunks leads to almost exactly the same results as my initial determination, confirming it and proving that work is distributed efficiently.
To further see how the system scales, I've added two more machines, bringing the total processing power up to 3.29.
machine | duration (s) | power |
---|---|---|
desktop | 1373 | 1.00 |
c1 | 2129 | 0.64 |
old1 | 2571 | 0.53 |
c2 | 3022 | 0.45 |
old2 | 3292 | 0.42 |
laptop | 5572 | 0.25 |
total | - | 3.29 |
Encoding the video on these 6 machines in parallel took 3865 seconds, so 41.3% of the time using my fastest computer, or a 2.42x speedup. It's making use of the additionally available resources with a 74% efficiency here. As expected, while we can accelerate by adding more resources, we're looking at diminishing returns. Although the factor by which the efficiency decreases is not as bad as it could be.
While you can use your own ffmpeg
options to control how the video is
encoded, there is currently no such option for the audio, which is 192 kb/s
AAC by default.
Licensed under either of
at your option.
Unless you explicitly state otherwise, any contribution intentionally submitted for inclusion in the work by you, as defined in the Apache-2.0 license, shall be dual licensed as above, without any additional terms or conditions.