Crates.io | scrubby |
lib.rs | scrubby |
version | 0.4.0 |
source | src |
created_at | 2023-02-17 23:10:38.702736 |
updated_at | 2024-09-30 05:53:21.976998 |
description | Remove or extract background taxa for metagenomic diagnostic applications |
homepage | https://github.com/esteinig/scrubby |
repository | https://github.com/esteinig/scrubby |
max_upload_size | |
id | 787834 |
size | 256,867 |
Host background depletion for metagenomic diagnostics with benchmarks and optimisation for clinical sequencing protocols and application scenarios.
...
Scrubby is available as statically compiled binary release for Linux and macOS (x86_64
and aarch64
).
git clone https://github.com/esteinig/scrubby && cd scrubby
Compile default version, which requires classifier or aligner (and samtools
) as dependencies:
cargo build --release
Compile built-in minimap2-rs
version using the mm2
feature (experimental). Note that minimap2-rs
is
only tested for x86_64
(Linux/macOS) and will not compile for aarch64
(Linux/macOS).
cargo build --release --features mm2
Compile with htslib
feature for using scrubby alignment
command with {sam, bam, cram}
format:
cargo build --release --features htslib
YOLO pre-compiled binaries and execute :skull:
curl -L https://github.com/esteinig/scrubby/release/scrubby-v1.0.0-mm2-linux-x86-64.tar.gz | tar -xvj ./scrubby
Reads should be quality- and adapter-trimmed before applying Scrubby
.
Single or paired-end reads are supported with optional gz
input/output compression.
Paired-end reads are always depleted/extracted as a pair (no unpaired read output).
Default minimap2
presets are sr
for paired-end reads and map-ont
for single reads.
Multiple values can be specified consecutively or using multiple arguments (-T Metazoa -T Bacteria
)
List pre-built index names:
scrubby download --list
Download pre-built index by name for aligners and classifiers:
scrubby download --name chm13v2 --aligner bowtie2 minimap2 --classifier kraken2
More options for aligners and classifier index download:
scrubby download --help
Read depletion pipeline with Bowtie2
aligner (default for paired-end reads):
scrubby reads -i r1.fq r2.fq -o c1.fq c2.fq -I chm13v2
Use built-in minimap2-rs
if compiled with mm2
feature (default for paired-end and long reads):
scrubby reads -i r1.fq r2.fq -o c1.fq c2.fq -I chm13v2.fa.gz
Long reads with non-default preset and minimap2
aligner (default for long reads):
scrubby reads -i r.fq -o c.fq -I chm13v2.fa.gz --preset lr-hq
Single-end short reads requires explicit aligner and preset for minimap2
:
scrubby reads -i r1.fq -o c1.fq -I chm13v2.fa.gz --aligner minimap2 --preset sr
Use classifier Kraken2
or Metabuli
instead of aligner:
scrubby reads -i r1.fq r2.fq -o c1.fq c2.fq -T Chordata -D 9606 -I chm13v2_k2/ -c kraken2
Use different aligner strobealign
or minimap2
:
scrubby reads -i r1.fq r2.fq -o c1.fq c2.fq -I chm13v2.fa.gz -a strobealign
With report output and depleted read identifiers:
scrubby reads -i r1.fq r2.fq -o c1.fq c2.fq -I chm13v2 -j report.json -r reads.tsv
Input and output compressed reads, increase threads and set working directory:
scrubby reads -i r1.fq.gz r2.fq.gz -o c1.fq.gz c2.fq.gz -I chm13v2 -w /tmp -t 16
Read extraction instead of depletion:
scrubby reads -i r1.fq.gz r2.fq.gz -o c1.fq.gz c2.fq.gz -I chm13v2 -e
Classifier output cleaning for Kraken-style reports and read classification outputs (Kraken2, Metabuli):
scrubby classifier \
--input r1.fq r2.fq \
--output c1.fq c2.fq\
--report kraken2.report \
--reads kraken2.reads \
--taxa Chordata \
--taxa-direct 9606
Alignment output cleaning (.sam|.bam|.cram|.paf|.gaf) or read identifier list (.txt). Alignment format is recognized from file extension or can be explicitly set with --format
. Alignment can be '-' for reading from stdin
with explicit format argument. PAF and TXT formats can be compressed (.gz|.xz|.bz) unless reading
from stdin
.
scrubby alignment \
--input r1.fq r2.fq \
--output c1.fq c2.fq\
--alignment alignment.paf \
--min-len 50 \
--min-cov 0.5 \
--min-mapq 50
minimap2 -x map-ont ref.fa r.fq | scrubby alignment -a - -f paf -i r.fq -o c.fq
Add the --extract
(-e
) flag to any of the above tasks to reverse read depletion for read extraction:
scrubby reads --extract ...
Difference between input and output reads with optional counts and read identifier summaries:
scrubby diff -i r1.fq r2.fq -o c1.fq c2.fq -j counts.json -r reads.tsv
{
"version": "0.7.0",
"date": "2024-07-30T06:50:15Z",
"command": "scrubby reads -i smoke_R1.fastq.gz -i smoke_R2.fastq.gz -o test_R1.fq.gz -o test_R2.fq.gz --index /data/opt/scrubby_indices/chm13v2 --threads 16 --workdir /tmp/test --json test.json",
"input": [
"smoke_R1.fastq.gz",
"smoke_R2.fastq.gz"
],
"output": [
"test_R1.fq.gz",
"test_R2.fq.gz"
],
"reads_in": 6678,
"reads_out": 3346,
"reads_removed": 3332,
"reads_extracted": 0,
"settings": {
"aligner": "bowtie2",
"classifier": null,
"index": "/data/opt/scrubby_indices/chm13v2",
"alignment": null,
"reads": null,
"report": null,
"taxa": [],
"taxa_direct": [],
"classifier_args": null,
"aligner_args": null,
"min_len": 0,
"min_cov": 0.0,
"min_mapq": 0,
"extract": false,
"preset": null
}
}
In this example, the settings.aligner
is null
if a --classifier
is set.
scrubby 0.7.0
Eike Steinig (@esteinig)
Taxonomic read depletion for clinical metagenomic diagnostics
Usage: scrubby [OPTIONS] <COMMAND>
Commands:
reads Deplete or extract reads using aligners or classifiers
classifier Deplete or extract reads from classifier outputs (Kraken2/Metabuli)
alignment Deplete or extract reads from aligner output with additional filters (SAM/BAM/PAF/TXT)
download List available indices and download files for aligners and classfiers
diff Get read counts and identifiers of the difference between input and output read files
help Print this message or the help of the given subcommand(s)
Options:
-l, --log-file <LOG_FILE> Output logs to file instead of terminal
-h, --help Print help (see more with '--help')
-V, --version Print version
List available indices and download files for aligners and classfiers
Usage: scrubby download [OPTIONS] --name [<NAME>...]
Options:
-n, --name [<NAME>...] Index name to download [possible values: chm13v2]
-o, --outdir <OUTDIR> Output directory for index download [default: .]
-a, --aligner [<ALIGNER>...] Download index for one or more aligners [possible values: bowtie2, minimap2, strobealign]
-c, --classfier [<CLASSFIER>...] Download index for one or more classifiers [possible values: kraken2, metabuli]
-l, --list List available index names and exit
-t, --timeout <TIMEOUT> Download timeout in minutes - increase for large files and slow connections [default: 360]
-h, --help Print help (see more with '--help')
Deplete or extract reads using aligners or classifiers
Usage: scrubby reads [OPTIONS] --index <INDEX>
Options:
-i, --input [<INPUT>...] Input read files (optional .gz)
-o, --output [<OUTPUT>...] Output read files (optional .gz)
-I, --index <INDEX> Reference index for aligner or classifier
-e, --extract Read extraction instead of depletion
-a, --aligner <ALIGNER> Aligner to use, default is 'bowtie2' or 'minimap2' [possible values: bowtie2, minimap2, strobealign, minimap2-rs]
-c, --classifier <CLASSIFIER> Classifier to use [possible values: kraken2, metabuli]
-T, --taxa [<TAXA>...] Taxa and all sub-taxa to deplete using classifiers
-D, --taxa-direct [<TAXA_DIRECT>...] Taxa to deplete directly using classifiers
-t, --threads <THREADS> Number of threads to use for aligner and classifier [default: 4]
-j, --json <JSON> Summary output file (.json)
-w, --workdir <WORKDIR> Optional working directory
-r, --read-ids <READ_IDS> Read identifier file (.tsv)
-h, --help Print help (see more with '--help')
Deplete or extract reads from classifier outputs (Kraken2, Metabuli)
Usage: scrubby classifier [OPTIONS] --report <REPORT> --reads <READS> --classifier <CLASSIFIER>
Options:
-i, --input [<INPUT>...] Input read files (optional .gz)
-o, --output [<OUTPUT>...] Output read files (optional .gz)
-e, --extract Read extraction instead of depletion
-k, --report <REPORT> Kraken-style report output from classifier
-j, --reads <READS> Kraken-style read classification output
-c, --classifier <CLASSIFIER> Classifier output style [possible values: kraken2, metabuli]
-T, --taxa [<TAXA>...] Taxa and all sub-taxa to deplete using classifiers
-D, --taxa-direct [<TAXA_DIRECT>...] Taxa to deplete directly using classifiers
-j, --json <JSON> Summary output file (.json)
-w, --workdir <WORKDIR> Optional working directory
-r, --read-ids <READ_IDS> Read identifier file (.tsv)
-h, --help Print help (see more with '--help')
Deplete or extract reads from aligner output with additional filters (SAM/BAM/PAF)
Usage: scrubby alignment [OPTIONS] --alignment <ALIGNMENT>
Options:
-i, --input [<INPUT>...] Input read files (can be compressed with .gz)
-o, --output [<OUTPUT>...] Output read files (can be compressed with .gz)
-e, --extract Read extraction instead of depletion
-a, --alignment <ALIGNMENT> Alignment file in SAM/BAM/PAF/TXT format
-f, --format <FORMAT> Explicit alignment format [possible values: sam, bam, cram, paf, txt]
-l, --min-len <MIN_LEN> Minimum query alignment length filter [default: 0]
-c, --min-cov <MIN_COV> Minimum query alignment coverage filter [default: 0]
-q, --min-mapq <MIN_MAPQ> Minimum mapping quality filter [default: 0]
-j, --json <JSON> Summary output file (.json)
-w, --workdir <WORKDIR> Optional working directory
-r, --read-ids <READ_IDS> Read identifier file (.tsv)
-h, --help Print help (see more with '--help')
Get read counts and identifiers of the difference between input and output read files
Usage: scrubby diff [OPTIONS]
Options:
-i, --input [<INPUT>...] Input read files (.gz | .xz | .bz)
-o, --output [<OUTPUT>...] Output read files (.gz | .xz | .bz)
-j, --json <JSON> Summary output file (.json)
-r, --read-ids <READ_IDS> Read identifier file (.tsv)
-h, --help Print help (see more with '--help')
You can use Scrubby with the builder structs from the prelude:
use scrubby::prelude::*;
// Example running Minimap2 on long reads
let scrubby_mm2_ont = Scrubby::builder(
"/path/to/reads_in.fastq",
"/path/to/reads_out.fastq"
)
.json("/path/to/report.json")
.extract(false)
.threads(16)
.index("/path/to/reference.fasta")
.aligner(Aligner::Minimap2)
.preset(Preset::MapOnt)
.build();
scrubby_mm2_ont.clean();
// Example running Minimap2 on paired-end reads
let scrubby_mm2_sr = Scrubby::builder(
vec!["/path/to/reads_in_R1.fastq", "/path/to/reads_in_R2.fastq"]
vec!["/path/to/reads_out_R1.fastq", "/path/to/reads_out_R2.fastq"]
)
.json("/path/to/report.json")
.extract(false)
.threads(16)
.index("/path/to/reference.fasta")
.aligner(Aligner::Minimap2)
.preset(Preset::Sr)
.build();
scrubby_mm2_sr.clean();
// Example running Kraken2, depleting Metazoa
let scrubby_kraken2_metazoa = Scrubby::builder(
"/path/to/reads_in.fastq",
"/path/to/reads_out.fastq"
)
.json("/path/to/report.json")
.extract(false)
.threads(16)
.index("/path/to/kraken/index")
.classifier(Classifier::Kraken2)
.taxa(vec!["Metazoa"])
.build();
scrubby_kraken2_metazoa.clean();
// Example from Kraken2 outputs, depleting Metazoa
let scrubby_kraken2_output_metazoa = Scrubby::builder(
"/path/to/reads_in.fastq",
"/path/to/reads_out.fastq"
)
.json("/path/to/report.json")
.extract(false)
.classifier(Classifier::Kraken2)
.report("/path/to/kraken/report")
.reads("/path/to/kraken/read/classifications")
.taxa(vec!["Metazoa"])
.build();
scrubby_kraken2_output_metazoa.clean();
// Example from alignment output file with filters
let scrubby_paf_output_filters = Scrubby::builder(
"/path/to/reads_in.fastq",
"/path/to/reads_out.fastq"
)
.json("/path/to/report.json")
.extract(false)
.alignment("/path/to/aln.paf")
.min_query_length(50)
.min_query_coverage(0.5)
.min_mapq(50)
.build();
scrubby_paf_output_filters.clean();
// Downloader example
let scrubby_dl = ScrubbyDownloader::builder(
"/path/to/download/directory",
vec![ScrubbyIndex::Chm13v2]
)
.timeout(180)
.aligner(vec![
Aligner::Minimap2, Aligner::Bowtie2
])
.classifier(vec![
Classifier::Kraken2, Classifier::Metabuli
])
.build();
scrubby_dl.list();
scrubby_dl.download_index();
Rust libraries:
Aligners and classifiers: