seq_io

Crates.ioseq_io
lib.rsseq_io
version0.3.2
sourcesrc
created_at2017-07-31 11:59:44.300844
updated_at2023-08-07 14:24:30.331573
descriptionFast FASTA and FASTQ readers
homepage
repositoryhttps://github.com/markschl/seq_io
max_upload_size
id25803
size136,470
crate-owners (github:itdesigners:crate-owners)

documentation

https://docs.rs/seq_io

README

FASTA and FASTQ parsing and writing in Rust.

docs.rs crates.io Build status

This library provides an(other) attempt at parsing of the sequence formats FASTA and FASTQ, as well as writing.

Features:

  • Fast readers that minimize the use of allocations and copying of memory
  • Flexible methods for writing FASTA and FASTQ
  • Informative errors
  • Support for seeking
  • Serde support (for owned data structures)
  • Functions for parallel processing
  • Tested using fuzzing techniques see here

The FASTA parser can read and write multi-line files and allows iterating over the sequence lines without doing any allocation or copying. The FASTQ parser does not support multiple sequence / quality lines.

Simple example: Reads FASTA sequences from STDIN and writes them to STDOUT if long enough. Otherwise it prints a message. This should be very fast because the sequence is not allocated (seq_lines()).

use seq_io::fasta::{Reader,Record};
use std::io;

let mut reader = Reader::new(io::stdin());
let mut stdout = io::stdout();

while let Some(result) = reader.next() {
    let record = result.unwrap();
    // determine sequence length
    let seqlen = record.seq_lines()
                       .fold(0, |l, seq| l + seq.len());
    if seqlen > 100 {
        record.write_wrap(&mut stdout, 80).unwrap();
    } else {
        eprintln!("{} is only {} long", record.id().unwrap(), seqlen);
    }
}

Records are directly borrowing data from the internal buffered reader, therefore the while let is required. By default, the buffer will automatically grow if a record is too large to fit in. How it grows can be configured, it is also possible to set a size limit. Iterators over owned records are also provided.

Note: Make sure to add lto = true to the release profile in Cargo.toml because calls to functions of the underlying buffered reader (buf_redux) are not inlined otherwise.

View documentation

Multi-threaded processing

The parallel module contains functions for sending FASTQ/FASTA records to a thread pool where expensive calculations are done. Sequences are processed in batches (RecordSet) because sending across channels has a performance impact. FASTA/FASTQ records can be accessed in both the 'worker' function and (after processing) a function running in the main thread.

Similar projects in Rust

  • Rust-Bio: Binformatics library that provides simple FASTA and FASTQ readers.
  • fastq-rs: FASTQ parser with comparable performance (see below). seq_io was inspired by fastq_rs.
  • Needletail has a FASTA parser.
  • fasten (FASTQ parser)

Performance comparisons

The FASTQ reader from this crate performs similar to the fastq-rs reader. The rust-bio readers are slower due to allocations, copying, and UTF-8 validity checks.

All comparisons were run on a set of 100,000 auto-generated, synthetic sequences with lengths normally distributed around 500 bp and loaded into memory. The parsers from this crate (seq_io) are compared with fastq-rs (fastq_rs) and Rust-Bio (bio). The bars represent the throughput in GB/s (+/- standard error of the mean). Run on a Thinkpad X1 Carbon (i7-5500U) with a fixed frequency of 2.3 GHz using Rust 1.31 nightly

benchmark results

Explanation of labels:

  • Top bars: Iteration over all records without further action.
  • owned: An owned copy of each record is created for comparison with Rust-Bio, which does not provide zero copy parsing.
  • multiline: The FASTA sequence is split into 5 x 100 bp lines.
  • recordset: Records are parsed into record sets using read_record_set() (involves some copying).
  • parallel: Record sets are are sent to worker threads for parallel processing where they are being iterated over and then sent back to the main thread where there is another iteration over the records (the latter only in seq_io)
Commit count: 123

cargo fmt