sift_stream

Crates.iosift_stream
lib.rssift_stream
version0.7.2
created_at2025-04-01 19:12:51.607955+00
updated_at2026-01-16 22:34:11.131912+00
descriptionA robust Sift telemetry streaming library
homepagehttps://github.com/sift-stack/sift
repositoryhttps://github.com/sift-stack/sift
max_upload_size
id1615311
size623,789
Benji Nguyen (solidiquis)

documentation

README

sift_stream

Crates.io docs.rs

Overview

SiftStream is a Rust-based telemetry streaming system that provides reliable, high-throughput data ingestion to the Sift platform. The architecture is built around a task-based system with multiple async tasks communicating through channels and control messages to ensure data reliability and fault tolerance.

Features highlights:

  • Builtin retries with default or custom retry policies in the case of a Sift outage or a client-side network outage.
  • Periodic checkpointing to confirm that all data within a particular period has been received by Sift.
  • Optional automated backups to mitigate data-loss in the case of misc. outages.
  • Optional tracing/logging to monitor the health of your stream and view various ingestion performance metrics.

See the examples directory for demonstrations on how to stream data to Sift using sift_stream.

Core Architecture Components

Main Entry Points

  • SiftStreamBuilder: Configures and builds SiftStream instances with various options
  • SiftStream<E, T>: Generic streaming interface where E: Encoder (encodes data) and T: Transport (transmits data)
  • SiftStream<IngestionConfigEncoder, LiveStreaming>: Main streaming interface for real-time streaming to Sift (default)
  • SiftStream<IngestionConfigEncoder, FileBackup>: Streaming interface for writing data only to backup files (no live streaming)

Architecture Overview

SiftStream uses a separation of concerns between encoding (how data is structured) and transport (how data is transmitted):

  • Encoder trait: Defines how data is encoded/structured. The encoder is responsible for:

    • Converting user-provided data (e.g., Flow messages) into the appropriate message format
    • Managing flow descriptors and ingestion configuration
    • Providing metrics snapshots
    • Currently implemented by IngestionConfigEncoder for ingestion-config-based encoding
  • Transport trait: Defines how encoded messages are transmitted. The transport is responsible for:

    • Sending encoded messages to their destination
    • Managing the transmission mechanism (gRPC streams, file writing, etc.)
    • Handling cleanup and shutdown
    • Currently implemented by LiveStreaming (gRPC streaming) and FileBackup (file writing)
  • Encodeable trait: Types that can be encoded by an encoder (e.g., Flow, FlowBuilder)

This design allows for future extensibility: new encoding schemes or transport mechanisms can be added independently without affecting the other component.

Transport Modes

SiftStream supports two different transport modes:

  • LiveStreaming: The default transport that streams data directly to Sift via gRPC. This transport supports real-time streaming, optional disk backups, checkpointing, and retry policies. Use SiftStreamBuilder::build() to create a stream with this transport.

  • FileBackup: A specialized transport that only writes telemetry data to backup files on disk without streaming to Sift. This transport is useful for offline data collection, batch processing, or scenarios with unreliable network connectivity. Use SiftStreamBuilder::build_file_backup() to create a stream with this transport. Note that this transport requires a RecoveryStrategy::RetryWithBackups configuration.

Task System

The SiftStream architecture when using LiveStreaming transport consists of three main async tasks that work together to provide reliable data streaming:

  1. Backup Manager Task - Handles backup file creation and management
  2. Ingestion Task - Manages gRPC streaming to Sift
  3. Re-ingestion Task - Handles re-ingestion of backup files when failures occur

Note: FileBackup transport does not use the task system architecture, as it only writes to disk files without streaming to Sift.

Control Messages

Control messages are low-frequency messages sent between tasks via broadcast channels to coordinate checkpointing, error handling, and shutdown processes.

Task Responsibilities and Control Message Flow

1. Backup Manager Task

Responsibilities:

  • Receives data messages via backup_tx channel
  • Writes messages to backup files in PBFS format
  • Manages backup file rotation based on size/count limits
  • Handles checkpoint completion and cleanup
  • Tracks message IDs to determine which messages have been successfully committed to Sift
  • Skips backing up messages that have already been confirmed as successfully streamed (when backups are behind ingestion)

Control Messages Sent:

  • BackupFull - When backup files reach maximum count limit
  • ReingestBackups - When checkpoint fails and backup files need re-ingestion

Control Messages Received:

  • CheckpointComplete { first_message_id, last_message_id } - Signals checkpoint completion with the range of message IDs in the checkpoint
  • CheckpointNeedsReingestion { first_message_id, last_message_id } - Signals the current checkpoint will need re-ingestion with the range of message IDs
  • Shutdown - Initiates graceful shutdown

Conditions for Sending Messages:

  • BackupFull: Triggered when the number of backup files being tracked reaches the configured maximum
  • ReingestBackups: Triggered when a message (control or data) has indicated the current checkpoint should be re-ingested

2. Ingestion Task

Responsibilities:

  • Receives data messages via ingestion_tx channel
  • Creates and manages gRPC streams to Sift
  • Implements retry logic with exponential backoff
  • Handles checkpoint timing and completion

Control Messages Sent:

  • SignalNextCheckpoint - When a new stream is desired to verify messages sent have been successfully received
  • CheckpointComplete { first_message_id, last_message_id } - When the current stream has concluded at the end of a checkpoint, includes the range of message IDs in the checkpoint
  • CheckpointNeedsReingestion { first_message_id, last_message_id } - When gRPC stream fails, includes the range of message IDs that need re-ingestion

Control Messages Received:

  • BackupFull - Triggers immediate checkpoint completion, reseting the normal checkpoint interval
  • Shutdown - Initiates graceful shutdown with final stream completion

Conditions for Sending Messages:

  • SignalNextCheckpoint:
    • Timer expires (checkpoint_interval reached)
    • Backup full signal received
  • CheckpointComplete { first_message_id, last_message_id }:
    • When the existing stream has completed at the end of a checkpoint
    • During shutdown process
    • Includes the range of message IDs that were successfully streamed in this checkpoint
  • CheckpointNeedsReingestion { first_message_id, last_message_id }: When gRPC stream fails with error, includes the range of message IDs that need to be re-ingested

3. Re-ingestion Task

Responsibilities:

  • Receives backup file paths via ReingestBackups control message
  • Reads backup files and re-ingests data to Sift
  • Implements retry logic for failed re-ingestion attempts
  • Manages backup file cleanup after successful re-ingestion

Control Messages Received:

  • ReingestBackups - Contains list of backup files to re-ingest
  • Shutdown - Initiates graceful shutdown

Channel Architecture

Data Channels

  • ingestion_tx/ingestion_rx: Bounded channel for high-frequency data messages
  • backup_tx/backup_rx: Bounded channel for backup data messages

IMPORTANT: Data reliability is among the most important goals of sift-stream, however if the backup system falls behind (slow disk writes), a warning will be emitted but data will still be sent for ingestion. Streaming data in this situation is preferred over preventing it entirely. It is highly recommended however that write speeds be sufficiently high for the given data stream being backed up to ensure data is reliably backed up.

The message ID-based checkpoint system helps mitigate the impact of backups falling behind: messages that have already been confirmed as successfully streamed (via successful checkpoints) are automatically skipped during backup, allowing the backup system to catch up more efficiently.

In contrast, if the primary ingestion channel becomes full, the oldest data will be removed in favor of streaming newer data. The data removed during this process will be forwarded to the backup system and will be re-ingested at the next checkpoint.

Control Channel

  • control_tx/control_rx: Broadcast channel for low-frequency control messages

Channel Capacities

The default capacities are as follows:

  • data: 102,400
  • control: 1024

These can be configured however, based on individual streaming needs.

Checkpoint System

The checkpoint system ensures data reliability by periodically creating checkpoints that can be used for recovery. The system uses message IDs to precisely track which messages have been successfully streamed to Sift, enabling efficient backup management even when ingestion and backup processes are out of sync.

Message ID-Based Tracking

Each message in the stream is assigned a unique, monotonically increasing message ID. Checkpoints track the range of message IDs (first_message_id to last_message_id) that were included in the checkpoint, allowing the backup manager to:

  • Quickly drain committed messages: Messages that have already been confirmed as successfully streamed (message ID ≤ committed message ID) are skipped during backup, improving efficiency when backups are behind ingestion

  • Precisely match files to checkpoints: Backup files track the range of message IDs they contain, enabling exact matching with checkpoint ranges

  • Handle out-of-order scenarios: The system correctly handles cases where ingestion completes before backups catch up, or where backups are ahead of ingestion

Checkpoint Triggers

  1. Timer-based: Regular intervals (configurable via checkpoint_interval)
  2. Backup full: When backup files reach maximum count limit
  3. Manual: During shutdown process

Checkpoint Process

  1. Checkpoint Signal: Timer expires or backup full signal received
  2. Stream Completion: Current gRPC stream completes sending all buffered data
  3. Checkpoint Complete: CheckpointComplete { first_message_id, last_message_id } control message sent with the range of message IDs in the checkpoint
  4. Backup Cleanup: Backup files either deleted (success) or queued for re-ingestion (failure) based on matching message ID ranges

When a stream completes, an "ok" gRPC status from Sift indicates all messages for that stream have been received. The checkpoint includes the range of message IDs that were successfully streamed.

Backup files can also be retained regardless of successful checkpoints and re-ingestion processes.

Backup File Management

The backup system manages files through a rotation policy that balances data reliability with storage efficiency. Understanding how backup files are handled is crucial for optimizing checkpoint behavior and minimizing unnecessary re-ingestion.

Backup File Lifecycle

  1. File Creation: New backup files are created when the first message arrives that needs backing up
  2. Data Writing: Messages are written to the current backup file, with each file tracking the range of message IDs it contains (first_message_id to last_message_id)
  3. File Rotation: Files are rotated when they exceed size limits
  4. Checkpoint Completion: Files are processed based on checkpoint message ID ranges - files containing only committed messages are deleted (success) or queued for re-ingestion (failure)
  5. Re-ingestion: Failed checkpoints trigger re-ingestion of all files whose message ID ranges overlap with the failed checkpoint range

Message ID-Based File Processing

The backup manager processes checkpoints in order as backup files catch up to them. When a checkpoint completes:

  • Files containing message IDs that are fully within a successful checkpoint range are deleted

  • Files containing message IDs that overlap with a failed checkpoint range are marked for re-ingestion

  • Files containing message IDs beyond the current checkpoint are retained until their checkpoints complete

Backup File Configuration

The backup system is configured through the DiskBackupPolicy with two key parameters:

  • max_file_size: Maximum size per backup file before rotation
  • max_file_count: Maximum number of backup files to maintain before forcing a checkpoint

Selecting Optimal Backup File Parameters

Key Principle: Smaller, more frequent checkpoints reduce the amount of data that needs to be re-ingested when failures occur, but increase the overhead of checkpoint management and creating new gRPC streams to Sift.

Generally, the default configuration should be good for most use cases, and is the recommended configuration.

Error Handling and Recovery

Error Scenarios

  1. gRPC Stream Failure:

    • Ingestion task sends CheckpointNeedsReingestion
    • Backup manager queues the current checkpoint's backup files for re-ingestion
    • Re-ingestion task processes backup files asynchronously/concurrently to regular ingestion
  2. gRPC Re-Ingest Failure:

    • Backoff retries for automated recovery
    • Data remains persisted to disk if all retries fail
  3. Channel Overflow:

    • Data channels are bounded to prevent unbounded memory growth
    • Messages may be dropped if channels are full

Retry Logic

  • Exponential Backoff: Configurable retry policy with exponential backoff
  • Retry Counters: Metrics track retry attempts and failures
  • Backup Recovery: Failed data is automatically re-ingested from backup files

Shutdown Process

The shutdown process ensures graceful termination with data preservation:

Shutdown Sequence

  1. Shutdown Signal: Shutdown control message sent to all tasks
  2. Channel Closure: Data channels (ingestion_tx, backup_tx) are dropped
  3. Final Stream: Ingestion task completes final gRPC stream
  4. Checkpoint Complete: Final CheckpointComplete message sent
  5. Task Completion: All tasks complete and return results
  6. Cleanup: Backup files are cleaned up based on success/failure status

In order to shutdown quickly and not block the calling application, re-ingestion will be halted, though backup files not re-ingested will remain on disk for manual ingestion.

The data channel will be drained, and the final backup file will be sync'd/flushed to disk to preserve all data.

IMPORTANT: Calling applications must ensure graceful shutdown by calling finish() prior to dropping the SiftStream, otherwise data loss may occur.

Data Flow Architecture

Normal Operation Flow

  1. User sends dataSiftStream::send() with an Encodeable type (e.g., Flow)
  2. EncodingEncoder converts the data to the appropriate message format
  3. TransportTransport sends the encoded message
    • For LiveStreaming: Message ID assignment → Dual routing to ingestion_tx and backup_tx channels
    • For FileBackup: Direct writing to backup files
  4. Parallel processing (LiveStreaming only):
    • Ingestion task → gRPC stream → Sift
    • Backup task → Backup files
  5. Checkpoint completion → Cleanup or re-ingestion (LiveStreaming only)

Failure Recovery Flow

  1. gRPC failureCheckpointNeedsReingestion { first_message_id, last_message_id } signal with message ID range
  2. Backup manager → Matches backup files to the failed checkpoint range based on message IDs and queues them for re-ingestion
  3. Re-ingestion task → Reads backup files and re-streams to Sift
  4. Success → Backup files deleted
  5. Failure → Retry with exponential backoff

Metrics and Monitoring

The system provides comprehensive metrics for monitoring:

  • Message counts: Messages sent, received, dropped
  • Byte counts: Bytes sent, received
  • Checkpoint metrics: Checkpoint success/failure counts, timing
  • Retry metrics: Retry counts, backoff timing
  • Backup metrics: Backup file counts, sizes, rotation events, committed message ID, queued checkpoints, queued file contexts

Key Design Principles

  1. Fault Tolerance: Multiple layers of error handling and recovery
  2. Data Reliability: Backup system ensures no data loss
  3. High Throughput: Bounded channels prevent memory issues
  4. Graceful Shutdown: Clean termination with data preservation
  5. Observability: Comprehensive metrics and logging
  6. Configurability: Flexible configuration for different use cases

Conclusion

The SiftStream architecture provides a robust, fault-tolerant system for streaming telemetry data to Sift. The task-based design with control message coordination ensures reliable data delivery while maintaining high performance and providing comprehensive error recovery mechanisms.

Commit count: 411

cargo fmt