| Crates.io | evlib |
| lib.rs | evlib |
| version | 0.8.7 |
| created_at | 2025-05-20 18:39:10.677644+00 |
| updated_at | 2025-10-23 20:56:16.427555+00 |
| description | Event Camera Data Processing Library |
| homepage | https://github.com/tallamjr/evlib |
| repository | https://github.com/tallamjr/evlib |
| max_upload_size | |
| id | 1681944 |
| size | 2,475,524 |
|
|
An event camera processing library with Rust backend and Python bindings, designed for scalable data processing with real-world event camera datasets.
In Development: Advanced neural network processing (hopefully with Rust
backend, maybe Candle) Real-time visualization (Only simulated working at the
moment — see wasm-evlib)
Note*: The Rust backend currently focuses on data loading and processing, with Python modules providing advanced features like filtering and representations.
Event cameras (also called neuromorphic or dynamic vision sensors) differ fundamentally from traditional frame-based cameras. Instead of capturing images at fixed frame rates, they operate asynchronously, with each pixel independently reporting changes in brightness as they occur.
Each event is represented as a 4-tuple:
$$e = (x, y, t, p)$$
Where:
An event is triggered when the logarithmic brightness change exceeds a threshold:
$$\log(L(x,y,t)) - \log(L(x,y,t_{last})) > \pm C$$
where $L(x,y,t)$ is the brightness at pixel $(x,y)$ at time $t$, and $C$ is the contrast threshold.
Key advantages:
Event cameras excel at tracking fast motion, operating in challenging lighting conditions, and applications requiring precise temporal information like robotics, autonomous vehicles, and augmented reality.
Below is an overview of what the data "looks" like...
import evlib
# Load events from any supported format (automatic detection)
df = evlib.load_events("data/prophersee/samples/evt2/80_balls.raw").collect(engine='streaming')
# Or load as LazyFrame for memory-efficient processing
lf = evlib.load_events("data/prophersee/samples/evt2/80_balls.raw")
# Basic event information
print(f"Loaded {len(df)} events")
print(f"Resolution: {df['x'].max()} x {df['y'].max()}")
print(f"Duration: {df['t'].max() - df['t'].min()}")
# Convert to NumPy arrays for compatibility
x_coords = df['x'].to_numpy()
y_coords = df['y'].to_numpy()
timestamps = df['t'].to_numpy()
polarities = df['polarity'].to_numpy()
import evlib
import polars as pl
# Load events as LazyFrame for efficient processing
events = evlib.load_events("data/prophersee/samples/evt3/pedestrians.raw")
# Time filtering using Polars operations
time_filtered = events.with_columns([
(pl.col('t').dt.total_microseconds() / 1_000_000).alias('time_seconds')
]).filter(
(pl.col('time_seconds') >= 0.1) & (pl.col('time_seconds') <= 0.5)
)
# Spatial filtering (Region of Interest)
spatial_filtered = time_filtered.filter(
(pl.col('x') >= 100) & (pl.col('x') <= 500) &
(pl.col('y') >= 100) & (pl.col('y') <= 400)
)
# Polarity filtering
polarity_filtered = spatial_filtered.filter(pl.col('polarity') == 1)
# Collect final results
filtered_df = polarity_filtered.collect()
print(f"Filtered to {len(filtered_df)} events")
evlib provides comprehensive event representation functions for computer vision and neural network applications:
import evlib
import evlib.representations as evr
import polars as pl
# Load events and create representations
events = evlib.load_events("data/prophersee/samples/hdf5/pedestrians.hdf5")
events_df = events.collect()
# Create stacked histogram (replaces RVT preprocessing)
hist = evr.create_stacked_histogram(
events_df,
height=180, width=240,
bins=5, window_duration_ms=50.0,
_count_cutoff=5
)
print(f"Created stacked histogram with {len(hist)} spatial bins")
# Create mixed density stack representation
density = evr.create_mixed_density_stack(
events_df,
height=180, width=240,
window_duration_ms=50.0
)
print(f"Created mixed density stack with {len(density)} entries")
# Create voxel grid representation
voxel = evr.create_voxel_grid(
events_df,
height=180, width=240,
n_time_bins=3
)
print(f"Created voxel grid with {len(voxel)} voxels")
# Advanced representations (require data type conversion)
# Convert timestamp and ensure proper dtypes for advanced functions
small_events = events.limit(10000).collect()
converted_events = small_events.with_columns([
pl.col('t').dt.total_microseconds().cast(pl.Float64).alias('t_microseconds'),
pl.col('x').cast(pl.Int64),
pl.col('y').cast(pl.Int64),
pl.col('polarity').cast(pl.Int64)
]).select(['x', 'y', 't_microseconds', 'polarity']).rename({'t_microseconds': 't'})
# Create time surface representation
time_surface = evr.create_timesurface(
converted_events,
height=180, width=240,
dt=50000.0, # time step in microseconds
tau=10000.0 # decay constant in microseconds
)
print(f"Created time surface with {len(time_surface)} pixels")
# Create averaged time surface
avg_time_surface = evr.create_averaged_timesurface(
converted_events,
height=180, width=240,
cell_size=1, surface_size=1,
time_window=50000.0, tau=10000.0
)
print(f"Created averaged time surface with {len(avg_time_surface)} pixels")
RVT processing example:
# Let's load in some un-procedded RVT data, i.e. gen4_1mpx_original
In [5]: events = evlib.load_events("/Users/tallam/github/tallamjr/origin/evlib/data/gen4_1mpx
⋮ _original/val/moorea_2019-02-21_000_td_2257500000_2317500000_td.h5")
In [6]: events
Out[6]: <LazyFrame at 0x11D6EBE30>
# How many events in this window
In [7]: events.select(pl.len()).collect(engine="streaming")
Out[7]:
shape: (1, 1)
┌───────────┐
│ len │
│ --- │
│ u32 │
╞═══════════╡
│ 540124055 │
└───────────┘
# That's 500+ million events!
# Now let's process it and create stacked histograms ready for the RVT model
In [8]: hist = evr.create_stacked_histogram(
...: events,
...: height=480, width=640,
...: bins=10, window_duration_ms=50.0
...: )
...: print(f"Created stacked histogram with {len(hist)} spatial bins")
Created stacked histogram with 1519652 spatial bins
# 500M -> 1.5M in seconds :-)
In [9]: hist
Out[9]:
shape: (1_519_652, 5)
┌──────────┬──────────┬─────┬─────┬───────┐
│ time_bin ┆ polarity ┆ y ┆ x ┆ count │
│ --- ┆ --- ┆ --- ┆ --- ┆ --- │
│ i32 ┆ i8 ┆ i16 ┆ i16 ┆ u32 │
╞══════════╪══════════╪═════╪═════╪═══════╡
│ 0 ┆ 1 ┆ 0 ┆ 0 ┆ 4 │
│ 0 ┆ 1 ┆ 0 ┆ 1 ┆ 3 │
│ 0 ┆ 1 ┆ 0 ┆ 2 ┆ 5 │
│ 0 ┆ 1 ┆ 0 ┆ 3 ┆ 6 │
│ 0 ┆ 1 ┆ 0 ┆ 4 ┆ 5 │
│ … ┆ … ┆ … ┆ … ┆ … │
│ 9 ┆ 1 ┆ 479 ┆ 563 ┆ 1 │
│ 9 ┆ 1 ┆ 479 ┆ 624 ┆ 1 │
│ 9 ┆ 1 ┆ 479 ┆ 626 ┆ 1 │
│ 9 ┆ 1 ┆ 479 ┆ 638 ┆ 1 │
│ 9 ┆ 1 ┆ 479 ┆ 639 ┆ 1 │
└──────────┴──────────┴─────┴─────┴───────┘
pip install evlib
# For Polars DataFrame support (recommended)
pip install evlib[polars]
# For PyTorch integration
pip install evlib[pytorch]
We recommend using uv for fast, reliable Python package management:
# Install uv (if not already installed)
curl -LsSf https://astral.sh/uv/install.sh | sh
# Clone the repository
git clone https://github.com/tallamjr/evlib.git
cd evlib
# Create virtual environment and install dependencies
uv venv --python 3.12
source .venv/bin/activate # On Windows: .venv\Scripts\activate
uv pip install -e ".[dev,polars]"
# Build the Rust extensions
maturin develop
# Ubuntu/Debian
sudo apt install libhdf5-dev pkg-config
# macOS
brew install hdf5 pkg-config
# Windows
# No additional system dependencies required
# HDF5 file operations use pure Python h5py library
For optimal performance, ensure you have the recommended system configuration:
System Requirements:
Installation for Performance:
# Install with Polars support (recommended)
pip install "evlib[polars]"
# For development with all performance features (using uv)
uv pip install "evlib[dev,polars]"
# Verify installation with benchmark
python -c "import evlib; print('evlib installed successfully')"
python benchmark_memory.py # Test memory efficiency
Optional Performance Dependencies:
# For advanced memory monitoring
uv pip install psutil
# For parallel processing (already included in dev)
uv pip install multiprocessing-logging
Windows:
save_events_to_hdf5) is not available due to Rust HDF5 library limitationsimport h5py
import numpy as np
with h5py.File("output.h5", "w") as f:
f.create_dataset("events/x", data=x_coords)
f.create_dataset("events/y", data=y_coords)
f.create_dataset("events/t", data=timestamps)
f.create_dataset("events/p", data=polarities)
Linux/macOS:
evlib provides comprehensive Polars DataFrame support for high-performance event data processing:
import evlib
# Load as LazyFrame (recommended)
events = evlib.load_events("data/prophersee/samples/evt2/80_balls.raw")
df = events.collect() # Collect to DataFrame when needed
# Automatic format detection and optimization
events = evlib.load_events("data/prophersee/samples/evt2/80_balls.raw") # EVT2 format automatically detected
print(f"Format: {evlib.formats.detect_format('data/prophersee/samples/evt2/80_balls.raw')}")
print(f"Description: {evlib.formats.get_format_description('EVT2')}")
import evlib
import polars as pl
# Chain operations with LazyFrames for optimal performance
events = evlib.load_events("data/prophersee/samples/hdf5/pedestrians.hdf5")
result = events.filter(pl.col("polarity") == 1).with_columns([
pl.col("t").dt.total_microseconds().alias("time_us"),
(pl.col("x") + pl.col("y")).alias("diagonal_pos")
]).collect()
# Memory-efficient temporal analysis
time_stats = events.with_columns([
pl.col("t").dt.total_microseconds().alias("time_us")
]).group_by([
(pl.col("time_us") // 1_000_000).alias("time_second") # Group by second
]).agg([
pl.len().alias("event_count"),
pl.col("polarity").mean().alias("avg_polarity")
]).collect()
# Complex filtering operations with Polars
filtered = events.with_columns([
(pl.col('t').dt.total_microseconds() / 1_000_000).alias('time_seconds')
]).filter(
(pl.col('time_seconds') >= 0.1) & (pl.col('time_seconds') <= 0.5)
)
analysis = filtered.with_columns([
pl.col("t").dt.total_microseconds().alias("time_us")
]).collect()
import evlib
import polars as pl
import evlib.filtering as evf
# Built-in format detection
format_info = evlib.formats.detect_format("data/prophersee/samples/evt3/pedestrians.raw")
print(f"Detected format: {format_info}")
# Spatial filtering using Polars operations
events = evlib.load_events("data/prophersee/samples/evt3/pedestrians.raw")
spatial_filtered = events.filter(
(pl.col("x") >= 100) & (pl.col("x") <= 200) &
(pl.col("y") >= 50) & (pl.col("y") <= 150)
)
# Chain multiple filters efficiently
complex_filtered = events.filter(
(pl.col("x") >= 100) & (pl.col("x") <= 200) &
(pl.col("y") >= 50) & (pl.col("y") <= 150) &
(pl.col("polarity") == 1)
)
# Temporal analysis with Polars operations
rates = events.with_columns([
pl.col("t").dt.total_microseconds().alias("time_us")
]).group_by([
(pl.col("time_us") // 10_000).alias("time_10ms") # Group by 10ms
]).agg([
pl.len().alias("event_rate"),
pl.col("polarity").mean().alias("avg_polarity")
]).collect()
# Save processed data (working example)
processed = events.with_columns([
(pl.col('t').dt.total_microseconds() / 1_000_000).alias('time_seconds')
]).filter(
(pl.col('time_seconds') >= 0.1) & (pl.col('time_seconds') <= 0.5)
)
processed_df = processed.collect()
data_arrays = processed_df.select(["x", "y", "t", "polarity"]).to_numpy()
x, y, t_us, p = data_arrays.T
# Convert Duration microseconds to seconds for save function
t = t_us.astype('float64') / 1_000_000
evlib.formats.save_events_to_hdf5(x.astype('int16'), y.astype('int16'), t, p.astype('int8'), "output.h5")

Benchmark Results:
Run performance benchmarks to verify optimizations:
# Verify README performance claims and generate plots
python benches/benchmark_performance_readme.py
# Memory efficiency benchmark
python benches/benchmark_memory.py
# Test with your own data
python -c "
import evlib
import time
start = time.time()
events = evlib.load_events('data/prophersee/samples/evt2/80_balls.raw')
df = events.collect()
print(f'Loaded {len(df):,} events in {time.time()-start:.2f}s')
print(f'Format: {evlib.detect_format(\"data/prophersee/samples/evt2/80_balls.raw\")}')
print(f'Memory per event: {df.estimated_size() / len(df):.1f} bytes')
"
import evlib
import evlib.filtering as evf
import polars as pl
# Small files (<5M events) - Direct loading
events_small = evlib.load_events("data/prophersee/samples/evt2/80_balls.raw")
df_small = events_small.collect()
# Large files (>5M events) - Automatic streaming
events_large = evlib.load_events("data/prophersee/samples/hdf5/pedestrians.hdf5")
# Same API, automatically uses streaming for memory efficiency
# Memory-efficient filtering on large datasets using Polars
filtered = events_large.with_columns([
(pl.col('t').dt.total_microseconds() / 1_000_000).alias('time_seconds')
]).filter(
(pl.col('time_seconds') >= 1.0) & (pl.col('time_seconds') <= 2.0)
)
positive_events = filtered.filter(pl.col("polarity") == 1)
# Collect only when needed for memory efficiency
result_df = positive_events.collect()
print(f"Filtered to {len(result_df)} events")
import evlib
import psutil
import os
def monitor_memory():
process = psutil.Process(os.getpid())
return process.memory_info().rss / 1024 / 1024 # MB
# Monitor memory usage during loading
initial_mem = monitor_memory()
events = evlib.load_events("data/prophersee/samples/evt2/80_balls.raw")
df = events.collect()
peak_mem = monitor_memory()
print(f"Memory used: {peak_mem - initial_mem:.1f} MB")
print(f"Memory per event: {(peak_mem - initial_mem) * 1024 * 1024 / len(df):.1f} bytes")
print(f"Polars DataFrame size: {df.estimated_size() / 1024 / 1024:.1f} MB")
benchmark_memory.py to track usage.lazy() for complex filtering chainsIssue: Out of memory errors
import evlib
import evlib.filtering as evf
# Solution: Use filtering before collecting (streaming activates automatically)
events = evlib.load_events("data/prophersee/samples/hdf5/pedestrians.hdf5")
# Streaming activates automatically for files >5M events
# Apply filtering before collecting to reduce memory usage
filtered = events.with_columns([
(pl.col('t').dt.total_microseconds() / 1_000_000).alias('time_seconds')
]).filter(
(pl.col('time_seconds') >= 0.1) & (pl.col('time_seconds') <= 0.5)
)
df = filtered.collect() # Only collect when needed
# Or stream to disk using Polars
filtered.sink_parquet("filtered_events.parquet")
Issue: Slow loading performance
import evlib
import evlib.filtering as evf
import polars as pl
# Solution: Use LazyFrame for complex operations
events = evlib.load_events("data/prophersee/samples/evt2/80_balls.raw")
# Use Polars operations for optimized filtering
result = events.filter(
(pl.col("x") >= 0) & (pl.col("x") <= 640) &
(pl.col("y") >= 0) & (pl.col("y") <= 480)
)
df = result.collect()
# Or chain Polars operations
result = events.filter(pl.col("polarity") == 1).select(["x", "y", "t"]).collect()
Issue: Memory usage higher than expected
import evlib
# Solution: Monitor and verify optimization
events = evlib.load_events("data/prophersee/samples/evt3/pedestrians.raw")
df = events.collect()
print(f"Memory efficiency: {df.estimated_size() / len(df)} bytes/event")
print(f"DataFrame schema: {df.schema}")
print(f"Number of events: {len(df):,}")
# Check format detection
format_info = evlib.formats.detect_format("data/prophersee/samples/evt3/pedestrians.raw")
print(f"Format: {format_info}")
evlib provides several Python modules for different aspects of event processing:
evlib.formats: Direct Rust access for format loading and detectionevlib.filtering: High-performance event filtering with Polarsevlib.representations: Event representations (stacked histograms, voxel grids)evlib.models: Neural network model loading and inference (Under construction)import evlib
import evlib.filtering as evf
import evlib.representations as evr
# Core event loading (returns Polars LazyFrame)
events = evlib.load_events("data/prophersee/samples/hdf5/pedestrians.hdf5")
# Format detection and description
format_info = evlib.formats.detect_format("data/prophersee/samples/hdf5/pedestrians.hdf5")
description = evlib.formats.get_format_description("HDF5")
# Advanced filtering using Polars operations
filtered = events.with_columns([
(pl.col('t').dt.total_microseconds() / 1_000_000).alias('time_seconds')
]).filter(
(pl.col('time_seconds') >= 0.1) & (pl.col('time_seconds') <= 0.5)
)
time_filtered = filtered.collect()
# Event representations (working examples)
events_df = events.collect()
hist = evr.create_stacked_histogram(events_df, height=180, width=240, bins=5)
voxel = evr.create_voxel_grid(events_df, height=180, width=240, n_time_bins=3)
# Advanced representations (with proper data conversion)
small_events = events.limit(10000).collect()
converted_events = small_events.with_columns([
pl.col('t').dt.total_microseconds().cast(pl.Float64).alias('t_microseconds'),
pl.col('x').cast(pl.Int64),
pl.col('y').cast(pl.Int64),
pl.col('polarity').cast(pl.Int64)
]).select(['x', 'y', 't_microseconds', 'polarity']).rename({'t_microseconds': 't'})
time_surface = evr.create_timesurface(converted_events, height=180, width=240, dt=50000.0, tau=10000.0)
# Neural network models (limited functionality)
from evlib.models import ModelConfig # If available
# Data saving (working examples)
df = events.collect()
data_arrays = df.select(["x", "y", "t", "polarity"]).to_numpy()
x, y, t_us, p = data_arrays.T
# Convert Duration microseconds to seconds for save functions
t = t_us.astype('float64') / 1_000_000
evlib.formats.save_events_to_hdf5(x.astype('int16'), y.astype('int16'), t, p.astype('int8'), "output.h5")
evlib.formats.save_events_to_text(x.astype('int16'), y.astype('int16'), t, p.astype('int8'), "output.txt")
evlib includes an optimized PyTorch dataloader implementation that showcases best practices for event camera data processing:
.to_torch() conversion for zero-copy data transfer# New: Use the built-in PyTorch integration
import evlib
import torch
from torch.utils.data import DataLoader
from evlib.pytorch import create_dataloader, load_rvt_data, PolarsDataset, create_rvt_transform
# Option 1: Raw event data (since RVT data not available in CI)
events = evlib.load_events("data/slider_depth/events.txt")
dataloader = create_dataloader(events, data_type="events", batch_size=256)
# Option 2: Manual setup for custom transforms (for RVT data when available)
# lazy_df = load_rvt_data("data/gen4_1mpx_processed_RVT/val/moorea_2019-02-21_000_td_2257500000_2317500000")
# Option 3: Raw event data from various formats
# events = evlib.load_events("data/eTram/h5/val_2/val_night_007_td.h5") # eTram dataset
# events = evlib.load_events("data/prophersee/samples/hdf5/pedestrians.hdf5") # Prophesee format
events = evlib.load_events("data/slider_depth/events.txt") # Text format
dataloader = create_dataloader(events, data_type="events")
# Option 4: Advanced - Custom transform using provided functions (for RVT data)
# if lazy_df is not None:
# # Use the built-in RVT transform
# transform = create_rvt_transform()
# dataset = PolarsDataset(lazy_df, batch_size=256, shuffle=True,
# transform=transform, drop_last=True)
# dataloader = DataLoader(dataset, batch_size=None, num_workers=0)
# Option 5: Custom transform (if you need to modify the feature extraction)
def custom_split_features_labels(batch):
"""Custom transform to separate RVT features and labels from Polars batch"""
feature_tensors = []
# Add all temporal bin features (mean, std, max, nonzero for each bin)
for bin_idx in range(20):
for stat in ["mean", "std", "max", "nonzero"]:
key = f"bin_{bin_idx:02d}_{stat}"
if key in batch:
feature_tensors.append(batch[key])
# Add bounding box features
for key in ["bbox_x", "bbox_y", "bbox_w", "bbox_h", "bbox_area"]:
if key in batch:
feature_tensors.append(batch[key])
# Add activity features
for key in ["total_activity", "active_pixels", "temporal_center"]:
if key in batch:
feature_tensors.append(batch[key])
# Add normalized features (note: actual feature name is "t_norm", not "timestamp_norm")
for key in ["t_norm", "bbox_area_norm", "activity_norm"]:
if key in batch:
feature_tensors.append(batch[key])
# Stack into feature matrix and extract labels
features = torch.stack(feature_tensors, dim=1) # Shape: (batch_size, 91)
labels = batch["label"].long() # Shape: (batch_size,)
return {"features": features, "labels": labels}
# Train with real event camera data
for batch in dataloader:
features = batch["features"] # Shape: (256, 91) - 91 statistical features
labels = batch["labels"] # Shape: (256,) - object class labels
# Your PyTorch training loop here
# outputs = model(features)
# loss = criterion(outputs, labels)
# ... backward pass, optimizer step, etc.
print(f"Batch features shape: {features.shape}, labels shape: {labels.shape}")
break # Just show the data format
RVT HDF5 Data → Feature Extraction → Polars LazyFrame → .to_torch() → PyTorch Training
The dataloader demonstrates:
See examples/polars_pytorch_simplified.py for the complete implementation and adapt it for your own event camera datasets.
evlib includes a complete pipeline for converting standard video files to event camera data using the ESIM algorithm, with support for Mac GPU acceleration via MPS (Metal Performance Shaders).
Use the ESIM (Event-based Simulator) algorithm to convert any video file to event data:
# Basic conversion with automatic device selection (MPS on Mac, CUDA on Linux/Windows, CPU fallback)
python scripts/esim_convert.py sample.mp4 --cp 0.3 --cn 0.3 --width 640 --height 480
# Explicitly use Mac GPU acceleration
python scripts/esim_convert.py sample.mp4 --device mps --cp 0.3 --cn 0.3 --width 640 --height 480
# Show video information and processing configuration
python scripts/esim_convert.py sample.mp4 --video_info --cp 0.3 --cn 0.3
# Process specific time range
python scripts/esim_convert.py sample.mp4 --start_time 1.0 --end_time 3.0 --cp 0.3 --cn 0.3
# Estimate event count before full processing
python scripts/esim_convert.py sample.mp4 --estimate_only --sample_frames 50
Parameters:
--cp, --positive_threshold: Positive contrast threshold (default: 0.4)--cn, --negative_threshold: Negative contrast threshold (default: 0.4)--device: Computing device (auto, cuda, mps, cpu)--width, --height: Output resolution (default: 640x480)--fps: Override video FPS--refractory_period: Minimum time between events at same pixel (ms)Performance: Achieves 490,000+ events/second processing speed with MPS acceleration on Mac.
from evlib.simulation import ESIMConfig, VideoConfig, VideoToEvents
# Configure ESIM algorithm
esim_config = ESIMConfig(
device="auto", # Automatically selects MPS on Mac, CUDA on Linux/Windows
positive_threshold=0.3,
negative_threshold=0.3,
refractory_period_ms=0.1
)
# Configure video processing
video_config = VideoConfig(
width=640,
height=480,
grayscale=True
)
# Convert video to events
processor = VideoToEvents(esim_config, video_config)
x, y, t, polarity = processor.process_video("sample.mp4")
# Save as HDF5 file
import evlib
evlib.formats.save_events_to_hdf5(x, y, t, polarity, "events.h5")
Convert the generated event data to a visualization video:
# Basic visualization of converted events
python scripts/visualize_etram.py --input h5/events_esim.h5 --output events_visualization.mp4
# High-quality visualization with custom parameters
python scripts/visualize_etram.py --input h5/events_esim.h5 --output events_hq.mp4 \
--fps 60 --decay 50 --resolution 1280x720
# Thermal colormap visualization
python scripts/visualize_etram.py --input h5/events_esim.h5 --output events_thermal.mp4 \
--colormap --colormap-type jet --fps 60
# Process time range
python scripts/visualize_etram.py --input h5/events_esim.h5 --output events_clip.mp4 \
--start-time 1.0 --duration 3.0
Complete Workflow Example:
# 1. Convert video to events (generates ~18M events from 5.7s video)
python scripts/esim_convert.py sample.mp4 --cp 0.3 --cn 0.3 --width 640 --height 480
# Output: h5/events_esim.h5 (240MB HDF5 file)
# 2. Visualize the events as a video
python scripts/visualize_etram.py --input h5/events_esim.h5 --output sample_events.mp4 --fps 30
# Output: sample_events.mp4 (event visualization video)
This pipeline allows you to:
Run examples:
# Test all notebooks
pytest --nbmake examples/
# Run specific examples
python examples/simple_example.py
python examples/filtering_demo.py
python examples/stacked_histogram_demo.py
# Run the high-performance PyTorch dataloader example
python examples/polars_pytorch_simplified.py
# Run all tests (Python and Rust)
pytest
cargo test
# Test specific modules
pytest tests/test_filtering.py
pytest tests/test_representations.py
pytest tests/test_evlib_exact_match.py
# Test notebooks (including examples)
pytest --nbmake examples/
# Test with coverage
pytest --cov=evlib
All code examples in the documentation are automatically tested to ensure they work correctly:
# Test all documentation examples
pytest --markdown-docs docs/
# Test specific documentation file
pytest --markdown-docs docs/getting-started/quickstart.md
# Use the convenient test script
python scripts/test_docs.py --list # List testable files
python scripts/test_docs.py --report # Generate report
# Test specific documentation section
pytest --markdown-docs docs/user-guide/
pytest --markdown-docs docs/getting-started/
# Format code
black python/ tests/ examples/
cargo fmt
# Run linting
ruff check python/ tests/
cargo clippy
# Check types
mypy python/evlib/
rust-toolchain.toml)# Development build
maturin develop --features python # python required to register python modules
# Build with features
maturin develop --features polars
maturin develop --features pytorch
# Release build
maturin build --release

MIT License - see LICENSE.md for details.