| Crates.io | rotoml |
| lib.rs | rotoml |
| version | 0.1.2 |
| created_at | 2025-04-11 22:11:39.486832+00 |
| updated_at | 2026-01-03 09:46:46.198952+00 |
| description | A native Rust AutoML pipeline toolkit |
| homepage | |
| repository | https://github.com/okanyenigun/rotoml |
| max_upload_size | |
| id | 1630387 |
| size | 88,948 |
A native Rust AutoML toolkit for machine learning pipelines with powerful data analysis and manipulation capabilities.
🚀 Version 0.1.2: Enhanced data operations and comprehensive analysis features.
cargo install rotoml
Or add to your Cargo.toml:
[dependencies]
rotoml = "0.1.2"
# Analyze CSV file
rotoml --file data.csv
# Analyze Parquet file
rotoml --file data.parquet
use rotoml::data_loader::DataLoader;
use rotoml::data_operations::DataOperations;
use rotoml::data_reporter::DataReporter;
// Load data
let df = DataLoader::load("data.csv")?;
// Detect duplicates
let (dup_count, dup_indexes) = DataOperations::count_duplicate_rows(&df)?;
let duplicate_columns = DataOperations::detect_duplicate_columns(&df)?;
// Drop columns
let df = DataOperations::drop_columns(df, &["col1", "col2"])?;
// Drop rows
let df = DataOperations::drop_rows(df, &[0, 5, 10])?;
// Generate report
DataReporter::generate_data_report(&df, "data.csv", "report.md")?;
load(file_path) - Auto-detect and load CSV or Parquetload_csv(file_path) - Load CSV fileload_parquet(file_path) - Load Parquet filedrop_column(df, column_name) - Drop a single columndrop_columns(df, column_names) - Drop multiple columnsdrop_rows(df, indexes) - Drop rows by indexdetect_duplicate_columns(df) - Find duplicate columnscount_duplicate_rows(df) - Count and list duplicate rowsgenerate_data_report(df, file_name, output_path) - Generate comprehensive analysis reportThe generated report includes:
RotoML is evolving into a complete automated machine learning pipeline:
Contributions are welcome! Please feel free to submit a Pull Request.
MIT License - see LICENSE file for details.
Okan Yenigün (okanyenigun@gmail.com)