cubek-attention

Crates.iocubek-attention
lib.rscubek-attention
version0.1.1
created_at2025-12-05 17:13:38.608748+00
updated_at2026-01-23 17:03:14.354469+00
descriptionCubeK: Attention Kernels
homepage
repositoryhttps://github.com/tracel-ai/cubek/tree/main/crates/cubek-attention
max_upload_size
id1968745
size428,412
Nathaniel Simard (nathanielsimard)

documentation

README


Discord Current Crates.io Version Minimum Supported Rust Version Test Status license


CubeK: high-performance multi-platform kernels in CubeCL

Algorithms

Algorithms Variants
Random bernoulli normal uniform
Quantization symmetric per-block per-tensor q2 q4 q8 fp4
Reduction mean sum prod max min arg[max|min] per-cube per-plane
Matmul mma unit tma multi-stage specialization ordered multi-rows
Convolution mma unit tma multi-stage im2col
Attention mma unit multi-rows

Contributing

If you want to contribute new kernels, please read the GUIDE.md.

Commit count: 390

cargo fmt