Crates.io | burstq |
lib.rs | burstq |
version | 0.1.0 |
source | src |
created_at | 2024-04-06 22:42:33.670476 |
updated_at | 2024-04-06 22:42:33.670476 |
description | A multi-producer, multi-consumer channel that supports sending or receiving multiple values in a single operation. |
homepage | |
repository | https://github.com/tedsta/burstq |
max_upload_size | |
id | 1198608 |
size | 43,021 |
A multi-producer, multi-consumer channel that supports sending or receiving multiple items in a single operation.
Currently, only async and busy-waiting modes are supported. If you want to block the current thread, you'll need to use a minimal async executor such as pollster.
If you want to send and receive items in relatively large batches (e.g. upwards of 10 items per batch) but allow dynamic and uneven batch sizes, this may be the crate for you.
If you are sending and receiving one item at a time, or, more generally, have an exact batch size that is the same at both senders and receivers, you may be better off using one of the many other channel crates.
let (tx, rx) = mpmc::<u32>(5);
let thread = std::thread::spawn(move || {
pollster::block_on(async move {
let mut next = 0;
let payload: Vec<_> = (0..10).collect();
while next < 10 {
let n = tx.send(10 - next, |w| {
let len = w.len();
w.write_slice(&payload[next..next + len]);
})
.await
.unwrap();
next += n;
}
});
});
let received = pollster::block_on(async move {
let mut received = Vec::new();
while received.len() < 10 {
rx.recv(10 - received.len(), |r| {
received.extend(r);
})
.await
.unwrap();
}
received
});
thread.join().unwrap();
assert_eq!((0..10).sum::<u32>(), received.iter().sum());
The core lock-free enqueue / dequeue algorithm is based on DPDK's rte_ring
. In particular, it implements the "burst" (as opposed to "bulk") behavior of rte_ring
where if not all requested items can be enqueued/dequeued, as many as is currently possible will be.
The async-ness of burstq is achieved using the async-event crate.
MIT