Crates.io | naja_async_runtime |
lib.rs | naja_async_runtime |
version | 0.5.0-deprecated.1 |
source | src |
created_at | 2019-07-01 14:20:18.111597 |
updated_at | 2020-12-08 21:11:56.803494 |
description | A lightweight universal runtime for futures. |
homepage | |
repository | https://github.com/najamelan/async_runtime |
max_upload_size | |
id | 145175 |
size | 171,667 |
A lightweight runtime for global spawning of futures.
DEPRECATED: This crate has not been updated in a long time. It would need a rewrite, but it turns out most people are more interested an a parameter based approach rather than a global executor approach. I have in the mean time released async_executors and async_nursery wich allow libraries to take an agnostik executor without setting a global one. There is another crate called agnostik which does more or less the same thing. I have come to firmly believe that this is better design and there has been no demand for a global executor solution since. Thus I won't update this crate unless there is some serious demand.
The purpose of async_runtime
is to make it convenient to spawn and run futures.
It allows library authors to call rt::spawn( future );
rather than having to take a T: Executor
,
yet let client code decide what kind of executor is used. It avoids pulling in the entire network stack/reactor of
several runtime crates like runtime, tokio, async-std just because you want to use a few libraries that provide
an async interface and they all depend on their favorite runtime. async_runtime
tries to keep it fast and light.
Some key features:
macro attributes to turn async fn in sync fn (can be used on main, tests or arbitrary async fn)
tries to do one thing and do it good (does not pull network/timer dependencies)
support a variety of executors, including single threaded ones that allow spawning !Send
futures.
lightweight with very few dependencies
library authors can spawn, application authors can decide which executor is used on each thread.
wasm support (this means that a library which compiles on wasm does not need any specific code, just use rt::spawn
just the same)
With cargo add:
cargo add async_runtime
With cargo yaml:
dependencies:
async_runtime: { version: ^0.4, package: naja_async_runtime }
With raw Cargo.toml
[dependencies]
async_runtime = { version = "^0.4", package = "naja_async_runtime" }
Minimum required rustc version: 1.39.
Please check out the changelog when upgrading.
These features enable extra functionality:
macros
: proc macro attributes to turn an async fn into a sync one.juliex
: the juliex executor.async_std
: the async-std executor.localpool
: the localpool.bindgen
: the wasm-bindgen backed executor.Note for library authors. You should not enable any features on async_runtime
. The per thread executor is chosen by the application developer (exception: your library is creating the threads).
This crate has few dependiencies. Cargo will automatically handle it's dependencies for you, except that
when you are an application developer, you must enable at least one executor and you might want to enable
the macros
feature.
The basic concept is this. When you create a thread, you call [init
] to decide which executor will be used
for calls to spawn
and friends in the thread. async_runtime
makes sure that worker threads in threadpools
are set up to continue spawning on the same threadpool. All top level functions in this library will work correctly
regardless of the chosen executor. One exception are functions ending in _local
. Those are not available on threadpools
and will return an error.
If [spawn
]* gets called on a thread for which no executor has been chosen, an error is returned.
Warning: Some executors have specific modules (like rt::localpool
) which make available functionality
specific to this particular executor. These exist for 2 reasons:
rt::spawn_handle
).
Since async-std provides a JoinHandle
, there is [async_std::spawn_handle
] to recover that instead
of a boxed variant. On other executors you can use remote_handle
from the futures library to avoid boxing.localpool
, enabled by default on non WASM targets#[ rt::localpool ]
rt::Config::LocalPool
The localpool executor, being a single threaded executor has a specific design you should be aware of. If it would poll futures immediately after spawning, the thread would be occupied by this and your code that called spawn would not return immediately. Therefor, there is a process of spawning first and then calling blocking methods in order to run the executor.
Four methods are available on the executor to run it:
I'm not going to go into detail here about what these functions do, go check their documentation. Currently
async_runtime only exposes the run
function from these. If you run into specific issues that require to
access one of the other three functions, please file an issue.
Bindgen
, enabled by default on WASM targets#[ rt::bindgen ]
rt::Config::Bindgen
Currently the only executor available on WASM. It functions like a multithreaded executor in that spawned
futures will start to be polled immediately and no run
method must be called to start the executor.
juliex
#[ rt::juliex ]
rt::Config::Juliex
A threadpool. Worker threads created will automatically have juliex set as the thread executor. This
cannot be changed through the API async_runtime
exposes right now. Futures will be polled immediately.
If you have a top level future that you block on, or that is being waited on by the macro attribute, as soon as that future is done, the progam will end, even if there are still tasks in the thread pool that haven't finished yet.
async_runtime
provides the [spawn_handle
] method to wait on your futures, but
that requires boxing the returned handle. Otherwise you can add your own synchronization like channels or
join_all
from the futures library to wait on your tasks. The futures library also provides remote_handle
.
async_std
#[ rt::async_std ]
rt::Config::AsyncStd
A threadpool. Worker threads created cannot have AsyncStd set automatically as the default executor. This means
that rt::spawn
will have some overhead to make sure the worker thread is initialized properly.
Warning: async-std does not have optional dependencies, so you will be pulling in all their dependencies, including network libraries, mio, ... which will cause bloat if you don't use them.
Futures will be polled immediately. If you have a top level future that you block on, or that is being waited on by the macro attribute, as soon as that future is done, the progam will end, even if there are still tasks in the thread pool that haven't finished yet.
async_runtime
provides the [spawn_handle
] method to wait on your futures, but
that requires boxing the returned handle. Otherwise you can add your own synchronization like channels or
join_all
from the futures library to wait on your tasks. The futures library also provides remote_handle
.
futures::executor::block_on
Please read the documentation of [block_on].
Please have a look in the examples directory of the repository.
use async_runtime as rt;
// in library code:
//
fn do_something_in_parallel() -> Result<(), RtErr>
{
rt::spawn( async
{
println!( "I'm running in async context" );
})
}
// In client code we might decide that this runs in a LocalPool, instead of a threadpool:
//
[ rt::localpool ]
//
fn main()
{
// Please look at the documentation for spawn for the possible errors here.
//
do_something_in_parallel().expect( "Spawn futures" );
}
// In this example we run a bunch of tasks in parallel. To verify that
// they run on different threads we make them all sleep for a second and
// measure the time passed when they finish. If they run on a threadpool
// time passed for all of them should be around 1 second.
#![ feature( duration_constants ) ]
use
{
async_runtime as rt,
std :: { time::{ Duration, Instant }, thread::sleep } ,
futures :: { future::{ FutureExt, join_all } } ,
};
#[ rt::juliex ]
//
async fn main()
{
let start = Instant::now();
let mut tasks = Vec::new();
for i in 0..4
{
// There isn't currently a convenient way to run tasks on a threadpool until all tasks have
// finished, or until some shutdown signal is given.
//
// This is one of the ways tasks can synchronize and wait on eachother. Another way is to wait
// on channels.
//
let (fut, handle) = async move
{
sleep( Duration::SECOND );
println!( "Time elapsed at task {} end: {} second(s).", i, start.elapsed().as_secs() );
}.remote_handle();
rt::spawn( fut ).expect( "spawn task" );
tasks.push( handle );
}
join_all( tasks ).await;
}
To use the crate in wasm, please have a look at the example in the examples directory of the repository.
The only executor available on WASM is currently bindgen.
Api documentation can be found on docs.rs.
This repository accepts contributions. Ideas, questions, feature requests and bug reports can be filed through github issues.
Pull Requests are welcome on github. By commiting pull requests, you accept that your code might be modified and reformatted to fit the project coding style or to improve the implementation. Please discuss what you want to see modified before filing a pull request if you don't want to be doing work that might be rejected.
Please file pull requests against the dev branch.
Any of the behaviors described in point 4 "Unacceptable Behavior" of the Citizens Code of Conduct are not welcome here and might get you banned. If anyone including maintainers and moderators of the project fail to respect these/your limits, you are entitled to call them out.