Crates.io | miden-proving-service |
lib.rs | miden-proving-service |
version | 0.9.2 |
source | src |
created_at | 2025-01-23 11:05:23.791922+00 |
updated_at | 2025-06-10 07:16:11.378981+00 |
description | Miden rollup proving service |
homepage | https://miden.xyz |
repository | https://github.com/0xMiden/miden-base |
max_upload_size | |
id | 1527712 |
size | 295,287 |
A service for generating Miden proofs on-demand. The binary enables spawning workers and a proxy for Miden's remote proving service. It currently supports proving individual transactions, transaction batches, and blocks.
A worker is a gRPC service that can receive transaction witnesses, proposed batches, or proposed blocks, prove them, and return the generated proofs. It can handle only one request at a time and will return an error if it is already in use. Each worker is specialized on startup to handle exactly one type of proof requests - transactions, batches, or blocks.
The proxy uses Cloudflare's Pingora crate, which provides features to create a modular proxy. It is meant to handle multiple workers with a queue, assigning a worker to each request and retrying if the worker is not available. Further information about Pingora and its features can be found in the official GitHub repository.
Install the Debian package:
set -e
sudo wget https://github.com/0xMiden/miden-base/releases/download/v0.8/miden-prover-v0.8-arm64.deb -O prover.deb
sudo wget -q -O - https://github.com/0xMiden/miden-base/releases/download/v0.8/miden-prover-v0.8-arm64.deb.checksum | awk '{print $1}' | sudo tee prover.checksum
sudo sha256sum prover.deb | awk '{print $1}' > prover.sha256
sudo diff prover.sha256 prover.checksum
sudo dpkg -i prover.deb
sudo rm prover.deb
Edit the configuration file /lib/systemd/system/miden-prover.service.env
Run the service:
sudo systemctl daemon-reload
sudo systemctl enable miden-prover
sudo systemctl start miden-prover
set -e
sudo wget https://github.com/0xMiden/miden-base/releases/download/v0.8/miden-prover-proxy-v0.8-arm64.deb -O prover-proxy.deb
sudo wget -q -O - https://github.com/0xMiden/miden-base/releases/download/v0.8/miden-prover-proxy-v0.8-arm64.deb.checksum | awk '{print $1}' | sudo tee prover-proxy.checksum
sudo sha256sum prover-proxy.deb | awk '{print $1}' > prover-proxy.sha256
sudo diff prover-proxy.sha256 prover-proxy.checksum
sudo dpkg -i prover-proxy.deb
sudo rm prover-proxy.deb
Edit the configuration file /lib/systemd/system/miden-prover-proxy.service.env
Edit the service file to specify workers /lib/systemd/system/miden-prover-proxy.service
Run the service:
sudo systemctl daemon-reload
sudo systemctl enable miden-prover-proxy
sudo systemctl start miden-prover-proxy
To build the service from a local version, from the root of the workspace you can run:
make install-proving-service
The CLI can be installed from the source code using specific git revisions with cargo install
or from crates.io with cargo install miden-proving-service
.
To start the worker service you will need to run:
miden-proving-service start-worker --port 8082 --prover-type transaction
This will spawn a worker using the port defined in the command option. The host will be 0.0.0.0 by default, or 127.0.0.1 if the --localhost flag is used. In case that the port is not provided, it will default to 50051
. This command will start a worker that can handle transaction and batch proving requests.
The --prover-type
flag is required and specifies which type of proof the worker will handle. The available options are:
transaction
: For transaction proofsbatch
: For batch proofsblock
: For block proofsEach worker can only handle one type of proof. If you need to handle multiple proof types, you should start multiple workers, each with a different proof type. Additionally, you can use the --localhost
flag to bind to 127.0.0.1 instead of 0.0.0.0.
The worker can be configured using the following environment variables:
Variable | Description | Default |
---|---|---|
MPS_WORKER_LOCALHOST |
Use localhost (127.0.0.1) | false |
MPS_WORKER_PORT |
The port number for the worker | 50051 |
MPS_WORKER_PROVER_TYPE |
The supported prover type | transaction |
For example:
export MPS_WORKER_LOCALHOST="true"
export MPS_WORKER_PORT="8082"
export MPS_WORKER_PROVER_TYPE="block"
miden-proving-service start-worker
To start the proxy service, you will need to run:
miden-proving-service start-proxy --prover-type transaction [worker1] [worker2] ... [workerN]
For example:
miden-proving-service start-proxy --prover-type transaction 0.0.0.0:8084 0.0.0.0:8085
This command will start the proxy using the workers passed as arguments. The workers should be in the format host:port
. If no workers are passed, the proxy will start without any workers and will not be able to handle any requests until one is added through the miden-proving-service add-worker
command.
The --prover-type
flag is required and specifies which type of proof the proxy will handle. The available options are:
transaction
: For transaction proofsbatch
: For batch proofsblock
: For block proofsThe proxy can only handle one type of proof at a time. When you add workers to the proxy, it will check their supported proof type. Workers that support a different proof type than the proxy will be marked as unhealthy and will not be used for proving requests.
For example, if you start a proxy with --prover-type transaction
and add these workers:
Only Worker 1 will be used for proving requests, while Workers 2 and 3 will be marked as unhealthy due to incompatible proof types.
You can customize the proxy service by setting environment variables. Possible customizations can be found by running miden-proving-service start-proxy --help
.
An example .env
file is provided in the crate's root directory. To use the variables from a file in any Unix-like operating system, you can run source <your-file>
.
At the moment, when a worker added to the proxy stops working and can not connect to it for a request, the connection is marked as retriable meaning that the proxy will try reaching another worker. The number of retries is configurable via the MPS_MAX_RETRIES_PER_REQUEST
environmental variable.
To update the workers on a running proxy, two commands are provided: add-worker
and remove-worker
. These commands will update the workers on the proxy and will not require a restart. To use these commands, you will need to run:
miden-proving-service add-worker --control-port <port> [worker1] [worker2] ... [workerN]
miden-proving-service remove-worker --control-port <port> [worker1] [worker2] ... [workerN]
For example:
# To add 0.0.0.0:8085 and 200.58.70.4:50051 to the workers list:
miden-proving-service add-workers --control-port 8083 0.0.0.0:8085 200.58.70.4:50051
# To remove 158.12.12.3:8080 and 122.122.6.6:50051 from the workers list:
miden-proving-service remove-workers --control-port 8083 158.12.12.3:8080 122.122.6.6:50051
The --control-port
flag is required to specify the port where the proxy is listening for updates. The workers are passed as arguments in the format host:port
. The port can be specified via the MPS_CONTROL_PORT
environment variable. For example:
export MPS_CONTROL_PORT="8083"
miden-proving-service add-workers 0.0.0.0:8085
Note that, in order to update the workers, the proxy must be running in the same computer as the command is being executed because it will check if the client address is localhost to avoid any security issues.
The worker service implements the gRPC Health Check standard, and includes the methods described in this official proto file.
The proxy service uses this health check to determine if a worker is available to receive requests. If a worker is not available, it will be removed from the set of workers that the proxy can use to send requests.
The worker service implements a custom status check that returns information about the worker's current state and supported proof type. The proxy service uses this status check to determine if a worker is available to receive requests and if it supports the required proof type. If a worker is not available or doesn't support the required proof type, it will be removed from the set of workers that the proxy can use to send requests.
The status check returns:
The proxy service exposes a status endpoint that provides information about the current state of the proxy and its workers. This endpoint can be accessed at http://<proxy_host>:<status_port>/status
.
The status endpoint returns a JSON response with the following information:
version
: The version of the proxysupported_proof_type
: The types of proof that the proxy supportsworkers
: A list of workers with their statusExample response:
{
"version": "0.8.0",
"prover_type": "Transaction",
"workers": [
{
"address": "0.0.0.0:50051",
"version": "0.8.0",
"status": {
"Unhealthy": {
"failed_attempts": 3,
"reason": "Unsupported prover type: Batch"
}
}
},
{
"address": "0.0.0.0:50052",
"version": "0.8.0",
"status": "Healthy"
},
]
}
The status port can be configured using the MPS_STATUS_PORT
environment variable or the --status-port
command-line argument when starting the proxy.
The service uses the tracing
crate for both logging and distributed tracing, providing structured, high-performance logs and trace data.
By default, logs are written to stdout
and the default logging level is info
. This can be changed via the RUST_LOG
environment variable. For example:
export RUST_LOG=debug
For tracing, we use OpenTelemetry protocol. By default, traces are exported to the endpoint specified by OTEL_EXPORTER_OTLP_ENDPOINT
environment variable. To consume and visualize these traces we can use Jaeger or any other OpenTelemetry compatible consumer.
The simplest way to install Jaeger is by using a Docker container. To do so, run:
docker run -d -p4317:4317 -p16686:16686 jaegertracing/all-in-one:latest
Then access the Jaeger UI at http://localhost:16686/
.
If Docker is not an option, Jaeger can also be set up directly on your machine or hosted in the cloud. See the Jaeger documentation for alternative installation methods.
The proxy includes a service that can optionally expose metrics to be consumed by Prometheus. This service is enabled by specifying a metrics port.
To enable Prometheus metrics, simply specify a port on which to expose the metrics. This can be done via environment variables or command-line arguments.
Set the following environment variable:
export MPS_METRICS_PORT=6192 # Set to enable metrics on port 6192
To disable metrics, simply don't set the MPS_METRICS_PORT environment variable.
Specify a metrics port using the --metrics-port
flag when starting the proxy:
miden-proving-service start-proxy --metrics-port 6192 [worker1] [worker2] ... [workerN]
If you don't specify a metrics port, metrics will be disabled.
When enabled, the Prometheus metrics will be available at http://0.0.0.0:<metrics_port>
(e.g., http://0.0.0.0:6192
).
The metrics architecture works by having the proxy expose metrics at an endpoint (/metrics
) in a format Prometheus can read. Prometheus periodically scrapes this endpoint, adds timestamps to the metrics, and stores them in its time-series database. Then, we can use tools like Grafana to query Prometheus and visualize these metrics in configurable dashboards.
The simplest way to install Prometheus and Grafana is by using Docker containers. To do so, run:
docker run \
-d \
-p 9090:9090 \
-v /path/to/prometheus.yml:/etc/prometheus/prometheus.yml \
prom/prometheus
docker run -d -p 3000:3000 --name grafana grafana/grafana-enterprise:latest
In case that Docker is not an option, Prometheus and Grafana can also be set up directly on your machine or hosted in the cloud. See the Prometheus documentation and Grafana documentation for alternative installation methods.
A prometheus configuration file is provided in this repository, you will need to modify the scrape_configs
section to include the URL of the proxy service (e.g., http://0.0.0.0:6192
).
Then, to add the new Prometheus collector as a datasource for Grafana, you can follow this tutorial. A Grafana dashboard under the name proxy_grafana_dashboard.json
is provided, see this link to import it. Otherwise, you can create your own dashboard using the metrics provided by the proxy and export it by following this link.
Description of this crate's feature:
Features | Description |
---|---|
concurrent |
Enables concurrent code to speed up runtime execution. |
This project is MIT licensed.