Crates.io | virtiofsd |
lib.rs | virtiofsd |
version | 1.13.0 |
source | src |
created_at | 2021-12-23 08:25:20.116084 |
updated_at | 2024-11-27 13:33:20.184155 |
description | A virtio-fs vhost-user device daemon |
homepage | https://virtio-fs.gitlab.io/ |
repository | https://gitlab.com/virtio-fs/virtiofsd |
max_upload_size | |
id | 502099 |
size | 751,711 |
A virtio-fs vhost-user device daemon written in Rust.
This project depends on libcap-ng and libseccomp. You can obtain those dependencies by building them for their respective sources, or by installing the correspondent development packages from your distribution, if available:
dnf install libcap-ng-devel libseccomp-devel
apt install libcap-ng-dev libseccomp-dev
virtiofsd is written in Rust, so you will have to install Rust in order to compile it, and it uses cargo to manage the project and its dependencies. After installing Rust, you can compile it to a binary by running:
cargo build --release
Every time new code is merged, the CI pipeline will upload a debug binary of virtiofsd. It is intended to be an accessible way for anyone to download and test virtiofsd without needing a Rust toolchain installed.
The debug binary is built only for x86_64 Linux-based systems.
Click here to download the latest build
See CONTRIBUTING.md
This program must be run as the root user or as a "fake" root inside a user namespace (see Running as non-privileged user).
The program drops privileges where possible during startup, although it must be able to create and access files with any uid/gid:
seccomp(2)
.capabilities(7)
are dropped. virtiofsd only retains the following capabilities:
CAP_CHOWN
, CAP_DAC_OVERRIDE
, CAP_FOWNER
, CAP_FSETID
, CAP_SETGID
, CAP_SETUID
,
CAP_MKNOD
, CAP_SETFCAP
(and CAP_DAC_READ_SEARCH
if --inode-file-handles
is used).virtiofsd [FLAGS] [OPTIONS] --fd <fd>|--socket-path <socket-path> --shared-dir <shared-dir>
-h, --help
Prints help information.
-V, --version
Prints version information.
--syslog
Log to syslog. Default: stderr.
--print-capabilities
Print vhost-user.json backend program capabilities and exit.
--allow-direct-io
Honor the O_DIRECT
flag passed down by guest applications.
--announce-submounts
Tell the guest which directories are mount points.
If multiple filesystems are mounted in the shared directory,
virtiofsd passes inode IDs directly to the guest, and because such IDs
are unique only on a single filesystem, it is possible that the guest
will encounter duplicates if multiple filesystems are mounted in the
shared directory.
--announce-submounts
solves that problem because it reports a different
device number for every submount it encounters.
In addition, when running with --announce-submounts
, the client sends one
SYNCFS
request per submount that is to be synced, so virtiofsd
will call syncfs()
on each submount.
On the other hand, when running without --announce-submounts
,
the client only sends a SYNCFS
request for the root mount,
this may lead to data loss/corruption.
--no-killpriv-v2
Disable KILLPRIV V2
support.
This is required if the shared directory is an NFS file system.
KILLPRIV V2
support is disabled by default.
--killpriv-v2
Enable KILLPRIV V2
support. It is disabled by default.
--no-readdirplus
Disable support for READDIRPLUS
operations.
--writeback
Enable writeback cache.
--xattr
Enable support for extended attributes.
--posix-acl
Enable support for posix ACLs (implies --xattr).
--security-label
Enable support for security label (SELinux).
--preserve-noatime
Always preserve O_NOATIME
.
By default virtiofsd will implicitly clean up O_NOATIME
to prevent potential
permission errors. The option --preserve-noatime
can be used to override this
behavior and preserve the O_NOATIME
flag specified by the client.
--readonly
Prevent write accesses from the guest. Note that this does not make the underlying shared directory an actual read-only mount, so e.g. the access time is still updated on accesses.
--shared-dir <shared-dir>
Shared directory path.
--tag <tag>
The tag that the virtio device advertises.
Setting this option will enable advertising of VHOST_USER_PROTOCOL_F_CONFIG. However, the vhost-user frontend of your hypervisor may not negotiate this feature and (or) ignore this value. Notably, QEMU currently (as of 8.1) ignores the CONFIG feature. QEMU versions from 7.1 to 8.0 will crash while attempting to log a warning about not supporting the feature.
--socket-group <socket-group>
Name of group for the vhost-user socket.
--socket-path <socket-path>
vhost-user socket path.
--fd <fd>
File descriptor for the listening (not yet connected) socket.
--log-level <log-level>
Log level (error, warn, info, debug, trace, off).
Default: info.
--thread-pool-size <thread-pool-size>
Maximum thread pool size. A value of "0" disables the pool.
Default: 0.
--rlimit-nofile <rlimit-nofile>
Set maximum number of file descriptors.
If the soft limit is greater than 1M or --rlimit-nofile=0
is passed
as parameter, the maximum number of file descriptors is not changed.
Default: min(1000000, /proc/sys/fs/nr_open
).
--modcaps=<modcaps>
Modify the list of capabilities, e.g., --modcaps=+sys_admin:-chown
.
Although it is not mandatory, it is recommended to always use the =
sign,
in other case, this will fail --modcaps -mknod
, because it will be
interpreted as two options, instead of the intended --modcaps=-mknod
.
--sandbox <sandbox>
Sandbox mechanism to isolate the daemon process (namespace, chroot, none).
namespace: The program switches into a new file system
namespace (namespaces(7)
) and invokes pivot_root(2)
to make the shared directory
tree its root. A new mount (mount_namespaces(7)
), pid (pid_namespaces(7)
) and
net namespace (network_namespaces(7)
) is also created to isolate the process.
chroot: The program invokes chroot(2)
to make the shared
directory tree its root. This mode is intended for container environments where
the container runtime has already set up the namespaces and the program does
not have permission to create namespaces itself.
none: Do not isolate the daemon (not recommended).
Both namespace and chroot sandbox modes prevent "file system escapes" due to symlinks and other file system objects that might lead to files outside the shared directory.
Default: namespace.
--seccomp <seccomp>
Action to take when seccomp finds a not allowed syscall (none, kill, log, trap).
Default: kill.
--cache <cache>
The caching policy the file system should use (auto, always, metadata, never).
Default: auto.
--allow-mmap
For shared directories with --cache={metadata, never}
, allow files contained in the shared directory to be mmap
'd.
Regardless of the selected cache policy, this option should only be enabled when the file system has exclusive access
to the directory.
--inode-file-handles=<inode-file-handles>
When to use file handles to reference inodes instead of O_PATH
file descriptors (never, prefer, mandatory).
never: Never use file handles, always use O_PATH
file descriptors.
prefer: Attempt to generate file handles, but fall back to O_PATH
file descriptors where the underlying
filesystem does not support file handles or CAP_DAC_READ_SEARCH
is not available.
Useful when there are various different filesystems under the shared directory and some of them do not support file handles.
mandatory: Always use file handles.
It will fail if the underlying filesystem does not support file handles or CAP_DAC_READ_SEARCH
is not available.
Using file handles reduces the number of file descriptors virtiofsd keeps open, which is not only helpful with resources, but may also be important in cases where virtiofsd should only have file descriptors open for files that are open in the guest, e.g. to get around bad interactions with NFS's silly renaming (see NFS FAQ, Section D2: "What is a "silly rename"?").
Default: never.
--xattrmap <xattrmap>
Add custom rules for translating extended attributes between host and guest (e.g., :map::user.virtiofs.:
).
For additional details please see Extended attribute mapping.
--uid-map=:namespace_uid:host_uid:count:
When running virtiofsd as non-root, map a range of UIDs from host to namespace.
In order to use this option, the range of subordinate user IDs must have been set up via
subuid(5)
. virtiofsd uses newuidmap(1)
for non-trivial cases, that requires a valid subuid,
to do the mapping. If this option is not provided, virtiofsd will set up a 1-to-1 mapping for current uid.
namespace_uid: Beginning of the range of UIDs inside the user namespace. host_uid: Beginning of the range of UIDs outside the user namespace. count: Length of the ranges (both inside and outside the user namespace).
For instance, let's assume the invoking UID is 1000 and the content of /etc/subuid is: 1000:100000:65536, which creates 65536 subuids starting at 100000, i.e. the (inclusive) range [100000, 165535], belonging to the actual UID 1000. This range can be mapped to the UIDs [0, 65535] in virtiofsd’s user namespace (i.e. as seen in the guest) via --uid-map=:0:100000:65536:. Alternatively, you can simply map your own UID to a single UID in the namespace: For example, --uid-map=:0:1000:1: would map UID 1000 to root’s UID in the namespace (and thus the guest).
--gid-map=:namespace_gid:host_gid:count:
When running virtiofsd as non-root, map a range of GIDs from host to namespace.
In order to use this option, the range of subordinate group IDs must have been set up via
subgid(5)
. virtiofsd uses newgidmap(1)
for non-trivial cases, that requires a valid subgid,
to do the mapping. If this option is not provided, virtiofsd will set up a 1-to-1 mapping for current gid.
namespace_gid: Beginning of the range of GIDs inside the user namespace. host_gid: Beginning of the range of GIDs outside the user namespace. count: Length of the ranges (both inside and outside the user namespace).
For instance, let's assume the invoking GID is 1000 and the content of /etc/subgid is: 1000:100000:65536, which creates 65536 subgids starting at 100000, i.e. the (inclusive) range [100000, 165535], belonging to the actual GID 1000. This range can be mapped to the GIDs [0, 65535] in virtiofsd’s user namespace (i.e. as seen in the guest) via --gid-map=:0:100000:65536:. Alternatively, you can simply map your own GID to a single GID in the namespace: For example, --gid-map=:0:1000:1: would map GID 1000 to root’s GID in the namespace (and thus the guest).
--translate-uid=guest:<guest base UID>:<host base UID>:<count>
--translate-uid=host:<host base UID>:<guest base UID>:<count>
--translate-uid=squash-guest:<guest base UID>:<host UID>:<count>
--translate-uid=squash-host:<host base UID>:<guest UID>:<count>
--translate-uid=forbid-guest:<guest base UID>:<count>
--translate-uid=map:<guest base UID>:<host base UID>:<count>
Set up a map for virtiofsd to internally translate between host and guest UIDs. As opposed to --uid-map
, this option
does not require a user namespace, and may freely be used regardless of whether virtiofsd runs as root or not.
Mapping from guest UIDs to host UIDs is independent from the reverse, i.e. setting up a guest or squash-guest mapping only instructs virtiofsd to follow this guest-to-host mapping, it does not imply any potentially corresponding host-to-guest mapping. The only exception is the prefix-less form, which sets up a bidirectional mapping.
guest:<guest base UID>:<host base UID>:<count>
: Maps the range [guest base UID, guest base UID + count) 1:1 to [host
base UID, host base UID + count), i.e. guest UID ↦ host base UID + (guest UID - guest base UID)
.host:<host base UID>:<guest base UID>:<count>
: Reverse of the above, i.e. maps the range [host base UID, host base
UID + count) to [guest base UID, guest base UID + count); host UID ↦ guest base UID + (host UID - host base UID)
.squash-guest:<guest base UID>:<host UID>:<count>
: Maps everything in the range [guest base UID, guest base UID +
count) to the single given host UID, i.e. guest UID ↦ host UID
.squash-host:<host base UID>:<guest UID>:<count>
: Reverse of the above, i.e. maps the range [host base UID, host base
UID + count) to the single given guest UID, i.e. host UID ↦ guest UID
.forbid-guest:<guest base UID>:<count>
: Prohibits use of guest UIDs in the given range: Returns an error to the guest
whenever it tries to use a UID in that range for a new file or assign such a UID to an existing file.map:<guest base UID>:<host base UID>:<count>
: Sets up a bidirectional 1:1 mapping between [guest base UID, guest
base UID + count) and [host base UID, host base UID + count), i.e. the same as passing both guest:<guest base UID>:<host base UID>:<count>
and host:<host base UID>:<guest base UID>:<count>
.When giving multiple mappings, their source ranges must not overlap.
Neither of --translate-uid
and --translate-gid
can be used together with --posix-acl
; translating UIDs or GIDs in
virtiofsd would break posix ACLs.
Example use case: virtiofsd runs unprivileged with UID:GID 1001:100. It cannot change its own UID/GID, so attempting to
let the guest create files with any other UID/GID combination will fail. By using --translate-uid
and
--translate-gid
, however, a mapping from guest UIDs/GIDs can be set up such that virtiofsd will create files under the
only combination that it can, which is 1001:100. For example, to allow any guest user to create a file, we can squash
everything to 1001:100, which will create all those files as 1001:100 on the host. In the guest, we may want to have
those files appear as 1000:1000, though, and all other UIDs and GIDs should be visible unchanged in the guest. That
would look like so:
virtiofsd [...] \
--translate-uid squash-guest:0:1001:4294967295 \
--translate-gid squash-guest:0:100:4294967295 \
--translate-uid host:1001:1000:1 \
--translate-gid host:100:1000:1
--translate-gid=<type>:<source base GID>:<target base GID>:<count>
Same as --translate-uid
, but for GIDs.
--migration-mode=<find-paths|file-handles>
Defines how to perform migration, i.e. how to represent the internal state to the destination instance, and how to obtain that representation. Note that (when using QEMU) QEMU version 8.2 or newer is required to use virtio-fs migration.
See doc/migration.md for a comprehensive explanation on how virtio-fs migration works, what its limitations are, and what configurations we recommend.
virtiofsd internally holds references to all inodes indexed or opened by the guest. During migration, these references need to be transferred to the destination; how that is done is determined with this switch:
find-paths: For all inodes held by the source instance, look up their paths by reading the symlinks in /proc/self/fd, transfer those paths to the destination, and let the destination instance open those paths. If any inode cannot be located this way, we fall back to iterating through the shared directory (exhaustive search) to find those paths. This allows migration without requiring special privileges, and regardless of whether source and destination use the same shared directory; but is vulnerable to third parties changing metadata in the shared directory while migration is ongoing (e.g. renaming, unlinking, removing permissions), which can potentially lead to data loss and/or corruption. In addition, the fall-back method of iterating through the shared directory is expensive in terms of I/O.
file-handles: Has the source instance generate a file handle for each inode, which is sent to
the destination and opened there. A file handle is data that uniquely identifies an inode on a
filesystem. Consequently, this migration mode requires source and destination to use the same
shared directory on the same filesystem; however, source and destination instance need not
necessarily be on the same host, if that filesystem is a network filesystem.
If the shared directory spans multiple filesystems, they must all be the same in source and
destination and have the same mount points inside of the shared directory.
Using file handles is comparatively cheap in terms of I/O, and it is resilient against inodes
being renamed or unlinked by any party while they are still in use by the guest, as long as the
virtiofsd source instance keeps running until migration is fully complete. They do however
require the destination instance to have the DAC_READ_SEARCH capability, which basically means
having to run it as root, and to pass the --modcaps=+dac_read_search
command line option to it
so it does not drop that capability at start-up.
This parameter is ignored on the destination side of migration.
--migration-on-error=<abort|guest-error>
Controls how to respond to errors during migration.
During migration, some inodes that the guest has indexed or opened may turn out not to be migrateable: Either the source instance cannot construct instructions on how the destination instance may be able to find/open some inode, or the destination instance finds itself unable to follow those instructions. In all cases, the destination instance is notified of these inodes, and then decides what to do depending on the value of this parameter:
abort: Whenever the destination instance sees any such error, it returns a hard error to the vhost-user front-end (e.g. QEMU), which aborts migration. Execution is to continue on the source VM.
guest-error: Migration is allowed to finish, but all affected inodes are marked as invalid. The guest will not be able to access any such inode, receiving only errors.
Note that this parameter is to be used purely for the destination instance; its value is ignored on the source side of migration.
--migration-verify-handles
Ensure that the migration destination opens the very same inodes as the source. This only works if source and destination are to use the same shared directory on the same filesystem.
On migration, the source instance informs the destination instance of all inodes the guest has indexed or opened, and has the destination re-open them. This switch makes the source generate a file handle for each such inode, and send it to the destination, allowing the destination to re-generate the same file handle for the inode it has opened and verify that it is equal, proving it is the same inode.
(File handles are per-filesystem unique identifiers for inodes that, besides the inode ID, also include a generation ID to protect against inode ID reuse.)
Using this option protects against external parties renaming or replacing inodes while migration is ongoing, which, without this option, can lead to data loss or corruption, so it should always be used when other processes besides virtiofsd have write access to the shared directory. However, again, it only works if both source and destination use the same shared directory; though in the case of network filesystems, this does not require them to run on the same host.
This parameter is ignored on the destination side of migration.
--migration-confirm-paths
Double-check the identity of inodes right before switching over to the destination, potentially making migration more resilient when third parties have write access to the shared directory.
When representing migrated inodes by their paths relative to the shared directory, double-check during switch-over to the destination that each path still matches the respective inode. If a path does not match, try to correct by consulting the respective symbolic link in /proc/self/fd.
Note that this option requires accessing each inode indexed or opened by the guest once during the switch-over stage of migration, when both the source and destination VM are paused, so can prolong that phase for an indeterminate amount of time.
This parameter is ignored on the destination side of migration.
Export /mnt
on vhost-user UNIX domain socket /tmp/vfsd.sock
:
host# virtiofsd --socket-path=/tmp/vfsd.sock --shared-dir /mnt \
--announce-submounts --inode-file-handles=mandatory &
host# qemu-system \
-blockdev file,node-name=hdd,filename=<your image> \
-device virtio-blk,drive=hdd \
-chardev socket,id=char0,path=/tmp/vfsd.sock \
-device vhost-user-fs-pci,queue-size=1024,chardev=char0,tag=myfs \
-object memory-backend-memfd,id=mem,size=4G,share=on \
-numa node,memdev=mem \
-accel kvm -m 4G
guest# mount -t virtiofs myfs /mnt
See FAQ for adding virtiofs config to an existing qemu command-line.
When run without root, virtiofsd requires a user namespace (see user_namespaces(7)
)
to be able to switch between arbitrary user/group IDs within the guest.
virtiofsd will fail in a user namespace where UIDs/GIDs have not been mapped
(i.e., uid_map
and gid_map
files have not been written).
There are many options to run virtiofsd inside a user namespace.
For instance:
Let's assume the invoking UID and GID is 1000 and the content of both /etc/subuid
and /etc/subgid
are:
1000:100000:65536
Using podman-unshare(1)
the user namespace will be configured so that the invoking user's UID
and primary GID (i.e., 1000) appear to be UID 0 and GID 0, respectively.
Any ranges which match that user and group in /etc/subuid
and /etc/subgid
are also
mapped in as themselves with the help of the newuidmap(1)
and newgidmap(1)
helpers:
host$ podman unshare -- virtiofsd --socket-path=/tmp/vfsd.sock --shared-dir /mnt \
--announce-submounts --sandbox chroot &
Using lxc-usernsexec(1)
, we could leave the invoking user outside the mapping, having
the root user inside the user namespace mapped to the user and group 100000:
host$ lxc-usernsexec -m b:0:100000:65536 -- virtiofsd --socket-path=/tmp/vfsd.sock \
--shared-dir /mnt --announce-submounts --sandbox chroot &
In order to have the same behavior as podman-unshare(1)
, we need to run
host$ lxc-usernsexec -m b:0:1000:1 -m b:1:100000:65536 -- virtiofsd --socket-path=/tmp/vfsd.sock \
--shared-dir /mnt --announce-submounts --sandbox chroot &
We could also select --sandbox none
instead of --sandbox chroot
.
Within the guest, it is not possible to create block or char device nodes in the shared directory.
virtiofsd can't use file handles (--inode-file-handles
requires CAP_DAC_READ_SEARCH
),
so a large number of file descriptors is required.
Additionally, on NFS, not using file handles may result in a hidden file lingering after some file is deleted
(see NFS FAQ, Section D2: "What is a "silly rename"?").
virtiofsd will not be able to increase RLIMIT_NOFILE
.
--readonly
switch to prevent write accesses
from the guest, for instance, exporting share
virtiofsd --shared-dir share --readonly ...
Or export a read-only mount point:
mkdir ro-share
mount -o bind,ro share ro-share
virtiofsd --shared-dir ro-share ...
share0
, share1
:mkdir -p share/{sh0,sh1}
mount -o bind share0 share/sh0
mount -o bind share1 share/sh1
virtiofsd --announce-submounts --shared-dir share ...
Note the use of --announce-submounts
to prevent data loss/corruption.
How to add virtiofs devices to an existing qemu command-line:
If -object memory-backend-memfd,id=mem
and either -numa node,memdev=mem
or a memory-backend=mem
property in the -machine
option
have not already been added to the command, add them.
If a different memory backend is already configured then it should be changed
to memory-backend-memfd
.
-object memory-backend-memfd
must have the option share=on
and size=
must match the memory size defined by -m
.
For each virtiofs device mount add a
-chardev socket,id=${MATCHING_ID},path=${VIRTIOFSD_SOCKET_PATH}
and
-device vhost-user-fs-pci,queue-size=1024,chardev=${MATCHING_ID},tag=${VIRTIOFS_TAG}
substituting appropriate values for the shell-style variables.
One can enable support for SELinux by running virtiofsd with option "--security-label". But this will try to save guest's security context in xattr security.selinux on host and it might fail if host's SELinux policy does not permit virtiofsd to do this operation.
Hence, it is recommended to remap guest's "security.selinux" xattr to say "trusted.virtiofs.security.selinux" on host. Add following option to command line.
"--xattrmap=:map:security.selinux:trusted.virtiofs.:"
This will make sure that guest and host's SELinux xattrs on same file remain separate and not interfere with each other. And will allow both host and guest to implement their own separate SELinux policies.
Setting trusted xattr on host requires CAP_SYS_ADMIN. So one will need add this capability to daemon. Add following option to command line.
"--modcaps=+sys_admin"
trusted xattrs are not namespaced. So virtiofsd needs to have CAP_SYS_ADMIN in init_user_ns. IOW, one should not be using user namespaces and virtiofsd should run with CAP_SYS_ADMIN.
Giving CAP_SYS_ADMIN increases the risk on system. Now virtiofsd is more powerful and if gets compromised, it can do lot of damage to host system. So keep this trade-off in my mind while making a decision.