Backup utility for libvirt
, using the latest changed block tracking features.
Create online, thin provisioned full and incremental or differential backups
of your kvm/qemu
virtual machines.
- About
- Prerequisites
- Installation
- Backup Format
- Backup Operation
- Supported disk formats / raw disks
- Backup Examples
- Restore examples
- Single file restore and instant recovery
- Extents
- Transient virtual machines: checkpoint persistency
- Hypervisors
- Authentication
- FAQ
- The thin provisioned backups are bigger than the original qcow images
- Backup fails with "Cannot store dirty bitmaps in qcow2 v2 files"
- Backup fails with "Timed out during operation: cannot acquire state change lock"
- Backup fails with "Failed to bind socket to /var/tmp/virtnbdbackup.XX: Permission denied"
- High memory usage during backup
- Test your backups!
- Links
- Contributing
Existing backup solutions or scripts for libvirt/kvm
usually depend on the
external snapshot feature to create backups, sometimes even require to
shutdown or pause the virtual machine.
Recent additions to both the libvirt
and qemu
projects have introduced new
capabilities that allow to create online (full and incremental) backups, by
using so called dirty bitmaps
(or changed block tracking).
virtnbdbackup
uses these features to create online full and incremental
or differential backups.
virtnbdrestore
can be used to re-construct the complete image from the
thin provisioned backups.
virtnbdmap
can be used to map an thin provisioned backup image into a
block device on-the-fly, for easy single file restore or even instant
boot from an backup image.
For backing up standalone qemu virtual machines not managed by libvirt, see this project: qmpbackup
Obviously you require a libvirt/qemu version that supports the incremental backup features. Following, you will find a short overview which distributions ship the required versions.
If you are using Debian Bullseye or Ubuntu 20.x, the included libvirt version already supports the features required, but does not enable them by default!
To enable the incremental backup feature on these libvirt versions
change the virtual machine config using virsh edit <vm>
like so: (the
first line must be changed, too!):
<domain type='kvm' id='1' xmlns:qemu='http://libvirt.org/schemas/domain/qemu/1.0'>
[..]
<qemu:capabilities>
<qemu:add capability='incremental-backup'/>
</qemu:capabilities>
[..]
</domain>
Note
:
You must restart the virtual machine after enabling the feature!
Up to RHEL/Centos8/Almalinux 8.5, libvirt packages from the advanced virtualization stream support all required features. To install libvirt from the stream use:
yum install centos-release-advanced-virtualization
yum makecache
yum module install virt
and enable the feature by adjusting the virtual machine config.
As of RHEL 8.6, the advanced virtualization stream has been deprecated, and all components supporting the new feature are included in the virt:rhel module, the feature is enabled by default. (Details)
- python libvirt module version >= 6.0.0 (yum install python3-libvirt)
- python libnbd bindings (https://github.com/libguestfs/libnbd) version >=
1.5.5
(yum install python3-libnbd) - The virtual machine should use qcow version 3 images to support the full feature set.
There are several ways to install the utility, below you will find an short description for each of them.
Note
:
Please consider to check past issues related to installation if you face any troubles before opening a new issue.
pip3 install -r requirements.txt
python3 setup.py install
Note
:
Do not install the "nbd" package available on PyPI, it does not provide the required nbd bindings (unfortunately has the same name). You have to additionally install the provided python3-libnbd packages by your distribution, or compile the libnbd bindings by yourself.
To create an RPM package from source suitable for installation:
To build the rpm package from source:
yum install epel-release # required for tqdm on centos 8
yum makecache
yum install rpm-build
python3.6 setup.py bdist_rpm --python /usr/bin/python3.6
yum install dist/virtnbdbackup-<version>-.noarch.rpm
To create a Debian package (Debian bullseye required) use:
sudo apt-get install python3-all python3-stdeb dh-python python3-libnbd python3-tqdm python3-lz4 python3-lxml
python3 setup.py --command-packages=stdeb.command bdist_deb
You can also use existing vagrant scripts to build the packages.
For setup within an virtualenv see venv scripts.
You can build an docker image using the existing Dockerfile
Currently, there are two output formats implemented:
stream
: the resulting backup image is saved in a streamlined format, where the backup file consists of metadata about offsets and lengths of zeroed or allocated contents of the virtual machines disk. This is the default. The resulting backup image is thin provisioned.raw
: The resulting backup image will be a full provisioned raw image, this should mostly be used for debugging any problems with the extent handler, it won't work with incremental backups.
Following backup modes can be used:
-
copy
: Full, thin provisioned backup of the virtual machine disks, no checkpoint is created for further incremental backups, existing checkpoints will be left untouched. This is the default mode and works with qcow images not supporting persistent bitmaps. -
full
: Full, thin provisioned backup of the virtual machine, a new checkpoint namedvirtnbdbackup
will be created, all existent checkpoints from prior backups matching this name will be removed: a new backup chain is created. -
inc
: Perform incremental backup, based on the last full or incremental backup. A checkpoint for each incremental backup is created and saved. -
diff
: Perform differential backup: saves the current delta to the last incremental or full backup. -
auto
: If the target folder is empty, attempt to execute full backup, otherwise switch to backup mode incremental: allows rotation of backup into monthly folders.
All required information for restore is stored to the same directory, including the latest virtual machine configuration, checkpoint information, disk data and logfiles.
The target directory must be rotated if a new backup set is created.
If the virtual domain is active and running, a backup job operation via
libvirt api
is started, which in turn initializes a new nbd server backend
listening on a local unix socket. This nbd backend provides consistent access
to the virtual machines, disk data and dirty blocks. After the backup process
finishes, the job is stopped and the nbd server quits operation.
Note
:
If the virtual domain is not in running state (powered off)
virtnbdbackup
supportscopy
andinc/diff
backup modes. Incremental and differential backups will then save the changed blocks since last created checkpoint. As no new checkpoints can be defined for offline domains, the Backup modefull
is changed to modecopy
.
It is possible to backup multiple virtual machines on the same host system at the same time, using separate calls to the application with a different target directory to store the data.
libvirt/qemu
supports dirty bitmaps, required for incremental backups only
with qcow(v3) based disk images. If you are using older image versions, you can
only create copy
backups, or consider converting the images to a newer
format using qemu-img
.
By default virtnbdbackup
will exclude all disks with format raw
. This
behavior can be changed if option --raw
is specified, raw disks will then be
included during a full
backup. This of course means that no thin provisioned
backup is created for these particular disks.
During restore, these files can be copied "as is" from the backup folder and
must not be processed using virtnbdrestore
.
- Start full backup of domain
vm1
, save data to/tmp/backupset
:
virtnbdbackup -d vm1 -l full -o /tmp/backupset
- Start incremental backup for domain
vm1
, backup only changed blocks to the last full backup:
virtnbdbackup -d vm1 -l inc -o /tmp/backupset
The resulting directory will contain all information for restoring the virtual machine, including logfiles that can be used for analyzing backup issues:
/tmp/backupset/
├── backup.full.05102021161752.log
├── backup.inc.05102021161813.log
├── backup.inc.05102021161814.log
├── checkpoints
│  ├── virtnbdbackup.0.xml
│  ├── virtnbdbackup.1.xml
│  └── virtnbdbackup.2.xml
├── sda.full.data
├── sda.inc.virtnbdbackup.1.data
├── sda.inc.virtnbdbackup.2.data
├── vm1.cpt
├── vmconfig.virtnbdbackup.0.xml
├── vmconfig.virtnbdbackup.1.xml
└── vmconfig.virtnbdbackup.2.xml
During backup virtnbdbackup
attempts to freeze all file systems within the
domain using the qemu guest agent filesystem freeze and thaw functions. In
case no qemu agent is installed or filesystem freeze fails, a warning is issued
during backup:
WARNING [..] Guest agent is not responding: QEMU guest agent is not connected
In case you receive this warning, check if the qemu agent is installed and running in the domain.
It is also possible to specify one or multiple mountpoints used within the virtual machine to freeze only specific filesystems, like so:
virtnbdbackup -d vm1 -l inc -o /tmp/backupset -F /mnt,/var
this way only the underlying filesystems on /mnt and /var are frozen and thawed.
Note:
It is highly recommended to have an qemu agent running in the virtual domain to ensure file system consistency during backup!
With backup mode auto
it is possible to have a monthly rotation/retention. If
the target folder is empty, backup mode auto will create an full backup. On the
following executions, it will automatically switch to backup mode incremental,
if the target folder already includes an full backup. Example:
virtnbdbackup -d vm1 -l auto -o /tmp/2022-06 -> creates full backup
virtnbdbackup -d vm1 -l auto -o /tmp/2022-06 -> creates inc backup
virtnbdbackup -d vm1 -l auto -o /tmp/2022-06 -> creates inc backup
virtnbdbackup -d vm1 -l auto -o /tmp/2022-07 -> creates full backup
virtnbdbackup -d vm1 -l auto -o /tmp/2022-07 -> creates inc backup
Option -x
can be used to exclude certain disks from the backup. The name of
the disk to be excluded must match the disks target device name as configured
in the domains xml definition, for example:
virtnbdbackup -d vm1 -l full -o /tmp/backupset -x sda
Special devices such as cdrom/floppy
or direct attached luns
are excluded
by default, as they are not supported by the changed block tracking layer.
It is also possible to only backup specific disks using the include option
(--include
, or -i
):
virtnbdbackup -d vm1 -l full -o /tmp/backupset -i sdf
Sometimes it can be useful to estimate the data size prior to executing the
next incremental
or differential
backup. This can be done by using the
option -p
which will query the virtual machine checkpoint information for the
current size:
virtnbdbackup -d vm1 -l inc -o /tmp/backupset -p
[..]
[..] INFO virtnbdbackup - handleCheckpoints [MainThread]: Using checkpoint name: [virtnbdbackup.1].
[..] INFO virtnbdbackup - main [MainThread]: Estimated checkpoint backup size: [24248320] Bytes
Note:
Not all libvirt versions support the flag required to read the checkpoint size. If the estimated checkpoint size is allways 0, your libvirt version might miss the required features.
If an incremental
or differential
backup is attempted and the virtual machine
is active, it is possible to specify an threshold for executing the backup
using the --threshold
option. The backup will then only be executed if the
amount of data changed meets the specified threshold (in bytes):
virtnbdbackup -d vm1 -l inc -o /tmp/backupset --threshold 3311264
[..]
[..] INFO virtnbdbackup - handleCheckpoints [MainThread]: Using checkpoint name: [virtnbdbackup.1].
[..] ]virtnbdbackup - main [MainThread]: Backup size [3211264] does not meet required threshold [3311264], skipping backup.
If virtnbdbackup
saves data to a regular target directory, it starts one
thread for each disk it detects to speed up the backup operation.
This behavior can be changed using the --worker
option to define an amount of
threads to be used for backup. Depending on how many disks your virtual machine
has attached, it might make sense to try a different amount of workers to see
which amount your hardware can handle best.
If standard output (-
) is defined as backup target, the amount of workers is
allways limited to 1, to ensure a valid Zip file format.
It is possible to enable compression for the stream
format via lz4
algorithm by using the --compress
option. The saved data is compressed inline
and the saveset file is appended with compression trailer including information
about the compressed block offsets.
During the restore, virtnbdrestore
will automatically detect such compressed
backup streams and attempts to decompress saved blocks accordingly.
Using compression will come with some CPU overhead, both lz4 checksums for block and original data are enabled.
It is also possible to backup remote libvirt systems. The most convinient way is to use ssh for initiating the libvirt connection (key authentication mandatory).
If the virtual machine has additional files configured, as described in Kernel/initrd and additional files, these files will be copied from the remote system via SSH(SFTP).
In order to backup virtual machines from a remote host, you must specify an libvirt URI to the remote system.
The following example saves the virtual machine src
from the remote libvirt
host hypervisor
to the local directory /tmp/backupset
, it uses the root
user for both the libvirt and ssh authentication:
virtnbdbackup -U qemu+ssh://root@hypervisor/system --ssh-user root -d src -o /tmp/backupset
See also: Authentication
By default disk data received from a remote system will be transferred via
regular NBD protocol. You can enable TLS for this connection, using the --tls
option. Before beeing able to use TLS, you must configure the required
certificates on both sides.
See the following documentation by the libvirt project for detailed instructions how setup:
https://wiki.libvirt.org/page/TLSCreateCACert
Note:
You should have installed at least version 1.12.6 of the libnbd library which makes the transfer via NBDS more stable full background
In case you want to use a dedicated network for the NBD transfer, you
can specify an specific IP address to bind the remote NBD service to
via --nbd-ip
option.
If the output target points to standard out (-
), virtnbdbackup
puts the
resulting backup data into an uncompessed zip archive.
A such, it is possible to transfer the backup data to different hosts, or pipe it to other programs.
However, keep in mind that in case you want to perform incremental backups, you must keep the checkpoint files on the host you are executing the backup utility from, until you create another full backup.
If output is set to standard out, virtnbdbackup
will create the required
checkpoint files in the directory it is executed from.
Here is an example:
# mkdir backup-weekly; cd backup-weekly
# virtnbdbackup -d vm1 -l full -o - | ssh root@remotehost 'cat > backup-full.zip'
# [..]
# INFO outputhelper - __init__: Writing zip file stream to stdout
# [..]
# INFO virtnbdbackup - main: Finished
# INFO virtnbdbackup - main: Adding vm config to zipfile
# [..]
Any subsequent incremental backup operations must be called from within this directory:
# cd backup-weekly
# virtnbdbackup -d vm1 -l inc -o - | ssh root@remotehost 'cat > backup-inc1.zip'
[..]
You may consider adding the created checkpoint files to some VCS system, like git, to have some kind of central backup history tracking.
During restore unzip the data from both zip files into a single directory:
(use virtnbdrestore
to reconstruct the virtual machine images):
# unzip -o -d restoredata backup-full.zip
# unzip -o -d restoredata backup-inc1.zip
If an domain has configured custom kernel, initrd, loader or nvram images (usually the case if the domain boots from OVM UEFI BIOS), these files will be saved to the backup folder aswell.
For restoring, virtnbdrestore
can be used. It reconstructs the streamed
backup format back into a usable qemu qcow image.
The restore process will create a qcow image with the original virtual size.
In a second step, the qcow image is then mapped to a ndb server instance where
all data blocks are sent to and are applied accordingly. The resulting image
can be mounted (using guestmount
) or attached to a running virtual machine in
order to recover required files.
As a first start, the dump
parameter can be used to dump the saveset
information of an existing backupset:
virtnbdrestore -i /tmp/backupset/ -o dump
INFO:root:Dumping saveset meta information
{'checkpointName': 'virtnbdbackup',
'dataSize': 704643072,
'date': '2020-11-15T20:50:36.448938',
'diskName': 'sda',
'incremental': False,
'parentCheckpoint': False,
'streamVersion': 1,
'virtualSize': 32212254720}
[..]
The output includes information about the thick and thin provisioned disk space that is required for recovery, date of the backup and checkpoint chain.
To restore all disks within the backupset into a usable qcow image use command:
virtnbdrestore -i /tmp/backupset/ -o /tmp/restore
All incremental backups found will be applied to the target images
in the output directory /tmp/restore
Note
:
The restore utility will copy the latest virtual machine config to the target directory, but wont alter its contents. You have to adjust the config file for the new pathes and/or excluded disks to be able to define and run it.
A single disk can be restored by using the option -d
, the disk name has
to match the virtual disks target name, for example:
virtnbdrestore -i /tmp/backupset/ -o /tmp/restore -d sda
Option --until
allows to perform a point in time restore up to the desired
checkpoint. The checkpoint name has to be specified as reported by the
dump output (field checkpointName
), for example:
virtnbdrestore -i /tmp/backupset/ -o /tmp/restore --until virtnbdbackup.2
It is also possible to specify the source data files specifically used for the
rollback via --sequence
option, but beware: you must be sure the sequence you
apply has the right order, otherwise the restored image might be errnous,
example:
virtnbdrestore -i /tmp/backupset/ -o /tmp/restore --sequence vdb.full.data,vdb.inc.virtnbdbackup.1.data
Option -c
can be used to adjust the virtual machine configuration during
restore accordingly, the following changes are done:
- UUID of the virtual machine is removed from the config file
- Name of the virtual machine is prefixed with "restore_" (use option
--name
to specify desired vm name) - The disk pathes to the virtual machine are changed to the new target directory.
- If virtual machine was operating on snapshots/backing store images, the references to the configured backing stores will be removed.
- Raw devices are removed from VM config if
--raw
is not specified, aswell as floppy or cdrom devices (which arent part of the backup).
Note:
If missing, Kernel, UEFI or NVRAM files are restored to their original location as set in the virtual machine configuration.
A restored virtual machine can then been defined and started right from the
restored directory (or use option -D
to define automatically):
virtnbdrestore -c -i /tmp/backupset/ -o /tmp/restore
[..]
[..] INFO virtnbdrestore - restoreConfig [MainThread]: Adjusted config placed in: [/tmp/restore/vmconfig.virtnbdbackup.0.xml]
[..] INFO virtnbdrestore - restoreConfig [MainThread]: Use 'virsh define /tmp/restore/vmconfig.virtnbdbackup.0.xml' to define VM
Restoring to a remote host is possible too, same options as during backup
apply. The following example will restore the virtual machine from the local
directory /tmp/backupset
to the remote system "hypervisor", alter its
configuration and register the virtual machine:
virtnbdrestore -U qemu+ssh://root@hypervisor/system --ssh-user root -cD -i /tmp/backupset -o /remote/target
The virtnbdmap
utility can be used to map uncompressed backup images from the
stream format into an accessible block device on the fly. This way, you can
restore single files or even boot from an existing backup image without having
to restore the complete dataset.
The utility requires nbdkit with the python plugin
to be installed on the
system along with required qemu tools (qemu-nbd
) and an loaded nbd kernel
module. It must be executed with superuser (root) rights or via sudo.
The following example maps an existing backup image to the network block
device /dev/nbd0
:
# modprobe nbd max_partitions=15
# virtnbdmap -f /backup/sda.full.data
[..] INFO virtnbdmap - <module> [MainThread]: Done mapping backup image to [/dev/nbd0]
[..] INFO virtnbdmap - <module> [MainThread]: Press CTRL+C to disconnect
While the process is running, you can access the backup image like a regular block device:
fdisk -l /dev/nbd0
Disk /dev/nbd0: 2 GiB, 2147483648 bytes, 4194304 sectors
You can also create an mapped "point in time" recovery image by passing a sequence of full and incremental backups as parameter. The changes from the incremental backups will then be replayed to the block device on the fly and the image will represent the latest state:
virtnbdmap -f /backup/sda.full.data,/backup/sda.inc.virtnbdbackup.1.data,/backup/sda.inc.virtnbdbackup.2.data
[..]
[..] INFO virtnbdmap - main [MainThread]: Need to replay incremental backups
[..] INFO virtnbdmap - main [MainThread]: Replaying offset 420 from /backup/sda.inc.virtnbdbackup.1.data
[..] INFO virtnbdmap - main [MainThread]: Replaying offset 131534 from /backup/sda.inc.virtnbdbackup.1.data
[..]
[..] INFO virtnbdmap - main [MainThread]: Replaying offset 33534 from /backup/sda.inc.virtnbdbackup.2.data
[..] INFO virtnbdmap - <module> [MainThread]: Done mapping backup image to [/dev/nbd0]
[..] INFO virtnbdmap - <module> [MainThread]: Press CTRL+C to disconnect
[..]
The original image will be left untouched as nbdkits copy on write filter is used to replay the changes.
Further you can create an overlay image via qemu-img
and boot from it right
away (or boot directly from the /dev/nbd0 device).
qemu-img create -b /dev/nbd0 -f qcow2 bootme.qcow2
qemu-system-x86_64 -enable-kvm -m 2000 -hda bootme.qcow2
To remove the mappings, stop the utility via "CTRL-C"
Note
:
If the virtual machine includes volume groups, the system will attempt to set them online as you create the mapping, because the copy on write device is writable by default. If your host system is using the same volume group names this could lead to issues (check
dmesg
orjournalctl
then). In case the volume groups are online, it is recommended to change them to offline just before you remove the mapping, to free all references to the mapped nbd device (vgchange -a n <vg_name>
)
Note
:
If you map the image device with the
--readonly
option you may need to pass certain options to the mount command (-o norecovery,ro) in order to be able to mount the filesystems. This may also be the case if no qemu agent was installed within the virtual machine during backup.
In order to save only used data from the images, dirty blocks are queried from
the NBD server. The behavior can be changed by using the option -q
to use
common qemu tools (nbdinfo). By default virtnbdbackup
uses a custom
implemented extent handler.
In case virtual machines are started in transient environments, such as using
cluster solutions like pacemaker
situations can appear where the checkpoints
for the virtual machine defined by libvirt are not in sync with the bitmap
information in the qcow files.
In case libvirt creates a checkpoint, the checkpoint information is stored in two places:
- var/lib/libvirt/qemu/checkpoint/<domain_name>
- In the bitmap file of the virtual machines qcow image.
Depending on the cluster solution, in case virtual machines are destroyed
on host A and are re-defined on host B, libvirt loses the information about
those checkpoints. Unfortunately libvirtd
scans the checkpoint only once
during startup.
This can result in a situation, where the bitmap is still defined in the qcow image, but libvirt doesn't know about the checkpoint, backup then fails with:
Unable to execute QEMU command 'transaction': Bitmap already exists
By default virtnbdbackup
attempts to store the checkpoint information in the
default backup directory, in situations where it detects a checkpoint is
missing, it attempts to redefine them from the prior backups.
In order to store the checkpoint information at some central place the option
--checkpointdir
can be used, this allows having persistent checkpoints
stored across multiple nodes:
As example:
- Create backup on host A, store checkpoints in a shared directory between
hosts in
/mnt/shared/vm1
:
virtnbdbackup -d vm1 -l full -o /tmp/backup_hosta --checkpointdir /mnt/shared/vm1
- After backup, the virtual machine is relocated to host B and loses its information about checkpoints and bitmaps, thus, the next full backup usually fails with:
virtnbdbackup -d vm1 -l full -o /tmp/backup_hostb
[..]
unable to execute QEMU command 'transaction': Bitmap already exists: virtnbdbackup.0
- Now pass the checkpoint dir and files written from host A, and virtnbdbackup will redefine missing checkpoints and execute a new full backup. As the new full backup removes all prior checkpoints the bitmap information is in sync after this operation and backup succeeds:
virtnbdbackup -d vm1 -l full -o /tmp/backup_hostb --checkpointdir /mnt/shared/vm1
[..]
redefineCheckpoints: Redefine missing checkpoint virtnbdbackup.0
[..]
See also: abbbi#10
virtnbdbackup
uses the lowest layer on top of libvirt to allow its
functionality, you can also use it with more advanced hypervisors solutions
such as ovirt, RHEV or OpenNebula, but please bear in
mind that it was not developed to target all of those solutions specifically!
If you are using the ovirt node based hypervisor hosts you should consider creating a virtualenv via the venv scripts and transferring it to the node system.
On regular centos/alma/rhel based nodes, installation via RPM package should be preferred. The incremental backup functionality can be enabled via ovirt management interface.
Usually ovirt restricts access to the libvirt daemon via different
authentication methods. Use the -U
parameter in order to specify an
authentication file:
virtnbdbackup -U qemu:///system?authfile=/etc/ovirt-hosted-engine/virsh_auth.conf -d src -o /tmp/backupset
Note:
virtnbdrestore
has not been adopted to cope with the ovirt specific domain xml format, so redefining and virtual machine on the node might not work.
See past issues
Both virtnbdbackup
and virtnbdrestore
commands support authenticating
against libvirtd with the usual URIs. Consider using the following options:
-U
: Specify an arbitary connection URI to use against libvirt
--user
: Username to use for the specified connection URI
--password
: Password to use for the specified connection URI.
It is also possible to specifiy the credentials stored as authentication file
like it would be possible using the virsh -c
option:
-U qemu:///system?authfile=/etc/virsh_auth.conf ..
Note:
The default connection URI used is
qemu:///system
which is usually the case if virtual machines operate as root user. Use theqemu:///session
URI to backup virtual machines as regular user.
Virtual machines using the qcow format do compress data. During backup, the image contents are exposed as NDB device which is a RAW device. The backup data will be at least as big as the used data within the virtual machine.
You can use the --compress
option or other tools to compress the backup
images in order to save storage space or consider using a deduplication capable
target file system.
If the backup fails with error:
ERROR [..] internal error: unable to execute QEMU command dirty bitmaps in qcow2 v2 files
consider migrating your qcow files to version 3 format. QEMU qcow image version
2 does not support storing advanced bitmap information, as such only backup
mode copy
is supported.
If backups fail with error:
ERROR [..] Timed out during operation: cannot acquire state change lock (held by monitor=remoteDispatchDomainBackupBegin)
there is still some block jobs operation active on the running domain, for
example a live migration or another backup job. It may also happen that
virtnbdbackup
crashes abnormally or is forcibly killed during backup
operation, unable to stop its own backup job.
You can use option -k
to forcibly kill any running active block jobs for the
domain, but use with care. It is better to check which operation is active with
the virsh domjobinfo
command first.
virtnbdbackup -d vm2 -l copy -k -o -
[..]
INFO virtnbdbackup - main: Stopping domain jobs
The issue is most likely an active apparmor
profile that prevents the qemu
daemon from creating its socket file for the nbd server. Try to disable
apparmor using the aa-teardown command for the current session you are
executing a backup or restore. You can also add the following lines:
/var/tmp/virtnbdbackup.* rw,
/var/tmp/backup.* rw,
to the configuration files (might not exist by default):
/etc/apparmor.d/usr.lib.libvirt.virt-aa-helper
/etc/apparmor.d/local/abstractions/libvirt-qemu
/etc/apparmor.d/local/usr.sbin.libvirtd
See also: abbbi#7
libnbd python implementation has had various memory leaks in older versions which cause such problems.
For centos 8 based distributions these fixes have been backported to libnbd
1.4.0.
The fix itself was released with libnbd 1.5.2, so be sure to use at least this
version if using virtnbdbackup
on any other distribution.
See also: abbbi#8
The utility is provided "as is", i take no responsibility or warranty if you face any issues recovering your data! The only way to ensure your backups are valid and your backup plan works correctly is to repeatedly test the integrity by restoring them! If you discover any issues, please do not hesitate to open an issue.
Backup howto for Debian Bullseye: https://abbbi.github.io/debian/
Short video: https://youtu.be/dOE0iB-CEGM
Active development happens in the development branch please base your pull requests on this branch.