Discussion:
[linux-lvm] LVM snapshot: in memory COW table
Francois Blondel
2017-01-25 13:46:36 UTC
Permalink
Hi all,

We currently use LVM snapshot as a solution to backups block devices.
The snapshot is not active for a long time, only for the time during
which we need to upload the backup to the backup server.

As I could read, for example here: http://www.nikhef.nl/~dennisvd/lvmcr
ap.html, snapshots multiply the IO needs of the origin Logical Volume.
We would like to avoid that.

As our servers have enough free RAM, and the COW table of the
snapshot remains quite small, I had the project to store the snapshot
(or its cow table) in RAM.
The goal there is to avoid additional IO aimed at the disk and redirect
them to memory.
Yes, this would lead to the loss of the snapshot in case of power
failure, but this is not an issue for our use case.

I tried to use lvmcache with an in memory block device for the caches,
but I could not create a snapshot using lvmcache.

Any hints or experience on that?

Many thanks,

François Blondel
Mike Snitzer
2017-01-25 14:38:25 UTC
Permalink
On Wed, Jan 25 2017 at 8:46am -0500,
Post by Francois Blondel
Hi all,
We currently use LVM snapshot as a solution to backups block devices.
The snapshot is not active for a long time, only for the time during
which we need to upload the backup to the backup server.
As I could read, for example here: http://www.nikhef.nl/~dennisvd/lvmcrap.html
snapshots multiply the IO needs of the origin Logical Volume.
We would like to avoid that.
The old dm-snapshot target has "transient" snapshot store already.
See drivers/md/dm-snap-transient.c

Use of transient isn't going to change the fact that you'll still need N
snapshot stores; and that any write will trigger an N-way copy-out
penality when N snapshots of the origin exist. All you'd be doing with
transient is getting implicit improved performance of memory speeds.

For improved scalability you'd do well to switch to using DM
thin-provisioning's snapshots. There are some customers/users who are
already making use of it for the backend for backups (using thin_delta
to know what the differences are between thin volumes and snapshots).
Post by Francois Blondel
As our servers have enough free RAM, and the COW table of the
snapshot remains quite small, I had the project to store the snapshot
(or its cow table) in RAM.
The goal there is to avoid additional IO aimed at the disk and redirect
them to memory.
Yes, this would lead to the loss of the snapshot in case of power
failure, but this is not an issue for our use case.
Documentation/device-mapper/snapshot touches on "transient" with:

*) snapshot <origin> <COW device> <persistent?> <chunksize>

A snapshot of the <origin> block device is created. Changed chunks of
<chunksize> sectors will be stored on the <COW device>. Writes will
only go to the <COW device>. Reads will come from the <COW device> or
from <origin> for unchanged data. <COW device> will often be
smaller than the origin and if it fills up the snapshot will become
useless and be disabled, returning errors. So it is important to monitor
the amount of free space and expand the <COW device> before it fills up.

<persistent?> is P (Persistent) or N (Not persistent - will not survive
after reboot). O (Overflow) can be added as a persistent store option
to allow userspace to advertise its support for seeing "Overflow" in the
snapshot status. So supported store types are "P", "PO" and "N".

The difference between persistent and transient is with transient
snapshots less metadata must be saved on disk - they can be kept in
memory by the kernel.
Stuart Gathman
2017-01-26 00:50:45 UTC
Permalink
Post by Mike Snitzer
On Wed, Jan 25 2017 at 8:46am -0500,
Post by Francois Blondel
Hi all,
We currently use LVM snapshot as a solution to backups block devices.
The snapshot is not active for a long time, only for the time during
which we need to upload the backup to the backup server.
As I could read, for example here: http://www.nikhef.nl/~dennisvd/lvmcrap.html
snapshots multiply the IO needs of the origin Logical Volume.
We would like to avoid that.
The old dm-snapshot target has "transient" snapshot store already.
See drivers/md/dm-snap-transient.c
Use of transient isn't going to change the fact that you'll still need N
snapshot stores; and that any write will trigger an N-way copy-out
penality when N snapshots of the origin exist. All you'd be doing with
transient is getting implicit improved performance of memory speeds.
The transient snapshots seem to only store metadata in memory. I think
what OP is proposing is to allocate, say, 2G of ram for the snapshot,
instead of allocating a 2G snapshot LV on disk. The snapshot would
simply disappear if the system crashes or reboots.

You could kludge this up now by adding a ramdisk PV to the volume group,
somehow avoid allocating normal LVs on the ramdisk PV, and force the
selected snapshot to be allocated on the ramdisk PV. Adding the
ramdisk to the VG immediately before allocating the snapshot on it (and
using up all the extents) would prevent normal LVs from getting
allocated on it.

The drawback to the above kludge is that on crash or reboot, the ramdisk
PV will be gone - requiring an extra step to remove it from the VG and
undo the snapshot remapping.

A special "use me for snapshots only" flag for PVs would be one way to
start implementing it.

Loading...