lvmvdo(top10.html) - phpMan

LVMVDO(7)                                                            LVMVDO(7)
NAME
       lvmvdo -- Support for Virtual Data Optimizer in LVM
DESCRIPTION
       VDO  is  software  that provides inline block-level deduplication, com-
       pression, and thin provisioning capabilities for primary storage.
       Deduplication is a technique for reducing the  consumption  of  storage
       resources  by eliminating multiple copies of duplicate blocks. Compres-
       sion takes the  individual  unique  blocks  and  shrinks  them.   These
       reduced  blocks  are  then  efficiently  packed  together into physical
       blocks. Thin provisioning manages the mapping from logical blocks  pre-
       sented  by  VDO  to where the data has actually been physically stored,
       and also eliminates any blocks of all zeroes.
       With deduplication, instead of writing the same data  more  than  once,
       VDO  detects  and  records  each  duplicate block as a reference to the
       original block. VDO maintains a mapping from  Logical  Block  Addresses
       (LBA) (used by the storage layer above VDO) to physical block addresses
       (used by the storage layer under VDO).  After  deduplication,  multiple
       logical  block  addresses  may  be  mapped  to  the same physical block
       address; these are called shared blocks and  are  reference-counted  by
       the software.
       With  compression,  VDO  compresses  multiple blocks (or shared blocks)
       with the fast LZ4 algorithm, and bins them together where  possible  so
       that multiple compressed blocks fit within a 4 KB block on the underly-
       ing storage. Mapping from LBA is to a physical block address and  index
       within  it  for  the desired compressed data. All compressed blocks are
       individually reference counted for correctness.
       Block sharing and block compression are invisible to applications using
       the  storage, which read and write blocks as they would if VDO were not
       present. When a shared block is overwritten, a new  physical  block  is
       allocated  for  storing the new block data to ensure that other logical
       block addresses that are mapped to the shared physical  block  are  not
       modified.
       To  use  VDO  with lvm(8), you must install the standard VDO user-space
       tools vdoformat(8) and the currently  non-standard  kernel  VDO  module
       "kvdo".
       The  "kvdo" module implements fine-grained storage virtualization, thin
       provisioning, block sharing, and compression.  The  "uds"  module  pro-
       vides  memory-efficient  duplicate identification. The user-space tools
       include vdostats(8) for extracting statistics from VDO volumes.
VDO TERMS
       VDODataLV
              VDO data LV
              A large hidden LV with the _vdata suffix. It is created in a VG
              used by the VDO kernel target to store  all  data  and  metadata
              blocks.
       VDOPoolLV
              VDO pool LV
              A  pool  for  virtual  VDOLV(s), which are the size of used VDO-
              DataLV.
              Only a single VDOLV is currently supported.
       VDOLV
              VDO LV
              Created from VDOPoolLV.
              Appears blank after creation.
VDO USAGE
       The primary methods for using VDO with lvm2:
   1. Create a VDOPoolLV and a VDOLV
       Create a VDOPoolLV that will hold VDO data, and a  virtual  size  VDOLV
       that the user can use. If you do not specify the virtual size, then the
       VDOLV is created with the maximum size that always fits into data  vol-
       ume  even  if  no  deduplication or compression can happen (i.e. it can
       hold the incompressible content of /dev/urandom).  If you do not  spec-
       ify  the  name  of  VDOPoolLV, it is taken from the sequence of vpool0,
       vpool1 ...
       Note: The performance of TRIM/Discard operations is slow for large vol-
       umes  of  VDO type. Please try to avoid sending discard requests unless
       necessary because it might take considerable amount of time  to  finish
       the discard operation.
       lvcreate --type vdo -n VDOLV -L DataSize -V LargeVirtualSize VG/VDOPoolLV
       lvcreate --vdo -L DataSize VG
       Example
       # lvcreate --type vdo -n vdo0 -L 10G -V 100G vg/vdopool0
       # mkfs.ext4 -E nodiscard /dev/vg/vdo0
   2. Convert an existing LV into VDOPoolLV
       Convert  an already created or existing LV into a VDOPoolLV, which is a
       volume that can hold data and metadata.  You will be prompted  to  con-
       firm  such  conversion  because it IRREVERSIBLY DESTROYS the content of
       such volume and the volume is immediately formatted by vdoformat(8)  as
       a  VDO  pool data volume. You can specify the virtual size of the VDOLV
       associated with this VDOPoolLV.  If you  do  not  specify  the  virtual
       size, it will be set to the maximum size that can keep 100% incompress-
       ible data there.
       lvconvert --type vdo-pool -n VDOLV -V VirtualSize VG/VDOPoolLV
       lvconvert --vdopool VG/VDOPoolLV
       Example
       # lvconvert --type vdo-pool -n vdo0 -V10G vg/ExistingLV
   3. Change the compression and deduplication of a VDOPoolLV
       Disable or enable the compression and deduplication for VDOPoolLV  (the
       volume that maintains all VDO LV(s) associated with it).
       lvchange --compression y|n --deduplication y|n VG/VDOPoolLV
       Example
       # lvchange --compression n  vg/vdopool0
       # lvchange --deduplication y vg/vdopool1
   4. Change the default settings used for creating a VDOPoolLV
       VDO  allows  to  set a large variety of options. Lots of these settings
       can be specified in lvm.conf or profile settings.  You  can  prepare  a
       number of different profiles in the /etc/lvm/profile directory and just
       specify the profile file name.  Check the output  of  lvmconfig  --type
       default --withcomments for a detailed description of all individual VDO
       settings.
       Example
       # cat <<EOF > /etc/lvm/profile/vdo_create.profile
       allocation {
              vdo_use_compression=1
              vdo_use_deduplication=1
              vdo_use_metadata_hints=1
              vdo_minimum_io_size=4096
              vdo_block_map_cache_size_mb=128
              vdo_block_map_period=16380
              vdo_use_sparse_index=0
              vdo_index_memory_size_mb=256
              vdo_slab_size_mb=2048
              vdo_ack_threads=1
              vdo_bio_threads=1
              vdo_bio_rotation=64
              vdo_cpu_threads=2
              vdo_hash_zone_threads=1
              vdo_logical_threads=1
              vdo_physical_threads=1
              vdo_write_policy="auto"
              vdo_max_discard=1
       }
       EOF
       # lvcreate --vdo -L10G --metadataprofile vdo_create vg/vdopool0
       # lvcreate --vdo -L10G --config 'allocation/vdo_cpu_threads=4' vg/vdopool1
   5. Set or change VDO settings with option --vdosettings
       Use the form 'option=value' or 'option1=value option2=value', or repeat
       --vdosettings  for  each  option  being set.  Options are listed in the
       Example section  above,  for  the  full  description  see  lvm.conf(5).
       Options  can  omit  'vdo_'  and  'vdo_use_' prefixes and all its under-
       scores.  So i.e.  vdo_use_metadata_hints=1   and   metadatahints=1  are
       equivalent.  To change the option for an already existing VDOPoolLV use
       lvchange(8) command. However not all option can be changed.  Only  com-
       pression  and  deduplication  options can be also changed for an active
       VDO LV.  Lowest priority options are specified with configuration file,
       then with --vdosettings and highest are expliction option --compression
       and --deduplication.
       Example
       # lvcreate --vdo -L10G --vdosettings 'ack_threads=1 hash_zone_threads=2' vg/vdopool0
       # lvchange --vdosettings 'bio_threads=2 deduplication=1' vg/vdopool0
   6. Checking the usage of VDOPoolLV
       To quickly check how much data on a VDOPoolLV is already consumed,  use
       lvs(8).  The  Data% field reports how much data is occupied in the con-
       tent of the virtual data for the VDOLV and how much  space  is  already
       consumed with all the data and metadata blocks in the VDOPoolLV.  For a
       detailed description, use the vdostats(8) command.
       Note: vdostats(8) currently understands only /dev/mapper device names.
       Example
       # lvcreate --type vdo -L10G -V20G -n vdo0 vg/vdopool0
       # mkfs.ext4 -E nodiscard /dev/vg/vdo0
       # lvs -a vg
         LV               VG Attr       LSize  Pool     Origin Data%
         vdo0             vg vwi-a-v--- 20.00g vdopool0        0.01
         vdopool0         vg dwi-ao---- 10.00g                 30.16
         [vdopool0_vdata] vg Dwi-ao---- 10.00g
       # vdostats --all /dev/mapper/vg-vdopool0-vpool
       /dev/mapper/vg-vdopool0 :
         version                             : 30
         release version                     : 133524
         data blocks used                    : 79
         ...
   7. Extending the VDOPoolLV size
       You can add more space to hold VDO data and metadata by  extending  the
       VDODataLV  using  the commands lvresize(8) and lvextend(8).  The exten-
       sion needs to add at least one new VDO slab. You can configure the slab
       size with the allocation/vdo_slab_size_mb setting.
       You  can  also enable automatic size extension of a monitored VDOPoolLV
       with   the   activation/vdo_pool_autoextend_percent   and   activation/
       vdo_pool_autoextend_threshold settings.
       Note: You cannot reduce the size of a VDOPoolLV.
       lvextend -L+AddingSize VG/VDOPoolLV
       Example
       # lvextend -L+50G vg/vdopool0
       # lvresize -L300G vg/vdopool1
   8. Extending or reducing the VDOLV size
       You  can  extend  or  reduce a virtual VDO LV as a standard LV with the
       lvresize(8), lvextend(8), and lvreduce(8) commands.
       Note: The reduction needs to process TRIM  for  reduced  disk  area  to
       unmap  used  data  blocks  from  the VDOPoolLV, which might take a long
       time.
       lvextend -L+AddingSize VG/VDOLV
       lvreduce -L-ReducingSize VG/VDOLV
       Example
       # lvextend -L+50G vg/vdo0
       # lvreduce -L-50G vg/vdo1
       # lvresize -L200G vg/vdo2
   9. Component activation of a VDODataLV
       You can activate a VDODataLV separately as a component LV for  examina-
       tion purposes. The activation of the VDODataLV activates the data LV in
       read-only mode, and the data LV cannot be modified.  If  the  VDODataLV
       is  active  as  a  component,  any upper LV using this volume CANNOT be
       activated. You have to deactivate the VDODataLV first  to  continue  to
       use the VDOPoolLV.
       Example
       # lvchange -ay vg/vpool0_vdata
       # lvchange -an vg/vpool0_vdata
VDO TOPICS
   1. Stacking VDO
       You can convert or stack a VDOPooLV with these currently supported vol-
       ume types: linear, stripe, raid, and cache with cachepool.
   2. VDOPoolLV on top of raid
       Using a raid type LV for a VDODataLV.
       Example
       # lvcreate --type raid1 -L 5G -n vdopool vg
       # lvconvert --type vdo-pool -V 10G vg/vdopool
   3. Caching a VDOPoolLV
       VDOPoolLV (accepts also VDODataLV volume name) caching provides a mech-
       anism to accelerate reads and writes of already compressed and dedupli-
       cated data blocks together with VDO metadata.
       Example
       # lvcreate --type vdo -L 5G -V 10G -n vdo1 vg/vdopool
       # lvcreate --type cache-pool -L 1G -n cachepool vg
       # lvconvert --cache --cachepool vg/cachepool vg/vdopool
       # lvconvert --uncache vg/vdopool
   4. Caching a VDOLV
       VDO LV cache allow you to  'cache'  a  device  for  better  performance
       before it hits the processing of the VDO Pool LV layer.
       Example
       # lvcreate --type vdo -L 5G -V 10G -n vdo1 vg/vdopool
       # lvcreate --type cache-pool -L 1G -n cachepool vg
       # lvconvert --cache --cachepool vg/cachepool vg/vdo1
       # lvconvert --uncache vg/vdo1
   5. Usage of Discard/TRIM with a VDOLV
       You can discard data on a VDO LV and reduce used blocks on a VDOPoolLV.
       However, the current performance of discard  operations  is  still  not
       optimal  and  takes  a considerable amount of time and CPU.  Unless you
       really need it, you should avoid using discard.
       When a block device is going to be rewritten, its blocks will be  auto-
       matically  reused  for  new data.  Discard is useful in situations when
       user knows that the given portion of a VDO LV is not going to  be  used
       and  the  discarded  space  can be used for block provisioning in other
       regions of the VDO LV.  For the same reason,  you  should  avoid  using
       mkfs  with  discard  for a freshly created VDO LV to save a lot of time
       that this operation would take otherwise as device is already  expected
       to be empty.
   6. Memory usage
       The VDO target requires 38 MiB of RAM and several variable amounts:
       o 1.15  MiB  of  RAM for each 1 MiB of configured block map cache size.
         The block map cache requires a minimum of 150 MiB RAM.
       o 1.6 MiB of RAM for each 1 TiB of logical space.
       o 268 MiB of RAM for each 1 TiB of physical storage managed by the vol-
         ume.
       UDS  requires  a  minimum  of 250 MiB of RAM, which is also the default
       amount that deduplication uses.
       The memory required for the UDS index is determined by the  index  type
       and  the required size of the deduplication window and is controlled by
       the allocation/vdo_use_sparse_index setting.
       With enabled UDS sparse indexing, it relies on the temporal locality of
       data  and  attempts  to  retain only the most relevant index entries in
       memory and can maintain a deduplication window that is ten times larger
       than with dense while using the same amount of memory.
       Although  the  sparse  index  provides the greatest coverage, the dense
       index provides more deduplication advice.  For  most  workloads,  given
       the  same  amount  of  memory,  the  difference  in deduplication rates
       between dense and sparse indexes is negligible.
       A dense index with 1 GiB of RAM maintains a 1 TiB deduplication window,
       while a sparse index with 1 GiB of RAM maintains a 10 TiB deduplication
       window.  In general, 1 GiB is sufficient for 4 TiB  of  physical  space
       with a dense index and 40 TiB with a sparse index.
   7. Storage space requirements
       You can configure a VDOPoolLV to use up to 256 TiB of physical storage.
       Only a certain part of the physical storage is usable  to  store  data.
       This  section provides the calculations to determine the usable size of
       a VDO-managed volume.
       The VDO target requires storage for two types of VDO metadata  and  for
       the UDS index:
       o The  first  type  of VDO metadata uses approximately 1 MiB for each 4
         GiB of physical storage plus an additional 1 MiB per slab.
       o The second type of VDO metadata consumes approximately 1.25  MiB  for
         each 1 GiB of logical storage, rounded up to the nearest slab.
       o The  amount of storage required for the UDS index depends on the type
         of index and the amount of RAM allocated to the index. For each 1 GiB
         of  RAM,  a  dense  UDS index uses 17 GiB of storage and a sparse UDS
         index will use 170 GiB of storage.
SEE ALSO
       lvm(8), lvm.conf(5), lvmconfig(8), lvcreate(8), lvconvert(8),
       lvchange(8), lvextend(8), lvreduce(8), lvresize(8), lvremove(8),
       lvs(8),
       vdo(8), vdoformat(8), vdostats(8),
       mkfs(8)
Red Hat, Inc        LVM TOOLS 2.03.14(2)-RHEL8 (2021-10-20)          LVMVDO(7)