Provided by: lvm2_2.02.133-1ubuntu10_amd64 bug


       lvcreate - create a logical volume in an existing volume group


       lvcreate [-a|--activate [a][e|l|s]{y|n}] [--addtag Tag] [--alloc
       AllocationPolicy] [-A|--autobackup {y|n}] [-H|--cache] [--cachemode
       {passthrough|writeback|writethrough}] [--cachepolicy policy]
       [--cachepool CachePoolLogicalVolume] [--cachesettings key=value]
       [-c|--chunksize ChunkSize] [--commandprofile ProfileName]
       [-C|--contiguous {y|n}] [-d|--debug] [--discards
       {ignore|nopassdown|passdown}] [--errorwhenfull {y|n}] [{-l|--extents
       LogicalExtentsNumber[%{FREE|PVS|VG}] | -L|--size LogicalVolumeSize}
       [-i|--stripes Stripes [-I|--stripesize StripeSize]]] [-h|-?|--help]
       [-K|--ignoreactivationskip] [--ignoremonitoring] [--minor minor
       [-j|--major major]] [--metadataprofile ProfileName] [-m|--mirrors
       Mirrors [--corelog|--mirrorlog {disk|core|mirrored}] [--nosync]
       [-R|--regionsize MirrorLogRegionSize]] [--monitor {y|n}] [-n|--name
       LogicalVolume] [--noudevsync] [-p|--permission {r|rw}] [-M|--persistent
       {y|n}] [--poolmetadatasize MetadataVolumeSize] [--poolmetadataspare
       {y|n}] [--[raid]maxrecoveryrate Rate] [--[raid]minrecoveryrate Rate]
       [-r|--readahead {ReadAheadSectors|auto|none}] [-k|--setactivationskip
       {y|n}] [-s|--snapshot] [-V|--virtualsize VirtualSize] [-t|--test]
       [-T|--thin] [--thinpool ThinPoolLogicalVolume] [--type SegmentType]
       [-v|--verbose] [-W|--wipesignatures {y|n}] [-Z|--zero {y|n}]

       lvcreate [-l|--extents LogicalExtentsNumber[%{FREE|ORIGIN|PVS|VG}] |
       -L|--size LogicalVolumeSize] [-c|--chunksize ChunkSize]
       [--commandprofile ProfileName] [--noudevsync] [--ignoremonitoring]
       [--metadataprofile ProfileName] [--monitor {y|n}] [-n|--name
       SnapshotLogicalVolume] -s|--snapshot|-H|--cache
       {[VolumeGroup/]OriginalLogicalVolume [-V|--virtualsize VirtualSize]}


       lvcreate   creates  a  new  logical  volume  in  a  volume  group  (see
       vgcreate(8), vgchange(8)) by allocating logical extents from  the  free
       physical  extent  pool  of  that volume group.  If there are not enough
       free physical extents then  the  volume  group  can  be  extended  (see
       vgextend(8))  with  other  physical  volumes  or  by  reducing existing
       logical volumes of this volume group in size (see lvreduce(8)).  If you
       specify  one  or  more  PhysicalVolumes, allocation of physical extents
       will be restricted to these volumes.
       The second form supports the creation of snapshot logical volumes which
       keep the contents of the original logical volume for backup purposes.


       See lvm(8) for common options.

       -a|--activate [a][l|e|s]{y|n}
              Controls  the  availability of the Logical Volumes for immediate
              use after the command finishes running.  By default, new Logical
              Volumes are activated (-ay).  If it is possible technically, -an
              will leave the new Logical Volume  inactive.  But  for  example,
              snapshots  of  active  origin  can only be created in the active
              state so -an cannot be used with --type snapshot. This does  not
              apply  to  thin  volume  snapshots, which are by default created
              with flag to skip their activation (-ky).  Normally the --zero n
              argument  has  to  be  supplied too because zeroing (the default
              behaviour) also requires activation.  If  autoactivation  option
              is  used  (-aay),  the  logical  volume  is activated only if it
              matches an item  in  the  activation/auto_activation_volume_list
              set in lvm.conf(5).  For autoactivated logical volumes, --zero n
              and --wipesignatures  n  is  always  assumed  and  it  can't  be
              overridden.  If  the  clustered  locking  is  enabled, -aey will
              activate exclusively on one node and -a{a|l}y will activate only
              on the local node.

              Creates  cache  or  cache  pool  logical volume.  Specifying the
              optional argument --extents or --size will cause the creation of
              the  cache  logical  volume.   When  the  Volume  group  name is
              specified together with existing logical volume  name  which  is
              NOT  a  cache  pool name, such volume is treated as cache origin
              volume and cache pool is created. In this case the --extents  or
              --size  is  used  to  specify  size  of  cache pool volume.  See
              lvmcache(7) for more info about caching support.  Note that  the
              cache  segment  type  requires  a dm-cache kernel module version
              1.3.0 or greater.

       --cachemode {passthrough|writeback|writethrough}
              Specifying a cache mode determines when the writes to a cache LV
              are  considered  complete.  When writeback is specified, a write
              is considered complete as soon as it is stored in the cache pool
              LV.  If writethough is specified, a write is considered complete
              only when it has been stored in the cache pool  LV  and  on  the
              origin  LV.   While writethrough may be slower for writes, it is
              more resilient if something should happen to a device associated
              with the cache pool LV.

       --cachepolicy policy
              Only  applicable  to  cached LVs; see also lvmcache(7). Sets the
              cache policy. mq is the basic policy name. smq is more  advanced
              version available in newer kernels.

       --cachepool CachePoolLogicalVolume{Name|Path}
              Specifies  the  name of cache pool volume name. The other way to
              specify pool name  is  to  append  name  to  Volume  group  name

       --cachesettings key=value
              Only  applicable  to  cached LVs; see also lvmcache(7). Sets the
              cache tunable settings. In most use-cases, default values should
              be adequate.  Special string value default switches setting back
              to its default kernel value and removes  it  from  the  list  of
              settings stored in lvm2 metadata.

       -c|--chunksize ChunkSize[b|B|s|S|k|K|m|M|g|G]
              Gives  the  size of chunk for snapshot, cache pool and thin pool
              logical volumes.  Default unit is in kilobytes.
              For snapshots the value must be power  of  2  between  4KiB  and
              512KiB and the default value is 4KiB.
              For cache pools the value must a multiple of 32KiB between 32KiB
              and 1GiB. The default is 64KiB.
              For thin pools the value must be a  multiple  of  64KiB  between
              64KiB and 1GiB.  Default value starts with 64KiB and grows up to
              fit the pool metadata size within 128MiB, if the  pool  metadata
              size    is    not    specified.     See    lvm.conf(5)   setting
              allocation/thin_pool_chunk_size_policy   to   select   different
              calculation policy.  Thin pool target version <1.4 requires this
              value to be a power of 2.  For target version  <1.5  discard  is
              not supported for non power of 2 values.

       -C|--contiguous {y|n}
              Sets  or  resets  the  contiguous  allocation policy for logical
              volumes. Default is no contiguous allocation  based  on  a  next
              free principle.

              This is shortcut for option --mirrorlog core.

       --discards {ignore|nopassdown|passdown}
              Sets discards behavior for thin pool.  Default is passdown.

       --errorwhenfull {y|n}
              Configures  thin  pool  behaviour  when data space is exhausted.
              Default is no.  Device will queue I/O  operations  until  target
              timeout (see dm-thin-pool kernel module option no_space_timeout)
              expires. Thus configured system has a time to  i.e.  extend  the
              size  of  thin  pool  data  device.   When  set  to yes, the I/O
              operation is immeditelly errored.

              Ignore the flag to skip Logical Volumes during activation.   Use
              --setactivationskip  option  to set or reset activation skipping
              flag persistently for logical volume.

              Make no attempt to interact with dmeventd  unless  --monitor  is

       -l|--extents LogicalExtentsNumber[%{VG|PVS|FREE|ORIGIN}]
              Gives  the  number  of  logical  extents to allocate for the new
              logical volume.  The total number of physical extents  allocated
              will  be  greater  than  this,  for  example,  if  the volume is
              mirrored.  The number can also be expressed as a  percentage  of
              the  total  space  in the Volume Group with the suffix %VG, as a
              percentage of the remaining free space in the Volume Group  with
              the  suffix  %FREE,  as a percentage of the remaining free space
              for the specified PhysicalVolume(s) with  the  suffix  %PVS,  or
              (for  a  snapshot)  as  a  percentage  of the total space in the
              Origin Logical Volume with the suffix %ORIGIN  (i.e.  100%ORIGIN
              provides  space  for  the  whole  origin).   When expressed as a
              percentage, the number is treated as an approximate upper  limit
              for  the  number  of physical extents to be allocated (including
              extents used by any mirrors, for example).

       -j|--major major
              Sets the major number.  Major numbers  are  not  supported  with
              pool  volumes.   This  option is supported only on older systems
              (kernel version 2.4) and is  ignored  on  modern  Linux  systems
              where major numbers are dynamically assigned.

       --metadataprofile ProfileName
              Uses  and  attaches the ProfileName configuration profile to the
              logical  volume  metadata.  Whenever  the  logical   volume   is
              processed  next  time,  the profile is automatically applied. If
              the volume group  has  another  profile  attached,  the  logical
              volume   profile   is   preferred.   See  lvm.conf(5)  for  more
              information about metadata profiles.

       --minor minor
              Sets the minor number.  Minor numbers  are  not  supported  with
              pool volumes.

       -m|--mirrors mirrors
              Creates  a  mirrored  logical  volume  with mirrors copies.  For
              example, specifying -m 1 would result  in  a  mirror  with  two-
              sides; that is, a linear volume plus one copy.

              Specifying   the  optional  argument  --nosync  will  cause  the
              creation of the mirror to skip  the  initial  resynchronization.
              Any  data  written afterwards will be mirrored, but the original
              contents will not be copied.  This  is  useful  for  skipping  a
              potentially long and resource intensive initial sync of an empty

              There are two implementations of mirroring which can be used and
              correspond  to  the  "raid1"  and  "mirror"  segment types.  The
              default is "raid1".  See the --type option for more  information
              if  you would like to use the legacy "mirror" segment type.  See
              lvm.conf(5)    settings     global/mirror_segtype_default    and
              global/raid10_segtype_default   to   configure   default  mirror
              segment type.  The options --mirrorlog and  --corelog  apply  to
              the legacy "mirror" segment type only.

       --mirrorlog {disk|core|mirrored}
              Specifies  the  type  of  log  to  be  used  for logical volumes
              utilizing the legacy "mirror" segment type.
              The default is disk, which is persistent and  requires  a  small
              amount  of  storage space, usually on a separate device from the
              data being mirrored.
              Using core means the mirror is regenerated by copying  the  data
              from the first device each time the logical volume is activated,
              like after every reboot.
              Using mirrored will create  a  persistent  log  that  is  itself

       --monitor {y|n}
              Starts  or  avoids  monitoring a mirrored, snapshot or thin pool
              logical volume with dmeventd, if it is installed.  If  a  device
              used  by a monitored mirror reports an I/O error, the failure is
              handled according  to  activation/mirror_image_fault_policy  and
              activation/mirror_log_fault_policy set in lvm.conf(5).

       -n|--name LogicalVolume{Name|Path}
              Sets the name for the new logical volume.
              Without  this option a default name of "lvol#" will be generated
              where # is the LVM internal number of the logical volume.

              Causes  the  creation  of  the  mirror  to  skip   the   initial

              Disables  udev  synchronisation.  The  process will not wait for
              notification from udev.  It will continue  irrespective  of  any
              possible udev processing in the background.  You should only use
              this if udev is not running or has rules that ignore the devices
              LVM2 creates.

       -p|--permission {r|rw}
              Sets access permissions to read only (r) or read and write (rw).
              Default is read and write.

       -M|--persistent {y|n}
              Set  to  y  to make the minor number specified persistent.  Pool
              volumes  cannot  have  persistent  major  and   minor   numbers.
              Defaults  to  yes  only when major or minor number is specified.
              Otherwise it is no.

       --poolmetadatasize MetadataVolumeSize[b|B|s|S|k|K|m|M|g|G]
              Sets the size of  pool's  metadata  logical  volume.   Supported
              values  are  in  range between 2MiB and 16GiB for thin pool, and
              upto 16GiB for cache pool. The minimum value  is  computed  from
              pool's  data size.  Default value for thin pool is (Pool_LV_size
              / Pool_LV_chunk_size * 64b).  Default unit is megabytes.

       --poolmetadataspare {y|n}
              Controls creation and maintanence of pool metadata spare logical
              volume  that will be used for automated pool recovery.  Only one
              such volume is maintained within a volume group with the size of
              the biggest pool metadata volume.  Default is yes.

       --[raid]maxrecoveryrate Rate[b|B|s|S|k|K|m|M|g|G]
              Sets  the maximum recovery rate for a RAID logical volume.  Rate
              is specified as an amount per second  for  each  device  in  the
              array.   If  no suffix is given, then KiB/sec/device is assumed.
              Setting the recovery rate to 0 means it will be unbounded.

       --[raid]minrecoveryrate Rate[b|B|s|S|k|K|m|M|g|G]
              Sets the minimum recovery rate for a RAID logical volume.   Rate
              is  specified  as  an  amount  per second for each device in the
              array.  If no suffix is given, then KiB/sec/device  is  assumed.
              Setting the recovery rate to 0 means it will be unbounded.

       -r|--readahead {ReadAheadSectors|auto|none}
              Sets read ahead sector count of this logical volume.  For volume
              groups with metadata in  lvm1  format,  this  must  be  a  value
              between  2  and 120.  The default value is auto which allows the
              kernel to  choose  a  suitable  value  automatically.   none  is
              equivalent to specifying zero.

       -R|--regionsize MirrorLogRegionSize[b|B|s|S|k|K|m|M|g|G]
              A  mirror is divided into regions of this size (in MiB), and the
              mirror log uses this granularity to track which regions  are  in

       -k|--setactivationskip {y|n}
              Controls  whether Logical Volumes are persistently flagged to be
              skipped during activation. By default, thin snapshot volumes are
              flagged     for     activation     skip.      See    lvm.conf(5)
              activation/auto_set_activation_skip how to  change  its  default
              behaviour.     To    activate    such    volumes,    an    extra
              --ignoreactivationskip option must be  used.  The  flag  is  not
              applied  during  deactivation.  Use lvchange --setactivationskip
              command to change the skip flag for existing  volumes.   To  see
              whether the flag is attached, use lvs command where the state of
              the flag is reported within lv_attr bits.

       -L|--size LogicalVolumeSize[b|B|s|S|k|K|m|M|g|G|t|T|p|P|e|E]
              Gives the size to allocate for the new logical volume.   A  size
              suffix  of  B  for  bytes,  S  for  sectors  as 512 bytes, K for
              kilobytes, M for megabytes, G for gigabytes, T for terabytes,  P
              for petabytes or E for exabytes is optional.
              Default unit is megabytes.

       -s|--snapshot OriginalLogicalVolume{Name|Path}
              Creates a snapshot logical volume (or snapshot) for an existing,
              so  called  original  logical  volume  (or  origin).   Snapshots
              provide a 'frozen image' of the contents of the origin while the
              origin can still be updated. They enable consistent backups  and
              online recovery of removed/overwritten data/files.
              Thin  snapshot  is  created when the origin is a thin volume and
              the size IS NOT specified.  Thin  snapshot  shares  same  blocks
              within  the thin pool volume.  The non thin volume snapshot with
              the specified size does not need the same amount of storage  the
              origin  has.  In  a typical scenario, 15-20% might be enough. In
              case the snapshot runs out of storage, use lvextend(8)  to  grow
              it.  Shrinking  a  snapshot is supported by lvreduce(8) as well.
              Run lvs(8) on the snapshot in order to check how  much  data  is
              allocated to it.  Note: a small amount of the space you allocate
              to the snapshot is used to track the locations of the chunks  of
              data,  so  you  should  allocate  slightly  more  space than you
              actually need and monitor (--monitor)  the  rate  at  which  the
              snapshot  data is growing so you can avoid running out of space.
              If --thinpool is specified, thin volume is created that will use
              given  original logical volume as an external origin that serves
              unprovisioned blocks.  Only read-only volumes  can  be  used  as
              external  origins.   To  make  the  volume  external origin, lvm
              expects the volume to be inactive.  External origin  volume  can
              be  used/shared  for  many thin volumes even from different thin
              pools. See lvconvert(8) for online conversion  to  thin  volumes
              with external origin.

       -i|--stripes Stripes
              Gives  the  number  of  stripes.  This is equal to the number of
              physical volumes to scatter the logical volume.  When creating a
              RAID 4/5/6 logical volume, the extra devices which are necessary
              for parity are internally accounted for.  Specifying -i 3  would
              use  3  devices  for striped logical volumes, 4 devices for RAID
              4/5, and 5 devices for RAID 6.  Alternatively, RAID  4/5/6  will
              stripe  across  all  PVs  in  the volume group or all of the PVs
              specified if the -i argument is omitted.

       -I|--stripesize StripeSize
              Gives the  number  of  kilobytes  for  the  granularity  of  the
              StripeSize must be 2^n (n = 2 to 9) for metadata in LVM1 format.
              For metadata in LVM2 format, the stripe size  may  be  a  larger
              power of 2 but must not exceed the physical extent size.

              Creates  thin  pool  or thin logical volume or both.  Specifying
              the  optional  argument  --size  or  --extents  will  cause  the
              creation  of  the  thin  pool  logical  volume.   Specifying the
              optional argument --virtualsize will cause the creation  of  the
              thin  logical  volume  from  given thin pool volume.  Specifying
              both arguments will cause the creation of  both  thin  pool  and
              thin volume using this pool.  See lvmthin(7) for more info about
              thin provisioning support.  Thin  provisioning  requires  device
              mapper kernel driver from kernel 3.2 or greater.

       --thinpool ThinPoolLogicalVolume{Name|Path}
              Specifies  the  name  of thin pool volume name. The other way to
              specify pool name  is  to  append  name  to  Volume  group  name

       --type SegmentType
              Creates  a  logical  volume  with  the  specified  segment type.
              Supported types are: cache, cache-pool, error,  linear,  mirror,
              raid1,  raid4, raid5_la, raid5_ls (= raid5), raid5_ra, raid5_rs,
              raid6_nc,  raid6_nr,  raid6_zr  (=  raid6),  raid10,   snapshot,
              striped,  thin,  thin-pool  or  zero.   Segment  type may have a
              commandline switch alias that will enable  its  use.   When  the
              type  is  not  explicitly specified an implicit type is selected
              from combination of options:  -H|--cache|--cachepool  (cache  or
              cachepool),    -T|--thin|--thinpool    (thin    or    thinpool),
              -m|--mirrors (raid1 or  mirror),  -s|--snapshot|-V|--virtualsize
              (snapshot  or  thin),  -i|--stripes  (striped).  Default segment
              type is linear.

       -V|--virtualsize VirtualSize[b|B|s|S|k|K|m|M|g|G|t|T|p|P|e|E]
              Creates a thinly provisioned device or a sparse  device  of  the
              given  size  (in  MiB  by  default).   See  lvm.conf(5) settings
              global/sparse_segtype_default  to   configure   default   sparse
              segment   type.    See  lvmthin(7)  for  more  info  about  thin
              provisioning support.  Anything written  to  a  sparse  snapshot
              will be returned when reading from it.  Reading from other areas
              of the device will return blocks  of  zeros.   Virtual  snapshot
              (sparse  snapshot)  is  implemented by creating a hidden virtual
              device of the requested size using the zero target.  A suffix of
              _vorigin  is used for this device.  Note: using sparse snapshots
              is  not  efficient  for  larger   device   sizes   (GiB),   thin
              provisioning should be used for this case.

       -W|--wipesignatures {y|n}
              Controls  wiping of detected signatures on newly created Logical
              Volume.  If this  option  is  not  specified,  then  by  default
              signature  wiping is done each time the zeroing ( -Z|--zero ) is
              done.   This   default   behaviour   can   be   controlled    by
              allocation/wipe_signatures_when_zeroing_new_lvs setting found in
              If blkid wiping is used allocation/use_blkid_wiping  setting  in
              lvm.conf(5))  and  LVM2  is  compiled with blkid wiping support,
              then blkid(8) library is used  to  detect  the  signatures  (use
              blkid  -k  command  to list the signatures that are recognized).
              Otherwise, native LVM2 code is used  to  detect  signatures  (MD
              RAID, swap and LUKS signatures are detected only in this case).
              Logical volume is not wiped if the read only flag is set.

       -Z|--zero {y|n}
              Controls  zeroing  of  the first 4KiB of data in the new logical
              volume.  Default  is  yes.   Snapshot  COW  volumes  are  always
              zeroed.   Logical  volume is not zeroed if the read only flag is
              Warning: trying to mount an unzeroed logical  volume  can  cause
              the system to hang.


       Creates  a striped logical volume with 3 stripes, a stripe size of 8KiB
       and a size of 100MiB in the  volume  group  named  vg00.   The  logical
       volume name will be chosen by lvcreate:

       lvcreate -i 3 -I 8 -L 100M vg00

       Creates a mirror logical volume with 2 sides with a useable size of 500
       MiB.  This  operation  would  require  3  devices  (or  option  --alloc
       anywhere) - two for the mirror devices and one for the disk log:

       lvcreate -m1 -L 500M vg00

       Creates a mirror logical volume with 2 sides with a useable size of 500
       MiB.  This operation would require 2 devices - the log is "in-memory":

       lvcreate -m1 --mirrorlog core -L 500M vg00

       Creates a snapshot logical volume named "vg00/snap" which has access to
       the  contents  of  the  original  logical  volume named "vg00/lvol1" at
       snapshot logical volume creation time. If the original  logical  volume
       contains a file system, you can mount the snapshot logical volume on an
       arbitrary directory in order to access the contents of  the  filesystem
       to run a backup while the original filesystem continues to get updated:

       lvcreate --size 100m --snapshot --name snap /dev/vg00/lvol1

       Creates  a  snapshot  logical  volume  named  "vg00/snap" with size for
       overwriting 20% of the original logical volume named "vg00/lvol1".:

       lvcreate -s -l 20%ORIGIN --name snap vg00/lvol1

       Creates a sparse device named /dev/vg1/sparse of size 1TiB  with  space
       for just under 100MiB of actual data on it:

       lvcreate --virtualsize 1T --size 100M --snapshot --name sparse vg1

       Creates  a  linear  logical  volume "vg00/lvol1" using physical extents
       /dev/sda:0-7 and /dev/sdb:0-7 for allocation of extents:

       lvcreate -L 64M -n lvol1 vg00 /dev/sda:0-7 /dev/sdb:0-7

       Creates a 5GiB RAID5 logical volume "vg00/my_lv", with 3 stripes  (plus
       a parity drive for a total of 4 devices) and a stripesize of 64KiB:

       lvcreate --type raid5 -L 5G -i 3 -I 64 -n my_lv vg00

       Creates  a  RAID5  logical  volume  "vg00/my_lv", using all of the free
       space in the VG and spanning all the PVs in the VG:

       lvcreate --type raid5 -l 100%FREE -n my_lv vg00

       Creates a 5GiB RAID10 logical volume "vg00/my_lv", with 2 stripes on  2
       2-way  mirrors.   Note that the -i and -m arguments behave differently.
       The -i specifies the number of stripes.  The -m specifies the number of
       additional copies:

       lvcreate --type raid10 -L 5G -i 2 -m 1 -n my_lv vg00

       Creates  100MiB  pool logical volume for thin provisioning build with 2
       stripes 64KiB and chunk size 256KiB together with 1TiB thin provisioned
       logical volume "vg00/thin_lv":

       lvcreate -i 2 -I 64 -c 256 -L100M -T vg00/pool -V 1T --name thin_lv

       Creates a thin snapshot volume "thinsnap" of thin volume "thinvol" that
       will share the same blocks within the thin pool.  Note: the  size  MUST
       NOT be specified, otherwise the non-thin snapshot is created instead:

       lvcreate -s vg00/thinvol --name thinsnap

       Creates  a  thin  snapshot volume of read-only inactive volume "origin"
       which then becomes the thin  external  origin  for  the  thin  snapshot
       volume in vg00 that will use an existing thin pool "vg00/pool":

       lvcreate -s --thinpool vg00/pool origin

       Create  a  cache  pool  LV  that can later be used to cache one logical

       lvcreate --type cache-pool -L 1G -n my_lv_cachepool vg /dev/fast1

       If there is an existing cache pool LV, create  the  large  slow  device
       (i.e.  the  origin  LV)  and  link  it  to  the supplied cache pool LV,
       creating a cache LV.

       lvcreate --cache -L 100G -n my_lv vg/my_lv_cachepool /dev/slow1

       If there is an existing logical volume, create the small and fast cache
       pool  LV  and link it to the supplied existing logical volume (i.e. the
       origin LV), creating a cache LV.

       lvcreate --type cache -L 1G -n my_lv_cachepool vg/my_lv /dev/fast1


       lvm(8),    lvm.conf(5),    lvmcache(7),    lvmthin(7),    lvconvert(8),
       lvchange(8), lvextend(8), lvreduce(8), lvremove(8), lvrename(8) lvs(8),
       lvscan(8), vgcreate(8), blkid(8)