Provided by: lvm2_2.02.133-1ubuntu10_amd64 

NAME
lvcreate - create a logical volume in an existing volume group
SYNOPSIS
lvcreate [-a|--activate [a][e|l|s]{y|n}] [--addtag Tag] [--alloc AllocationPolicy] [-A|--autobackup
{y|n}] [-H|--cache] [--cachemode {passthrough|writeback|writethrough}] [--cachepolicy policy]
[--cachepool CachePoolLogicalVolume] [--cachesettings key=value] [-c|--chunksize ChunkSize]
[--commandprofile ProfileName] [-C|--contiguous {y|n}] [-d|--debug] [--discards
{ignore|nopassdown|passdown}] [--errorwhenfull {y|n}] [{-l|--extents LogicalExtentsNumber[%{FREE|PVS|VG}]
| -L|--size LogicalVolumeSize} [-i|--stripes Stripes [-I|--stripesize StripeSize]]] [-h|-?|--help]
[-K|--ignoreactivationskip] [--ignoremonitoring] [--minor minor [-j|--major major]] [--metadataprofile
ProfileName] [-m|--mirrors Mirrors [--corelog|--mirrorlog {disk|core|mirrored}] [--nosync]
[-R|--regionsize MirrorLogRegionSize]] [--monitor {y|n}] [-n|--name LogicalVolume] [--noudevsync]
[-p|--permission {r|rw}] [-M|--persistent {y|n}] [--poolmetadatasize MetadataVolumeSize]
[--poolmetadataspare {y|n}] [--[raid]maxrecoveryrate Rate] [--[raid]minrecoveryrate Rate] [-r|--readahead
{ReadAheadSectors|auto|none}] [-k|--setactivationskip {y|n}] [-s|--snapshot] [-V|--virtualsize
VirtualSize] [-t|--test] [-T|--thin] [--thinpool ThinPoolLogicalVolume] [--type SegmentType]
[-v|--verbose] [-W|--wipesignatures {y|n}] [-Z|--zero {y|n}]
[VolumeGroup[/{ExternalOrigin|Origin|Pool}LogicalVolumeName [PhysicalVolumePath[:PE[-PE]]...]]
lvcreate [-l|--extents LogicalExtentsNumber[%{FREE|ORIGIN|PVS|VG}] | -L|--size LogicalVolumeSize]
[-c|--chunksize ChunkSize] [--commandprofile ProfileName] [--noudevsync] [--ignoremonitoring]
[--metadataprofile ProfileName] [--monitor {y|n}] [-n|--name SnapshotLogicalVolume]
-s|--snapshot|-H|--cache {[VolumeGroup/]OriginalLogicalVolume [-V|--virtualsize VirtualSize]}
DESCRIPTION
lvcreate creates a new logical volume in a volume group (see vgcreate(8), vgchange(8)) by allocating
logical extents from the free physical extent pool of that volume group. If there are not enough free
physical extents then the volume group can be extended (see vgextend(8)) with other physical volumes or
by reducing existing logical volumes of this volume group in size (see lvreduce(8)). If you specify one
or more PhysicalVolumes, allocation of physical extents will be restricted to these volumes.
The second form supports the creation of snapshot logical volumes which keep the contents of the original
logical volume for backup purposes.
OPTIONS
See lvm(8) for common options.
-a|--activate [a][l|e|s]{y|n}
Controls the availability of the Logical Volumes for immediate use after the command finishes
running. By default, new Logical Volumes are activated (-ay). If it is possible technically, -an
will leave the new Logical Volume inactive. But for example, snapshots of active origin can only
be created in the active state so -an cannot be used with --type snapshot. This does not apply to
thin volume snapshots, which are by default created with flag to skip their activation (-ky).
Normally the --zero n argument has to be supplied too because zeroing (the default behaviour) also
requires activation. If autoactivation option is used (-aay), the logical volume is activated
only if it matches an item in the activation/auto_activation_volume_list set in lvm.conf(5). For
autoactivated logical volumes, --zero n and --wipesignatures n is always assumed and it can't be
overridden. If the clustered locking is enabled, -aey will activate exclusively on one node and
-a{a|l}y will activate only on the local node.
-H|--cache
Creates cache or cache pool logical volume. Specifying the optional argument --extents or --size
will cause the creation of the cache logical volume. When the Volume group name is specified
together with existing logical volume name which is NOT a cache pool name, such volume is treated
as cache origin volume and cache pool is created. In this case the --extents or --size is used to
specify size of cache pool volume. See lvmcache(7) for more info about caching support. Note
that the cache segment type requires a dm-cache kernel module version 1.3.0 or greater.
--cachemode {passthrough|writeback|writethrough}
Specifying a cache mode determines when the writes to a cache LV are considered complete. When
writeback is specified, a write is considered complete as soon as it is stored in the cache pool
LV. If writethough is specified, a write is considered complete only when it has been stored in
the cache pool LV and on the origin LV. While writethrough may be slower for writes, it is more
resilient if something should happen to a device associated with the cache pool LV.
--cachepolicy policy
Only applicable to cached LVs; see also lvmcache(7). Sets the cache policy. mq is the basic policy
name. smq is more advanced version available in newer kernels.
--cachepool CachePoolLogicalVolume{Name|Path}
Specifies the name of cache pool volume name. The other way to specify pool name is to append name
to Volume group name argument.
--cachesettings key=value
Only applicable to cached LVs; see also lvmcache(7). Sets the cache tunable settings. In most use-
cases, default values should be adequate. Special string value default switches setting back to
its default kernel value and removes it from the list of settings stored in lvm2 metadata.
-c|--chunksize ChunkSize[b|B|s|S|k|K|m|M|g|G]
Gives the size of chunk for snapshot, cache pool and thin pool logical volumes. Default unit is
in kilobytes.
For snapshots the value must be power of 2 between 4KiB and 512KiB and the default value is 4KiB.
For cache pools the value must a multiple of 32KiB between 32KiB and 1GiB. The default is 64KiB.
For thin pools the value must be a multiple of 64KiB between 64KiB and 1GiB. Default value starts
with 64KiB and grows up to fit the pool metadata size within 128MiB, if the pool metadata size is
not specified. See lvm.conf(5) setting allocation/thin_pool_chunk_size_policy to select different
calculation policy. Thin pool target version <1.4 requires this value to be a power of 2. For
target version <1.5 discard is not supported for non power of 2 values.
-C|--contiguous {y|n}
Sets or resets the contiguous allocation policy for logical volumes. Default is no contiguous
allocation based on a next free principle.
--corelog
This is shortcut for option --mirrorlog core.
--discards {ignore|nopassdown|passdown}
Sets discards behavior for thin pool. Default is passdown.
--errorwhenfull {y|n}
Configures thin pool behaviour when data space is exhausted. Default is no. Device will queue
I/O operations until target timeout (see dm-thin-pool kernel module option no_space_timeout)
expires. Thus configured system has a time to i.e. extend the size of thin pool data device. When
set to yes, the I/O operation is immeditelly errored.
-K|--ignoreactivationskip
Ignore the flag to skip Logical Volumes during activation. Use --setactivationskip option to set
or reset activation skipping flag persistently for logical volume.
--ignoremonitoring
Make no attempt to interact with dmeventd unless --monitor is specified.
-l|--extents LogicalExtentsNumber[%{VG|PVS|FREE|ORIGIN}]
Gives the number of logical extents to allocate for the new logical volume. The total number of
physical extents allocated will be greater than this, for example, if the volume is mirrored. The
number can also be expressed as a percentage of the total space in the Volume Group with the
suffix %VG, as a percentage of the remaining free space in the Volume Group with the suffix %FREE,
as a percentage of the remaining free space for the specified PhysicalVolume(s) with the suffix
%PVS, or (for a snapshot) as a percentage of the total space in the Origin Logical Volume with the
suffix %ORIGIN (i.e. 100%ORIGIN provides space for the whole origin). When expressed as a
percentage, the number is treated as an approximate upper limit for the number of physical extents
to be allocated (including extents used by any mirrors, for example).
-j|--major major
Sets the major number. Major numbers are not supported with pool volumes. This option is
supported only on older systems (kernel version 2.4) and is ignored on modern Linux systems where
major numbers are dynamically assigned.
--metadataprofile ProfileName
Uses and attaches the ProfileName configuration profile to the logical volume metadata. Whenever
the logical volume is processed next time, the profile is automatically applied. If the volume
group has another profile attached, the logical volume profile is preferred. See lvm.conf(5) for
more information about metadata profiles.
--minor minor
Sets the minor number. Minor numbers are not supported with pool volumes.
-m|--mirrors mirrors
Creates a mirrored logical volume with mirrors copies. For example, specifying -m 1 would result
in a mirror with two-sides; that is, a linear volume plus one copy.
Specifying the optional argument --nosync will cause the creation of the mirror to skip the
initial resynchronization. Any data written afterwards will be mirrored, but the original
contents will not be copied. This is useful for skipping a potentially long and resource
intensive initial sync of an empty device.
There are two implementations of mirroring which can be used and correspond to the "raid1" and
"mirror" segment types. The default is "raid1". See the --type option for more information if
you would like to use the legacy "mirror" segment type. See lvm.conf(5) settings
global/mirror_segtype_default and global/raid10_segtype_default to configure default mirror
segment type. The options --mirrorlog and --corelog apply to the legacy "mirror" segment type
only.
--mirrorlog {disk|core|mirrored}
Specifies the type of log to be used for logical volumes utilizing the legacy "mirror" segment
type.
The default is disk, which is persistent and requires a small amount of storage space, usually on
a separate device from the data being mirrored.
Using core means the mirror is regenerated by copying the data from the first device each time the
logical volume is activated, like after every reboot.
Using mirrored will create a persistent log that is itself mirrored.
--monitor {y|n}
Starts or avoids monitoring a mirrored, snapshot or thin pool logical volume with dmeventd, if it
is installed. If a device used by a monitored mirror reports an I/O error, the failure is handled
according to activation/mirror_image_fault_policy and activation/mirror_log_fault_policy set in
lvm.conf(5).
-n|--name LogicalVolume{Name|Path}
Sets the name for the new logical volume.
Without this option a default name of "lvol#" will be generated where # is the LVM internal number
of the logical volume.
--nosync
Causes the creation of the mirror to skip the initial resynchronization.
--noudevsync
Disables udev synchronisation. The process will not wait for notification from udev. It will
continue irrespective of any possible udev processing in the background. You should only use this
if udev is not running or has rules that ignore the devices LVM2 creates.
-p|--permission {r|rw}
Sets access permissions to read only (r) or read and write (rw).
Default is read and write.
-M|--persistent {y|n}
Set to y to make the minor number specified persistent. Pool volumes cannot have persistent major
and minor numbers. Defaults to yes only when major or minor number is specified. Otherwise it is
no.
--poolmetadatasize MetadataVolumeSize[b|B|s|S|k|K|m|M|g|G]
Sets the size of pool's metadata logical volume. Supported values are in range between 2MiB and
16GiB for thin pool, and upto 16GiB for cache pool. The minimum value is computed from pool's data
size. Default value for thin pool is (Pool_LV_size / Pool_LV_chunk_size * 64b). Default unit is
megabytes.
--poolmetadataspare {y|n}
Controls creation and maintanence of pool metadata spare logical volume that will be used for
automated pool recovery. Only one such volume is maintained within a volume group with the size
of the biggest pool metadata volume. Default is yes.
--[raid]maxrecoveryrate Rate[b|B|s|S|k|K|m|M|g|G]
Sets the maximum recovery rate for a RAID logical volume. Rate is specified as an amount per
second for each device in the array. If no suffix is given, then KiB/sec/device is assumed.
Setting the recovery rate to 0 means it will be unbounded.
--[raid]minrecoveryrate Rate[b|B|s|S|k|K|m|M|g|G]
Sets the minimum recovery rate for a RAID logical volume. Rate is specified as an amount per
second for each device in the array. If no suffix is given, then KiB/sec/device is assumed.
Setting the recovery rate to 0 means it will be unbounded.
-r|--readahead {ReadAheadSectors|auto|none}
Sets read ahead sector count of this logical volume. For volume groups with metadata in lvm1
format, this must be a value between 2 and 120. The default value is auto which allows the kernel
to choose a suitable value automatically. none is equivalent to specifying zero.
-R|--regionsize MirrorLogRegionSize[b|B|s|S|k|K|m|M|g|G]
A mirror is divided into regions of this size (in MiB), and the mirror log uses this granularity
to track which regions are in sync.
-k|--setactivationskip {y|n}
Controls whether Logical Volumes are persistently flagged to be skipped during activation. By
default, thin snapshot volumes are flagged for activation skip. See lvm.conf(5)
activation/auto_set_activation_skip how to change its default behaviour. To activate such
volumes, an extra --ignoreactivationskip option must be used. The flag is not applied during
deactivation. Use lvchange --setactivationskip command to change the skip flag for existing
volumes. To see whether the flag is attached, use lvs command where the state of the flag is
reported within lv_attr bits.
-L|--size LogicalVolumeSize[b|B|s|S|k|K|m|M|g|G|t|T|p|P|e|E]
Gives the size to allocate for the new logical volume. A size suffix of B for bytes, S for
sectors as 512 bytes, K for kilobytes, M for megabytes, G for gigabytes, T for terabytes, P for
petabytes or E for exabytes is optional.
Default unit is megabytes.
-s|--snapshot OriginalLogicalVolume{Name|Path}
Creates a snapshot logical volume (or snapshot) for an existing, so called original logical volume
(or origin). Snapshots provide a 'frozen image' of the contents of the origin while the origin
can still be updated. They enable consistent backups and online recovery of removed/overwritten
data/files.
Thin snapshot is created when the origin is a thin volume and the size IS NOT specified. Thin
snapshot shares same blocks within the thin pool volume. The non thin volume snapshot with the
specified size does not need the same amount of storage the origin has. In a typical scenario,
15-20% might be enough. In case the snapshot runs out of storage, use lvextend(8) to grow it.
Shrinking a snapshot is supported by lvreduce(8) as well. Run lvs(8) on the snapshot in order to
check how much data is allocated to it. Note: a small amount of the space you allocate to the
snapshot is used to track the locations of the chunks of data, so you should allocate slightly
more space than you actually need and monitor (--monitor) the rate at which the snapshot data is
growing so you can avoid running out of space. If --thinpool is specified, thin volume is created
that will use given original logical volume as an external origin that serves unprovisioned
blocks. Only read-only volumes can be used as external origins. To make the volume external
origin, lvm expects the volume to be inactive. External origin volume can be used/shared for many
thin volumes even from different thin pools. See lvconvert(8) for online conversion to thin
volumes with external origin.
-i|--stripes Stripes
Gives the number of stripes. This is equal to the number of physical volumes to scatter the
logical volume. When creating a RAID 4/5/6 logical volume, the extra devices which are necessary
for parity are internally accounted for. Specifying -i 3 would use 3 devices for striped logical
volumes, 4 devices for RAID 4/5, and 5 devices for RAID 6. Alternatively, RAID 4/5/6 will stripe
across all PVs in the volume group or all of the PVs specified if the -i argument is omitted.
-I|--stripesize StripeSize
Gives the number of kilobytes for the granularity of the stripes.
StripeSize must be 2^n (n = 2 to 9) for metadata in LVM1 format. For metadata in LVM2 format, the
stripe size may be a larger power of 2 but must not exceed the physical extent size.
-T|--thin
Creates thin pool or thin logical volume or both. Specifying the optional argument --size or
--extents will cause the creation of the thin pool logical volume. Specifying the optional
argument --virtualsize will cause the creation of the thin logical volume from given thin pool
volume. Specifying both arguments will cause the creation of both thin pool and thin volume using
this pool. See lvmthin(7) for more info about thin provisioning support. Thin provisioning
requires device mapper kernel driver from kernel 3.2 or greater.
--thinpool ThinPoolLogicalVolume{Name|Path}
Specifies the name of thin pool volume name. The other way to specify pool name is to append name
to Volume group name argument.
--type SegmentType
Creates a logical volume with the specified segment type. Supported types are: cache, cache-pool,
error, linear, mirror, raid1, raid4, raid5_la, raid5_ls (= raid5), raid5_ra, raid5_rs, raid6_nc,
raid6_nr, raid6_zr (= raid6), raid10, snapshot, striped, thin, thin-pool or zero. Segment type
may have a commandline switch alias that will enable its use. When the type is not explicitly
specified an implicit type is selected from combination of options: -H|--cache|--cachepool (cache
or cachepool), -T|--thin|--thinpool (thin or thinpool), -m|--mirrors (raid1 or mirror),
-s|--snapshot|-V|--virtualsize (snapshot or thin), -i|--stripes (striped). Default segment type
is linear.
-V|--virtualsize VirtualSize[b|B|s|S|k|K|m|M|g|G|t|T|p|P|e|E]
Creates a thinly provisioned device or a sparse device of the given size (in MiB by default). See
lvm.conf(5) settings global/sparse_segtype_default to configure default sparse segment type. See
lvmthin(7) for more info about thin provisioning support. Anything written to a sparse snapshot
will be returned when reading from it. Reading from other areas of the device will return blocks
of zeros. Virtual snapshot (sparse snapshot) is implemented by creating a hidden virtual device
of the requested size using the zero target. A suffix of _vorigin is used for this device. Note:
using sparse snapshots is not efficient for larger device sizes (GiB), thin provisioning should be
used for this case.
-W|--wipesignatures {y|n}
Controls wiping of detected signatures on newly created Logical Volume. If this option is not
specified, then by default signature wiping is done each time the zeroing ( -Z|--zero ) is done.
This default behaviour can be controlled by allocation/wipe_signatures_when_zeroing_new_lvs
setting found in lvm.conf(5).
If blkid wiping is used allocation/use_blkid_wiping setting in lvm.conf(5)) and LVM2 is compiled
with blkid wiping support, then blkid(8) library is used to detect the signatures (use blkid -k
command to list the signatures that are recognized). Otherwise, native LVM2 code is used to
detect signatures (MD RAID, swap and LUKS signatures are detected only in this case).
Logical volume is not wiped if the read only flag is set.
-Z|--zero {y|n}
Controls zeroing of the first 4KiB of data in the new logical volume. Default is yes. Snapshot
COW volumes are always zeroed. Logical volume is not zeroed if the read only flag is set.
Warning: trying to mount an unzeroed logical volume can cause the system to hang.
Examples
Creates a striped logical volume with 3 stripes, a stripe size of 8KiB and a size of 100MiB in the volume
group named vg00. The logical volume name will be chosen by lvcreate:
lvcreate -i 3 -I 8 -L 100M vg00
Creates a mirror logical volume with 2 sides with a useable size of 500 MiB. This operation would
require 3 devices (or option --alloc anywhere) - two for the mirror devices and one for the disk log:
lvcreate -m1 -L 500M vg00
Creates a mirror logical volume with 2 sides with a useable size of 500 MiB. This operation would
require 2 devices - the log is "in-memory":
lvcreate -m1 --mirrorlog core -L 500M vg00
Creates a snapshot logical volume named "vg00/snap" which has access to the contents of the original
logical volume named "vg00/lvol1" at snapshot logical volume creation time. If the original logical
volume contains a file system, you can mount the snapshot logical volume on an arbitrary directory in
order to access the contents of the filesystem to run a backup while the original filesystem continues to
get updated:
lvcreate --size 100m --snapshot --name snap /dev/vg00/lvol1
Creates a snapshot logical volume named "vg00/snap" with size for overwriting 20% of the original logical
volume named "vg00/lvol1".:
lvcreate -s -l 20%ORIGIN --name snap vg00/lvol1
Creates a sparse device named /dev/vg1/sparse of size 1TiB with space for just under 100MiB of actual
data on it:
lvcreate --virtualsize 1T --size 100M --snapshot --name sparse vg1
Creates a linear logical volume "vg00/lvol1" using physical extents /dev/sda:0-7 and /dev/sdb:0-7 for
allocation of extents:
lvcreate -L 64M -n lvol1 vg00 /dev/sda:0-7 /dev/sdb:0-7
Creates a 5GiB RAID5 logical volume "vg00/my_lv", with 3 stripes (plus a parity drive for a total of 4
devices) and a stripesize of 64KiB:
lvcreate --type raid5 -L 5G -i 3 -I 64 -n my_lv vg00
Creates a RAID5 logical volume "vg00/my_lv", using all of the free space in the VG and spanning all the
PVs in the VG:
lvcreate --type raid5 -l 100%FREE -n my_lv vg00
Creates a 5GiB RAID10 logical volume "vg00/my_lv", with 2 stripes on 2 2-way mirrors. Note that the -i
and -m arguments behave differently. The -i specifies the number of stripes. The -m specifies the
number of additional copies:
lvcreate --type raid10 -L 5G -i 2 -m 1 -n my_lv vg00
Creates 100MiB pool logical volume for thin provisioning build with 2 stripes 64KiB and chunk size 256KiB
together with 1TiB thin provisioned logical volume "vg00/thin_lv":
lvcreate -i 2 -I 64 -c 256 -L100M -T vg00/pool -V 1T --name thin_lv
Creates a thin snapshot volume "thinsnap" of thin volume "thinvol" that will share the same blocks within
the thin pool. Note: the size MUST NOT be specified, otherwise the non-thin snapshot is created instead:
lvcreate -s vg00/thinvol --name thinsnap
Creates a thin snapshot volume of read-only inactive volume "origin" which then becomes the thin external
origin for the thin snapshot volume in vg00 that will use an existing thin pool "vg00/pool":
lvcreate -s --thinpool vg00/pool origin
Create a cache pool LV that can later be used to cache one logical volume.
lvcreate --type cache-pool -L 1G -n my_lv_cachepool vg /dev/fast1
If there is an existing cache pool LV, create the large slow device (i.e. the origin LV) and link it to
the supplied cache pool LV, creating a cache LV.
lvcreate --cache -L 100G -n my_lv vg/my_lv_cachepool /dev/slow1
If there is an existing logical volume, create the small and fast cache pool LV and link it to the
supplied existing logical volume (i.e. the origin LV), creating a cache LV.
lvcreate --type cache -L 1G -n my_lv_cachepool vg/my_lv /dev/fast1
SEE ALSO
lvm(8), lvm.conf(5), lvmcache(7), lvmthin(7), lvconvert(8), lvchange(8), lvextend(8), lvreduce(8),
lvremove(8), lvrename(8) lvs(8), lvscan(8), vgcreate(8), blkid(8)
Sistina Software UK LVM TOOLS 2.02.133(2) (2015-10-30) LVCREATE(8)