mirror of
				git://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
				synced 2025-09-04 20:19:47 +08:00 
			
		
		
		
	 751d5b2741
			
		
	
	
		751d5b2741
		
	
	
	
	
		
			
			Fix thirty five typos in dm-integrity.rst, dm-raid.rst, dm-zoned.rst, verity.rst, writecache.rst, tsx_async_abort.rst, md.rst, bttv.rst, dvb_references.rst, frontend-cardlist.rst, gspca-cardlist.rst, ipu3.rst, remote-controller.rst, mm/index.rst, numaperf.rst, userfaultfd.rst, module-signing.rst, imx-ddr.rst, intel-speed-select.rst, intel_pstate.rst, ramoops.rst, abi.rst, kernel.rst, vm.rst Signed-off-by: Andrew Klychkov <andrew.a.klychkov@gmail.com> Link: https://lore.kernel.org/r/20201204072848.GA49895@spblnx124.lan Signed-off-by: Jonathan Corbet <corbet@lwn.net>
		
			
				
	
	
		
			766 lines
		
	
	
		
			27 KiB
		
	
	
	
		
			ReStructuredText
		
	
	
	
	
	
			
		
		
	
	
			766 lines
		
	
	
		
			27 KiB
		
	
	
	
		
			ReStructuredText
		
	
	
	
	
	
| RAID arrays
 | |
| ===========
 | |
| 
 | |
| Boot time assembly of RAID arrays
 | |
| ---------------------------------
 | |
| 
 | |
| Tools that manage md devices can be found at
 | |
|    https://www.kernel.org/pub/linux/utils/raid/
 | |
| 
 | |
| 
 | |
| You can boot with your md device with the following kernel command
 | |
| lines:
 | |
| 
 | |
| for old raid arrays without persistent superblocks::
 | |
| 
 | |
|   md=<md device no.>,<raid level>,<chunk size factor>,<fault level>,dev0,dev1,...,devn
 | |
| 
 | |
| for raid arrays with persistent superblocks::
 | |
| 
 | |
|   md=<md device no.>,dev0,dev1,...,devn
 | |
| 
 | |
| or, to assemble a partitionable array::
 | |
| 
 | |
|   md=d<md device no.>,dev0,dev1,...,devn
 | |
| 
 | |
| ``md device no.``
 | |
| +++++++++++++++++
 | |
| 
 | |
| The number of the md device
 | |
| 
 | |
| ================= =========
 | |
| ``md device no.`` device
 | |
| ================= =========
 | |
|               0		md0
 | |
| 	      1		md1
 | |
| 	      2		md2
 | |
| 	      3		md3
 | |
| 	      4		md4
 | |
| ================= =========
 | |
| 
 | |
| ``raid level``
 | |
| ++++++++++++++
 | |
| 
 | |
| level of the RAID array
 | |
| 
 | |
| =============== =============
 | |
| ``raid level``  level
 | |
| =============== =============
 | |
| -1		linear mode
 | |
| 0		striped mode
 | |
| =============== =============
 | |
| 
 | |
| other modes are only supported with persistent super blocks
 | |
| 
 | |
| ``chunk size factor``
 | |
| +++++++++++++++++++++
 | |
| 
 | |
| (raid-0 and raid-1 only)
 | |
| 
 | |
| Set  the chunk size as 4k << n.
 | |
| 
 | |
| ``fault level``
 | |
| +++++++++++++++
 | |
| 
 | |
| Totally ignored
 | |
| 
 | |
| ``dev0`` to ``devn``
 | |
| ++++++++++++++++++++
 | |
| 
 | |
| e.g. ``/dev/hda1``, ``/dev/hdc1``, ``/dev/sda1``, ``/dev/sdb1``
 | |
| 
 | |
| A possible loadlin line (Harald Hoyer <HarryH@Royal.Net>)  looks like this::
 | |
| 
 | |
| 	e:\loadlin\loadlin e:\zimage root=/dev/md0 md=0,0,4,0,/dev/hdb2,/dev/hdc3 ro
 | |
| 
 | |
| 
 | |
| Boot time autodetection of RAID arrays
 | |
| --------------------------------------
 | |
| 
 | |
| When md is compiled into the kernel (not as module), partitions of
 | |
| type 0xfd are scanned and automatically assembled into RAID arrays.
 | |
| This autodetection may be suppressed with the kernel parameter
 | |
| ``raid=noautodetect``.  As of kernel 2.6.9, only drives with a type 0
 | |
| superblock can be autodetected and run at boot time.
 | |
| 
 | |
| The kernel parameter ``raid=partitionable`` (or ``raid=part``) means
 | |
| that all auto-detected arrays are assembled as partitionable.
 | |
| 
 | |
| Boot time assembly of degraded/dirty arrays
 | |
| -------------------------------------------
 | |
| 
 | |
| If a raid5 or raid6 array is both dirty and degraded, it could have
 | |
| undetectable data corruption.  This is because the fact that it is
 | |
| ``dirty`` means that the parity cannot be trusted, and the fact that it
 | |
| is degraded means that some datablocks are missing and cannot reliably
 | |
| be reconstructed (due to no parity).
 | |
| 
 | |
| For this reason, md will normally refuse to start such an array.  This
 | |
| requires the sysadmin to take action to explicitly start the array
 | |
| despite possible corruption.  This is normally done with::
 | |
| 
 | |
|    mdadm --assemble --force ....
 | |
| 
 | |
| This option is not really available if the array has the root
 | |
| filesystem on it.  In order to support this booting from such an
 | |
| array, md supports a module parameter ``start_dirty_degraded`` which,
 | |
| when set to 1, bypassed the checks and will allows dirty degraded
 | |
| arrays to be started.
 | |
| 
 | |
| So, to boot with a root filesystem of a dirty degraded raid 5 or 6, use::
 | |
| 
 | |
|    md-mod.start_dirty_degraded=1
 | |
| 
 | |
| 
 | |
| Superblock formats
 | |
| ------------------
 | |
| 
 | |
| The md driver can support a variety of different superblock formats.
 | |
| Currently, it supports superblock formats ``0.90.0`` and the ``md-1`` format
 | |
| introduced in the 2.5 development series.
 | |
| 
 | |
| The kernel will autodetect which format superblock is being used.
 | |
| 
 | |
| Superblock format ``0`` is treated differently to others for legacy
 | |
| reasons - it is the original superblock format.
 | |
| 
 | |
| 
 | |
| General Rules - apply for all superblock formats
 | |
| ------------------------------------------------
 | |
| 
 | |
| An array is ``created`` by writing appropriate superblocks to all
 | |
| devices.
 | |
| 
 | |
| It is ``assembled`` by associating each of these devices with an
 | |
| particular md virtual device.  Once it is completely assembled, it can
 | |
| be accessed.
 | |
| 
 | |
| An array should be created by a user-space tool.  This will write
 | |
| superblocks to all devices.  It will usually mark the array as
 | |
| ``unclean``, or with some devices missing so that the kernel md driver
 | |
| can create appropriate redundancy (copying in raid 1, parity
 | |
| calculation in raid 4/5).
 | |
| 
 | |
| When an array is assembled, it is first initialized with the
 | |
| SET_ARRAY_INFO ioctl.  This contains, in particular, a major and minor
 | |
| version number.  The major version number selects which superblock
 | |
| format is to be used.  The minor number might be used to tune handling
 | |
| of the format, such as suggesting where on each device to look for the
 | |
| superblock.
 | |
| 
 | |
| Then each device is added using the ADD_NEW_DISK ioctl.  This
 | |
| provides, in particular, a major and minor number identifying the
 | |
| device to add.
 | |
| 
 | |
| The array is started with the RUN_ARRAY ioctl.
 | |
| 
 | |
| Once started, new devices can be added.  They should have an
 | |
| appropriate superblock written to them, and then be passed in with
 | |
| ADD_NEW_DISK.
 | |
| 
 | |
| Devices that have failed or are not yet active can be detached from an
 | |
| array using HOT_REMOVE_DISK.
 | |
| 
 | |
| 
 | |
| Specific Rules that apply to format-0 super block arrays, and arrays with no superblock (non-persistent)
 | |
| --------------------------------------------------------------------------------------------------------
 | |
| 
 | |
| An array can be ``created`` by describing the array (level, chunksize
 | |
| etc) in a SET_ARRAY_INFO ioctl.  This must have ``major_version==0`` and
 | |
| ``raid_disks != 0``.
 | |
| 
 | |
| Then uninitialized devices can be added with ADD_NEW_DISK.  The
 | |
| structure passed to ADD_NEW_DISK must specify the state of the device
 | |
| and its role in the array.
 | |
| 
 | |
| Once started with RUN_ARRAY, uninitialized spares can be added with
 | |
| HOT_ADD_DISK.
 | |
| 
 | |
| 
 | |
| MD devices in sysfs
 | |
| -------------------
 | |
| 
 | |
| md devices appear in sysfs (``/sys``) as regular block devices,
 | |
| e.g.::
 | |
| 
 | |
|    /sys/block/md0
 | |
| 
 | |
| Each ``md`` device will contain a subdirectory called ``md`` which
 | |
| contains further md-specific information about the device.
 | |
| 
 | |
| All md devices contain:
 | |
| 
 | |
|   level
 | |
|      a text file indicating the ``raid level``. e.g. raid0, raid1,
 | |
|      raid5, linear, multipath, faulty.
 | |
|      If no raid level has been set yet (array is still being
 | |
|      assembled), the value will reflect whatever has been written
 | |
|      to it, which may be a name like the above, or may be a number
 | |
|      such as ``0``, ``5``, etc.
 | |
| 
 | |
|   raid_disks
 | |
|      a text file with a simple number indicating the number of devices
 | |
|      in a fully functional array.  If this is not yet known, the file
 | |
|      will be empty.  If an array is being resized this will contain
 | |
|      the new number of devices.
 | |
|      Some raid levels allow this value to be set while the array is
 | |
|      active.  This will reconfigure the array.   Otherwise it can only
 | |
|      be set while assembling an array.
 | |
|      A change to this attribute will not be permitted if it would
 | |
|      reduce the size of the array.  To reduce the number of drives
 | |
|      in an e.g. raid5, the array size must first be reduced by
 | |
|      setting the ``array_size`` attribute.
 | |
| 
 | |
|   chunk_size
 | |
|      This is the size in bytes for ``chunks`` and is only relevant to
 | |
|      raid levels that involve striping (0,4,5,6,10). The address space
 | |
|      of the array is conceptually divided into chunks and consecutive
 | |
|      chunks are striped onto neighbouring devices.
 | |
|      The size should be at least PAGE_SIZE (4k) and should be a power
 | |
|      of 2.  This can only be set while assembling an array
 | |
| 
 | |
|   layout
 | |
|      The ``layout`` for the array for the particular level.  This is
 | |
|      simply a number that is interpreted differently by different
 | |
|      levels.  It can be written while assembling an array.
 | |
| 
 | |
|   array_size
 | |
|      This can be used to artificially constrain the available space in
 | |
|      the array to be less than is actually available on the combined
 | |
|      devices.  Writing a number (in Kilobytes) which is less than
 | |
|      the available size will set the size.  Any reconfiguration of the
 | |
|      array (e.g. adding devices) will not cause the size to change.
 | |
|      Writing the word ``default`` will cause the effective size of the
 | |
|      array to be whatever size is actually available based on
 | |
|      ``level``, ``chunk_size`` and ``component_size``.
 | |
| 
 | |
|      This can be used to reduce the size of the array before reducing
 | |
|      the number of devices in a raid4/5/6, or to support external
 | |
|      metadata formats which mandate such clipping.
 | |
| 
 | |
|   reshape_position
 | |
|      This is either ``none`` or a sector number within the devices of
 | |
|      the array where ``reshape`` is up to.  If this is set, the three
 | |
|      attributes mentioned above (raid_disks, chunk_size, layout) can
 | |
|      potentially have 2 values, an old and a new value.  If these
 | |
|      values differ, reading the attribute returns::
 | |
| 
 | |
|         new (old)
 | |
| 
 | |
|      and writing will effect the ``new`` value, leaving the ``old``
 | |
|      unchanged.
 | |
| 
 | |
|   component_size
 | |
|      For arrays with data redundancy (i.e. not raid0, linear, faulty,
 | |
|      multipath), all components must be the same size - or at least
 | |
|      there must a size that they all provide space for.  This is a key
 | |
|      part or the geometry of the array.  It is measured in sectors
 | |
|      and can be read from here.  Writing to this value may resize
 | |
|      the array if the personality supports it (raid1, raid5, raid6),
 | |
|      and if the component drives are large enough.
 | |
| 
 | |
|   metadata_version
 | |
|      This indicates the format that is being used to record metadata
 | |
|      about the array.  It can be 0.90 (traditional format), 1.0, 1.1,
 | |
|      1.2 (newer format in varying locations) or ``none`` indicating that
 | |
|      the kernel isn't managing metadata at all.
 | |
|      Alternately it can be ``external:`` followed by a string which
 | |
|      is set by user-space.  This indicates that metadata is managed
 | |
|      by a user-space program.  Any device failure or other event that
 | |
|      requires a metadata update will cause array activity to be
 | |
|      suspended until the event is acknowledged.
 | |
| 
 | |
|   resync_start
 | |
|      The point at which resync should start.  If no resync is needed,
 | |
|      this will be a very large number (or ``none`` since 2.6.30-rc1).  At
 | |
|      array creation it will default to 0, though starting the array as
 | |
|      ``clean`` will set it much larger.
 | |
| 
 | |
|   new_dev
 | |
|      This file can be written but not read.  The value written should
 | |
|      be a block device number as major:minor.  e.g. 8:0
 | |
|      This will cause that device to be attached to the array, if it is
 | |
|      available.  It will then appear at md/dev-XXX (depending on the
 | |
|      name of the device) and further configuration is then possible.
 | |
| 
 | |
|   safe_mode_delay
 | |
|      When an md array has seen no write requests for a certain period
 | |
|      of time, it will be marked as ``clean``.  When another write
 | |
|      request arrives, the array is marked as ``dirty`` before the write
 | |
|      commences.  This is known as ``safe_mode``.
 | |
|      The ``certain period`` is controlled by this file which stores the
 | |
|      period as a number of seconds.  The default is 200msec (0.200).
 | |
|      Writing a value of 0 disables safemode.
 | |
| 
 | |
|   array_state
 | |
|      This file contains a single word which describes the current
 | |
|      state of the array.  In many cases, the state can be set by
 | |
|      writing the word for the desired state, however some states
 | |
|      cannot be explicitly set, and some transitions are not allowed.
 | |
| 
 | |
|      Select/poll works on this file.  All changes except between
 | |
|      Active_idle and active (which can be frequent and are not
 | |
|      very interesting) are notified.  active->active_idle is
 | |
|      reported if the metadata is externally managed.
 | |
| 
 | |
|      clear
 | |
|          No devices, no size, no level
 | |
| 
 | |
|          Writing is equivalent to STOP_ARRAY ioctl
 | |
| 
 | |
|      inactive
 | |
|          May have some settings, but array is not active
 | |
|          all IO results in error
 | |
| 
 | |
|          When written, doesn't tear down array, but just stops it
 | |
| 
 | |
|      suspended (not supported yet)
 | |
|          All IO requests will block. The array can be reconfigured.
 | |
| 
 | |
|          Writing this, if accepted, will block until array is quiessent
 | |
| 
 | |
|      readonly
 | |
|          no resync can happen.  no superblocks get written.
 | |
| 
 | |
|          Write requests fail
 | |
| 
 | |
|      read-auto
 | |
|          like readonly, but behaves like ``clean`` on a write request.
 | |
| 
 | |
|      clean
 | |
|          no pending writes, but otherwise active.
 | |
| 
 | |
|          When written to inactive array, starts without resync
 | |
| 
 | |
|          If a write request arrives then
 | |
|          if metadata is known, mark ``dirty`` and switch to ``active``.
 | |
|          if not known, block and switch to write-pending
 | |
| 
 | |
|          If written to an active array that has pending writes, then fails.
 | |
|      active
 | |
|          fully active: IO and resync can be happening.
 | |
|          When written to inactive array, starts with resync
 | |
| 
 | |
|      write-pending
 | |
|          clean, but writes are blocked waiting for ``active`` to be written.
 | |
| 
 | |
|      active-idle
 | |
|          like active, but no writes have been seen for a while (safe_mode_delay).
 | |
| 
 | |
|   bitmap/location
 | |
|      This indicates where the write-intent bitmap for the array is
 | |
|      stored.
 | |
| 
 | |
|      It can be one of ``none``, ``file`` or ``[+-]N``.
 | |
|      ``file`` may later be extended to ``file:/file/name``
 | |
|      ``[+-]N`` means that many sectors from the start of the metadata.
 | |
| 
 | |
|      This is replicated on all devices.  For arrays with externally
 | |
|      managed metadata, the offset is from the beginning of the
 | |
|      device.
 | |
| 
 | |
|   bitmap/chunksize
 | |
|      The size, in bytes, of the chunk which will be represented by a
 | |
|      single bit.  For RAID456, it is a portion of an individual
 | |
|      device. For RAID10, it is a portion of the array.  For RAID1, it
 | |
|      is both (they come to the same thing).
 | |
| 
 | |
|   bitmap/time_base
 | |
|      The time, in seconds, between looking for bits in the bitmap to
 | |
|      be cleared. In the current implementation, a bit will be cleared
 | |
|      between 2 and 3 times ``time_base`` after all the covered blocks
 | |
|      are known to be in-sync.
 | |
| 
 | |
|   bitmap/backlog
 | |
|      When write-mostly devices are active in a RAID1, write requests
 | |
|      to those devices proceed in the background - the filesystem (or
 | |
|      other user of the device) does not have to wait for them.
 | |
|      ``backlog`` sets a limit on the number of concurrent background
 | |
|      writes.  If there are more than this, new writes will by
 | |
|      synchronous.
 | |
| 
 | |
|   bitmap/metadata
 | |
|      This can be either ``internal`` or ``external``.
 | |
| 
 | |
|      ``internal``
 | |
|        is the default and means the metadata for the bitmap
 | |
|        is stored in the first 256 bytes of the allocated space and is
 | |
|        managed by the md module.
 | |
| 
 | |
|      ``external``
 | |
|        means that bitmap metadata is managed externally to
 | |
|        the kernel (i.e. by some userspace program)
 | |
| 
 | |
|   bitmap/can_clear
 | |
|      This is either ``true`` or ``false``.  If ``true``, then bits in the
 | |
|      bitmap will be cleared when the corresponding blocks are thought
 | |
|      to be in-sync.  If ``false``, bits will never be cleared.
 | |
|      This is automatically set to ``false`` if a write happens on a
 | |
|      degraded array, or if the array becomes degraded during a write.
 | |
|      When metadata is managed externally, it should be set to true
 | |
|      once the array becomes non-degraded, and this fact has been
 | |
|      recorded in the metadata.
 | |
| 
 | |
|   consistency_policy
 | |
|      This indicates how the array maintains consistency in case of unexpected
 | |
|      shutdown. It can be:
 | |
| 
 | |
|      none
 | |
|        Array has no redundancy information, e.g. raid0, linear.
 | |
| 
 | |
|      resync
 | |
|        Full resync is performed and all redundancy is regenerated when the
 | |
|        array is started after unclean shutdown.
 | |
| 
 | |
|      bitmap
 | |
|        Resync assisted by a write-intent bitmap.
 | |
| 
 | |
|      journal
 | |
|        For raid4/5/6, journal device is used to log transactions and replay
 | |
|        after unclean shutdown.
 | |
| 
 | |
|      ppl
 | |
|        For raid5 only, Partial Parity Log is used to close the write hole and
 | |
|        eliminate resync.
 | |
| 
 | |
|      The accepted values when writing to this file are ``ppl`` and ``resync``,
 | |
|      used to enable and disable PPL.
 | |
| 
 | |
|   uuid
 | |
|      This indicates the UUID of the array in the following format:
 | |
|      xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxxxxxx
 | |
| 
 | |
| 
 | |
| As component devices are added to an md array, they appear in the ``md``
 | |
| directory as new directories named::
 | |
| 
 | |
|       dev-XXX
 | |
| 
 | |
| where ``XXX`` is a name that the kernel knows for the device, e.g. hdb1.
 | |
| Each directory contains:
 | |
| 
 | |
|       block
 | |
|         a symlink to the block device in /sys/block, e.g.::
 | |
| 
 | |
| 	     /sys/block/md0/md/dev-hdb1/block -> ../../../../block/hdb/hdb1
 | |
| 
 | |
|       super
 | |
|         A file containing an image of the superblock read from, or
 | |
|         written to, that device.
 | |
| 
 | |
|       state
 | |
| 	A file recording the current state of the device in the array
 | |
| 	which can be a comma separated list of:
 | |
| 
 | |
| 	      faulty
 | |
| 			device has been kicked from active use due to
 | |
| 			a detected fault, or it has unacknowledged bad
 | |
| 			blocks
 | |
| 
 | |
| 	      in_sync
 | |
| 			device is a fully in-sync member of the array
 | |
| 
 | |
| 	      writemostly
 | |
| 			device will only be subject to read
 | |
| 			requests if there are no other options.
 | |
| 
 | |
| 			This applies only to raid1 arrays.
 | |
| 
 | |
| 	      blocked
 | |
| 			device has failed, and the failure hasn't been
 | |
| 			acknowledged yet by the metadata handler.
 | |
| 
 | |
| 			Writes that would write to this device if
 | |
| 			it were not faulty are blocked.
 | |
| 
 | |
| 	      spare
 | |
| 			device is working, but not a full member.
 | |
| 
 | |
| 			This includes spares that are in the process
 | |
| 			of being recovered to
 | |
| 
 | |
| 	      write_error
 | |
| 			device has ever seen a write error.
 | |
| 
 | |
| 	      want_replacement
 | |
| 			device is (mostly) working but probably
 | |
| 			should be replaced, either due to errors or
 | |
| 			due to user request.
 | |
| 
 | |
| 	      replacement
 | |
| 			device is a replacement for another active
 | |
| 			device with same raid_disk.
 | |
| 
 | |
| 
 | |
| 	This list may grow in future.
 | |
| 
 | |
| 	This can be written to.
 | |
| 
 | |
| 	Writing ``faulty``  simulates a failure on the device.
 | |
| 
 | |
| 	Writing ``remove`` removes the device from the array.
 | |
| 
 | |
| 	Writing ``writemostly`` sets the writemostly flag.
 | |
| 
 | |
| 	Writing ``-writemostly`` clears the writemostly flag.
 | |
| 
 | |
| 	Writing ``blocked`` sets the ``blocked`` flag.
 | |
| 
 | |
| 	Writing ``-blocked`` clears the ``blocked`` flags and allows writes
 | |
| 	to complete and possibly simulates an error.
 | |
| 
 | |
| 	Writing ``in_sync`` sets the in_sync flag.
 | |
| 
 | |
| 	Writing ``write_error`` sets writeerrorseen flag.
 | |
| 
 | |
| 	Writing ``-write_error`` clears writeerrorseen flag.
 | |
| 
 | |
| 	Writing ``want_replacement`` is allowed at any time except to a
 | |
| 	replacement device or a spare.  It sets the flag.
 | |
| 
 | |
| 	Writing ``-want_replacement`` is allowed at any time.  It clears
 | |
| 	the flag.
 | |
| 
 | |
| 	Writing ``replacement`` or ``-replacement`` is only allowed before
 | |
| 	starting the array.  It sets or clears the flag.
 | |
| 
 | |
| 
 | |
| 	This file responds to select/poll. Any change to ``faulty``
 | |
| 	or ``blocked`` causes an event.
 | |
| 
 | |
|       errors
 | |
| 	An approximate count of read errors that have been detected on
 | |
| 	this device but have not caused the device to be evicted from
 | |
| 	the array (either because they were corrected or because they
 | |
| 	happened while the array was read-only).  When using version-1
 | |
| 	metadata, this value persists across restarts of the array.
 | |
| 
 | |
| 	This value can be written while assembling an array thus
 | |
| 	providing an ongoing count for arrays with metadata managed by
 | |
| 	userspace.
 | |
| 
 | |
|       slot
 | |
|         This gives the role that the device has in the array.  It will
 | |
| 	either be ``none`` if the device is not active in the array
 | |
|         (i.e. is a spare or has failed) or an integer less than the
 | |
| 	``raid_disks`` number for the array indicating which position
 | |
| 	it currently fills.  This can only be set while assembling an
 | |
| 	array.  A device for which this is set is assumed to be working.
 | |
| 
 | |
|       offset
 | |
|         This gives the location in the device (in sectors from the
 | |
|         start) where data from the array will be stored.  Any part of
 | |
|         the device before this offset is not touched, unless it is
 | |
|         used for storing metadata (Formats 1.1 and 1.2).
 | |
| 
 | |
|       size
 | |
|         The amount of the device, after the offset, that can be used
 | |
|         for storage of data.  This will normally be the same as the
 | |
| 	component_size.  This can be written while assembling an
 | |
|         array.  If a value less than the current component_size is
 | |
|         written, it will be rejected.
 | |
| 
 | |
|       recovery_start
 | |
|         When the device is not ``in_sync``, this records the number of
 | |
| 	sectors from the start of the device which are known to be
 | |
| 	correct.  This is normally zero, but during a recovery
 | |
| 	operation it will steadily increase, and if the recovery is
 | |
| 	interrupted, restoring this value can cause recovery to
 | |
| 	avoid repeating the earlier blocks.  With v1.x metadata, this
 | |
| 	value is saved and restored automatically.
 | |
| 
 | |
| 	This can be set whenever the device is not an active member of
 | |
| 	the array, either before the array is activated, or before
 | |
| 	the ``slot`` is set.
 | |
| 
 | |
| 	Setting this to ``none`` is equivalent to setting ``in_sync``.
 | |
| 	Setting to any other value also clears the ``in_sync`` flag.
 | |
| 
 | |
|       bad_blocks
 | |
| 	This gives the list of all known bad blocks in the form of
 | |
| 	start address and length (in sectors respectively). If output
 | |
| 	is too big to fit in a page, it will be truncated. Writing
 | |
| 	``sector length`` to this file adds new acknowledged (i.e.
 | |
| 	recorded to disk safely) bad blocks.
 | |
| 
 | |
|       unacknowledged_bad_blocks
 | |
| 	This gives the list of known-but-not-yet-saved-to-disk bad
 | |
| 	blocks in the same form of ``bad_blocks``. If output is too big
 | |
| 	to fit in a page, it will be truncated. Writing to this file
 | |
| 	adds bad blocks without acknowledging them. This is largely
 | |
| 	for testing.
 | |
| 
 | |
|       ppl_sector, ppl_size
 | |
|         Location and size (in sectors) of the space used for Partial Parity Log
 | |
|         on this device.
 | |
| 
 | |
| 
 | |
| An active md device will also contain an entry for each active device
 | |
| in the array.  These are named::
 | |
| 
 | |
|     rdNN
 | |
| 
 | |
| where ``NN`` is the position in the array, starting from 0.
 | |
| So for a 3 drive array there will be rd0, rd1, rd2.
 | |
| These are symbolic links to the appropriate ``dev-XXX`` entry.
 | |
| Thus, for example::
 | |
| 
 | |
|        cat /sys/block/md*/md/rd*/state
 | |
| 
 | |
| will show ``in_sync`` on every line.
 | |
| 
 | |
| 
 | |
| 
 | |
| Active md devices for levels that support data redundancy (1,4,5,6,10)
 | |
| also have
 | |
| 
 | |
|    sync_action
 | |
|      a text file that can be used to monitor and control the rebuild
 | |
|      process.  It contains one word which can be one of:
 | |
| 
 | |
|        resync
 | |
| 		redundancy is being recalculated after unclean
 | |
|                 shutdown or creation
 | |
| 
 | |
|        recover
 | |
| 		a hot spare is being built to replace a
 | |
| 		failed/missing device
 | |
| 
 | |
|        idle
 | |
| 		nothing is happening
 | |
|        check
 | |
| 		A full check of redundancy was requested and is
 | |
|                 happening.  This reads all blocks and checks
 | |
|                 them. A repair may also happen for some raid
 | |
|                 levels.
 | |
| 
 | |
|        repair
 | |
| 		A full check and repair is happening.  This is
 | |
| 		similar to ``resync``, but was requested by the
 | |
|                 user, and the write-intent bitmap is NOT used to
 | |
| 		optimise the process.
 | |
| 
 | |
|       This file is writable, and each of the strings that could be
 | |
|       read are meaningful for writing.
 | |
| 
 | |
| 	``idle`` will stop an active resync/recovery etc.  There is no
 | |
| 	guarantee that another resync/recovery may not be automatically
 | |
| 	started again, though some event will be needed to trigger
 | |
| 	this.
 | |
| 
 | |
| 	``resync`` or ``recovery`` can be used to restart the
 | |
|         corresponding operation if it was stopped with ``idle``.
 | |
| 
 | |
| 	``check`` and ``repair`` will start the appropriate process
 | |
| 	providing the current state is ``idle``.
 | |
| 
 | |
|       This file responds to select/poll.  Any important change in the value
 | |
|       triggers a poll event.  Sometimes the value will briefly be
 | |
|       ``recover`` if a recovery seems to be needed, but cannot be
 | |
|       achieved. In that case, the transition to ``recover`` isn't
 | |
|       notified, but the transition away is.
 | |
| 
 | |
|    degraded
 | |
|       This contains a count of the number of devices by which the
 | |
|       arrays is degraded.  So an optimal array will show ``0``.  A
 | |
|       single failed/missing drive will show ``1``, etc.
 | |
| 
 | |
|       This file responds to select/poll, any increase or decrease
 | |
|       in the count of missing devices will trigger an event.
 | |
| 
 | |
|    mismatch_count
 | |
|       When performing ``check`` and ``repair``, and possibly when
 | |
|       performing ``resync``, md will count the number of errors that are
 | |
|       found.  The count in ``mismatch_cnt`` is the number of sectors
 | |
|       that were re-written, or (for ``check``) would have been
 | |
|       re-written.  As most raid levels work in units of pages rather
 | |
|       than sectors, this may be larger than the number of actual errors
 | |
|       by a factor of the number of sectors in a page.
 | |
| 
 | |
|    bitmap_set_bits
 | |
|       If the array has a write-intent bitmap, then writing to this
 | |
|       attribute can set bits in the bitmap, indicating that a resync
 | |
|       would need to check the corresponding blocks. Either individual
 | |
|       numbers or start-end pairs can be written.  Multiple numbers
 | |
|       can be separated by a space.
 | |
| 
 | |
|       Note that the numbers are ``bit`` numbers, not ``block`` numbers.
 | |
|       They should be scaled by the bitmap_chunksize.
 | |
| 
 | |
|    sync_speed_min, sync_speed_max
 | |
|      This are similar to ``/proc/sys/dev/raid/speed_limit_{min,max}``
 | |
|      however they only apply to the particular array.
 | |
| 
 | |
|      If no value has been written to these, or if the word ``system``
 | |
|      is written, then the system-wide value is used.  If a value,
 | |
|      in kibibytes-per-second is written, then it is used.
 | |
| 
 | |
|      When the files are read, they show the currently active value
 | |
|      followed by ``(local)`` or ``(system)`` depending on whether it is
 | |
|      a locally set or system-wide value.
 | |
| 
 | |
|    sync_completed
 | |
|      This shows the number of sectors that have been completed of
 | |
|      whatever the current sync_action is, followed by the number of
 | |
|      sectors in total that could need to be processed.  The two
 | |
|      numbers are separated by a ``/``  thus effectively showing one
 | |
|      value, a fraction of the process that is complete.
 | |
| 
 | |
|      A ``select`` on this attribute will return when resync completes,
 | |
|      when it reaches the current sync_max (below) and possibly at
 | |
|      other times.
 | |
| 
 | |
|    sync_speed
 | |
|      This shows the current actual speed, in K/sec, of the current
 | |
|      sync_action.  It is averaged over the last 30 seconds.
 | |
| 
 | |
|    suspend_lo, suspend_hi
 | |
|      The two values, given as numbers of sectors, indicate a range
 | |
|      within the array where IO will be blocked.  This is currently
 | |
|      only supported for raid4/5/6.
 | |
| 
 | |
|    sync_min, sync_max
 | |
|      The two values, given as numbers of sectors, indicate a range
 | |
|      within the array where ``check``/``repair`` will operate. Must be
 | |
|      a multiple of chunk_size. When it reaches ``sync_max`` it will
 | |
|      pause, rather than complete.
 | |
|      You can use ``select`` or ``poll`` on ``sync_completed`` to wait for
 | |
|      that number to reach sync_max.  Then you can either increase
 | |
|      ``sync_max``, or can write ``idle`` to ``sync_action``.
 | |
| 
 | |
|      The value of ``max`` for ``sync_max`` effectively disables the limit.
 | |
|      When a resync is active, the value can only ever be increased,
 | |
|      never decreased.
 | |
|      The value of ``0`` is the minimum for ``sync_min``.
 | |
| 
 | |
| 
 | |
| 
 | |
| Each active md device may also have attributes specific to the
 | |
| personality module that manages it.
 | |
| These are specific to the implementation of the module and could
 | |
| change substantially if the implementation changes.
 | |
| 
 | |
| These currently include:
 | |
| 
 | |
|   stripe_cache_size  (currently raid5 only)
 | |
|       number of entries in the stripe cache.  This is writable, but
 | |
|       there are upper and lower limits (32768, 17).  Default is 256.
 | |
| 
 | |
|   strip_cache_active (currently raid5 only)
 | |
|       number of active entries in the stripe cache
 | |
| 
 | |
|   preread_bypass_threshold (currently raid5 only)
 | |
|       number of times a stripe requiring preread will be bypassed by
 | |
|       a stripe that does not require preread.  For fairness defaults
 | |
|       to 1.  Setting this to 0 disables bypass accounting and
 | |
|       requires preread stripes to wait until all full-width stripe-
 | |
|       writes are complete.  Valid values are 0 to stripe_cache_size.
 | |
| 
 | |
|   journal_mode (currently raid5 only)
 | |
|       The cache mode for raid5. raid5 could include an extra disk for
 | |
|       caching. The mode can be "write-throuth" and "write-back". The
 | |
|       default is "write-through".
 | |
| 
 | |
|   ppl_write_hint
 | |
|       NVMe stream ID to be set for each PPL write request.
 |