What is Single Disk, JBOD, RAID 0, RAID 1, RAID 5, RAID 6, RAID 10?
Posted: 23 Mar 2021, 14:37
What is Single Disk, JBOD, RAID 0, RAID 1, RAID 5, RAID 6, RAID 10?
In the process of using TNAS, you may need to set the array mode for your hard drives. If you are not familiar with various array modes, please refer to the following content to understand the definition and working principle of various modes.
If the following definition conflicts with the information you know, in order not to prevent you from continuing to use the TNAS device, please follow the content of this article.
1. Single Disk
Single Disk: described multiple hard disk drives operated as individual independent hard disk drives. Single Disk is an architecture using multiple hard drives exposed as individual devices. Hard drives may be treated independently.
2. JBOD
JBOD (abbreviated from "Just a Bunch Of Disks"/"Just a Bunch Of Drives") is an architecture using multiple hard drives combined into one or more logical volumes using a volume manager like LVM or mdadm, or a device-spanning filesystem like btrfs; such volumes are usually called "spanned" or "linear | SPAN | BIG". A spanned volume provides no redundancy, so failure of a single hard drive amounts to failure of the whole logical volume. JBOD is not one of the numbered RAID levels, but it is a popular method for combining multiple physical disk drives into a single logical disk. It provides no data redundancy. Drives are merely concatenated together, end to beginning, so they appear to be a single large disk. It may be referred to as SPAN or BIG (meaning just the words "span" or "big", not as acronyms).
https://en.wikipedia.org/wiki/Non-RAID_ ... tures#JBOD
3. RAID 0
RAID 0 (also known as a stripe set or striped volume) splits ("stripes") data evenly across two or more disks, without parity information, redundancy, or fault tolerance. Since RAID 0 provides no fault tolerance or redundancy, the failure of one drive will cause the entire array to fail; as a result of having data striped across all disks, the failure will result in total data loss. This configuration is typically implemented having speed as the intended goal. RAID 0 is normally used to increase performance, although it can also be used as a way to create a large logical volume out of two or more physical disks. Normally, a RAID 0 setup can be created with disks of differing sizes, but the storage space added to the array by each disk is limited to the size of the smallest disk. For example, if a 1TB disk is striped together with a 3TB disk, the size of the array will be 1TB × 2 = 2TB. However, TerraMaster RAID implementations allow the remaining 2TB to be used for other purposes, this mean you will have total size of array 4TB.
https://en.wikipedia.org/wiki/Standard_ ... els#RAID_0
4. RAID 1
RAID 1 consists of an exact copy (or mirror) of a set of data on two or more disks; a classic RAID 1 mirrored pair contains two disks. This configuration offers no parity, striping, or spanning of disk space across multiple disks, since the data is mirrored on all disks belonging to the array, and the array can only be as big as the smallest member disk. This layout is useful when read performance or reliability is more important than write performance or the resulting data storage capacity. The array will continue to operate so long as at least one member drive is operational.
5. RAID 10
RAID 10, also called RAID 1+0 and sometimes RAID 1&0, is similar to RAID 01 with an exception that the two used standard RAID levels are layered in the opposite order; thus, RAID 10 is a stripe of mirrors. RAID 10, as recognized by the storage industry association and as generally implemented by RAID controllers, is a RAID 0 array of mirrors, which may be two- or three-way mirrors, and requires a minimum of four drives.
According to manufacturer specifications and official independent benchmarks, in most cases RAID 10 provides better throughput and latency than all other RAID levels except RAID 0 (which wins in throughput).Thus, it is the preferable RAID level for I/O-intensive applications such as database, email, and web servers, as well as for any other use requiring high disk performance.
https://en.wikipedia.org/wiki/Nested_RA ... ls#RAID_10
6. RAID 5
RAID 5 consists of block-level striping with distributed parity. Unlike in RAID 4, parity information is distributed among the drives. It requires that all drives but one be present to operate. Upon failure of a single drive, subsequent reads can be calculated from the distributed parity such that no data is lost. RAID 5 requires at least three disks. RAID 5's distributed parity evens out the stress of a dedicated parity disk among all RAID members. Additionally, write performance is increased since all RAID members participate in the serving of write requests. Although it will not be as efficient as a striping (RAID 0) setup, because parity must still be written, this is no longer a bottleneck.
7. RAID 6
RAID 6 extends RAID 5 by adding another parity block; thus, it uses block-level striping with two parity blocks distributed across all member disks. As in RAID 5, there are many layouts of RAID 6 disk arrays depending upon the direction the data blocks are written, the location of the parity blocks with respect to the data blocks and whether or not the first data block of a subsequent stripe is written to the same drive as the last parity block of the prior stripe. The figure to the right is just one of many such layouts
In the process of using TNAS, you may need to set the array mode for your hard drives. If you are not familiar with various array modes, please refer to the following content to understand the definition and working principle of various modes.
If the following definition conflicts with the information you know, in order not to prevent you from continuing to use the TNAS device, please follow the content of this article.
1. Single Disk
Single Disk: described multiple hard disk drives operated as individual independent hard disk drives. Single Disk is an architecture using multiple hard drives exposed as individual devices. Hard drives may be treated independently.
2. JBOD
JBOD (abbreviated from "Just a Bunch Of Disks"/"Just a Bunch Of Drives") is an architecture using multiple hard drives combined into one or more logical volumes using a volume manager like LVM or mdadm, or a device-spanning filesystem like btrfs; such volumes are usually called "spanned" or "linear | SPAN | BIG". A spanned volume provides no redundancy, so failure of a single hard drive amounts to failure of the whole logical volume. JBOD is not one of the numbered RAID levels, but it is a popular method for combining multiple physical disk drives into a single logical disk. It provides no data redundancy. Drives are merely concatenated together, end to beginning, so they appear to be a single large disk. It may be referred to as SPAN or BIG (meaning just the words "span" or "big", not as acronyms).
https://en.wikipedia.org/wiki/Non-RAID_ ... tures#JBOD
3. RAID 0
RAID 0 (also known as a stripe set or striped volume) splits ("stripes") data evenly across two or more disks, without parity information, redundancy, or fault tolerance. Since RAID 0 provides no fault tolerance or redundancy, the failure of one drive will cause the entire array to fail; as a result of having data striped across all disks, the failure will result in total data loss. This configuration is typically implemented having speed as the intended goal. RAID 0 is normally used to increase performance, although it can also be used as a way to create a large logical volume out of two or more physical disks. Normally, a RAID 0 setup can be created with disks of differing sizes, but the storage space added to the array by each disk is limited to the size of the smallest disk. For example, if a 1TB disk is striped together with a 3TB disk, the size of the array will be 1TB × 2 = 2TB. However, TerraMaster RAID implementations allow the remaining 2TB to be used for other purposes, this mean you will have total size of array 4TB.
https://en.wikipedia.org/wiki/Standard_ ... els#RAID_0
4. RAID 1
RAID 1 consists of an exact copy (or mirror) of a set of data on two or more disks; a classic RAID 1 mirrored pair contains two disks. This configuration offers no parity, striping, or spanning of disk space across multiple disks, since the data is mirrored on all disks belonging to the array, and the array can only be as big as the smallest member disk. This layout is useful when read performance or reliability is more important than write performance or the resulting data storage capacity. The array will continue to operate so long as at least one member drive is operational.
5. RAID 10
RAID 10, also called RAID 1+0 and sometimes RAID 1&0, is similar to RAID 01 with an exception that the two used standard RAID levels are layered in the opposite order; thus, RAID 10 is a stripe of mirrors. RAID 10, as recognized by the storage industry association and as generally implemented by RAID controllers, is a RAID 0 array of mirrors, which may be two- or three-way mirrors, and requires a minimum of four drives.
According to manufacturer specifications and official independent benchmarks, in most cases RAID 10 provides better throughput and latency than all other RAID levels except RAID 0 (which wins in throughput).Thus, it is the preferable RAID level for I/O-intensive applications such as database, email, and web servers, as well as for any other use requiring high disk performance.
https://en.wikipedia.org/wiki/Nested_RA ... ls#RAID_10
6. RAID 5
RAID 5 consists of block-level striping with distributed parity. Unlike in RAID 4, parity information is distributed among the drives. It requires that all drives but one be present to operate. Upon failure of a single drive, subsequent reads can be calculated from the distributed parity such that no data is lost. RAID 5 requires at least three disks. RAID 5's distributed parity evens out the stress of a dedicated parity disk among all RAID members. Additionally, write performance is increased since all RAID members participate in the serving of write requests. Although it will not be as efficient as a striping (RAID 0) setup, because parity must still be written, this is no longer a bottleneck.
7. RAID 6
RAID 6 extends RAID 5 by adding another parity block; thus, it uses block-level striping with two parity blocks distributed across all member disks. As in RAID 5, there are many layouts of RAID 6 disk arrays depending upon the direction the data blocks are written, the location of the parity blocks with respect to the data blocks and whether or not the first data block of a subsequent stripe is written to the same drive as the last parity block of the prior stripe. The figure to the right is just one of many such layouts