NextPreviousContents
- Clonezilla Software Raid 0 Performance Review
- Clonezilla Software Raid 0 Performance Analysis
- Clonezilla Software Raid
Clonezilla itself is licensed under the GNU General Public License (GPL) Version 2. However, to run Clonezilla, a lot of free and open source software, e.g. Raid cloning software - EaseUS Todo Backup, file-by-file and sector-by-sector disk cloning, both Windows application and WinPE bootable disk support RAID array. Can somebody post relative performance levels for: Raid-0 7200rpm drives Raid-0 5400rpm drives Single 7200 drive Single 5400 drive With Mobos coming with Raid on board, is it still worth the extra money for two 7200rpm drives when much larger capacity can be had at 5400? As SSDs have gotten faster, especially with the advent of NVMe technology, the vast majority of users don't need to worry about RAID 0. However, there are still some niche applications where combining the speed of multiple, very fast SSDs is helpful - so in this article we are going to look at the current state of NVMe RAID solutions on a variety of modern platforms from Intel and AMD. All versions of Clonezilla live support machine with legacy BIOS. If your machine comes with uEFI secure boot enabled, you have to use AMD64 (X86-64) version (either Debian-based or Ubuntu-based) of Clonezilla live. Checksum files, not iso or zip ones, are GPG signed by DRBL project, which has the fingerprint: 54C0 821A 4871 5DAF D61B FCAF 6678. On several systems with RAID-1 setups - Clonezilla does not see the single logical drive, but instead the two raw drives. The problem is that if you proceed, backing up one of the drives, the timesstamp will be altered and a rebuild will be initiated in the best case, in the worse, both stamps will be altered and the RAID-1 image broken.
Clonezilla Software Raid 0 Performance Review
8. Performance, Tools & General Bone-headed Questions
Clonezilla Software Raid 0 Performance Analysis
- Q:I've created a RAID-0 device on
/dev/sda2
and/dev/sda3
. The device is a lot slower than a single partition. Isn't md a pile of junk?A:To have a RAID-0 device running a full speed, you must have partitions from different disks. Besides, putting the two halves of the mirror on the same disk fails to give you any protection whatsoever against disk failure.
- Q:What's the use of having RAID-linear when RAID-0 will do the same thing, but provide higher performance?
A:It's not obvious that RAID-0 will always provide better performance; in fact, in some cases, it could make things worse.The ext2fs file system scatters files all over a partition,and it attempts to keep all of the blocks of a filecontiguous, basically in an attempt to prevent fragmentation. Thus, ext2fs behaves 'as if' there were a (variable-sized) stripe per file. If there are several disks concatenatedinto a single RAID-linear, this will result files beingstatistically distributed on each of the disks. Thus,at least for ext2fs, RAID-linear will behave a lot likeRAID-0 with large stripe sizes. Conversely, RAID-0 with small stripe sizes can cause excessive disk activityleading to severely degraded performance if several large filesare accessed simultaneously.
In many cases, RAID-0 can be an obvious win. For example,imagine a large database file. Since ext2fs attempts tocluster together all of the blocks of a file, chances are good that it will end up on only one drive if RAID-linearis used, but will get chopped into lots of stripes if RAID-0 isused. Now imagine a number of (kernel) threads all tryingto random access to this database. Under RAID-linear, allaccesses would go to one disk, which would not be as efficientas the parallel accesses that RAID-0 entails.
- Q:How does RAID-0 handle a situation where the different stripe partitions are different sizes? Are the stripes uniformlydistributed?
A:To understand this, lets look at an example with threepartitions; one that is 50MB, one 90MB and one 125MB.Lets call D0 the 50MB disk, D1 the 90MB disk and D2 the 125MBdisk. When you start the device, the driver calculates 'stripzones'. In this case, it finds 3 zones, defined like this:You can see that the total size of the zones is the size of thevirtual device, but, depending on the zone, the striping isdifferent. Z2 is rather inefficient, since there's only onedisk.Since
ext2fs
and most other Unix file systems distribute files all over the disk, youhave a 35/265 = 13% chance that a fill will end upon Z2, and not get any of the benefits of striping. (DOS tries to fill a disk from beginning to end, and thus,the oldest files would end up on Z0. However, thisstrategy leads to severe filesystem fragmentation,which is why no one besides DOS does it this way.) - Q:I have some Brand X hard disks and a Brand Y controller.and am considering using
md
.Does it significantly increase the throughput?Is the performance really noticeable?A:The answer depends on the configuration that you use.
- Linux MD RAID-0 and RAID-linear performance:
If the system is heavily loaded with lots of I/O,statistically, some of it will go to one disk, andsome to the others. Thus, performance will improve over a single large disk. The actual improvementdepends a lot on the actual data, stripe sizes, andother factors. In a system with low I/O usage,the performance is equal to that of a single disk.
- Linux MD RAID-1 (mirroring) read performance:
MD implements read balancing. That is, the RAID-1code will alternate between each of the (two or more) disks in the mirror, making alternate reads to each.In a low-I/O situation, this won't change performanceat all: you will have to wait for one disk to complete the read.But, with two disks in a high-I/O environment,this could as much as double the read performance,since reads can be issued to each of the disks in parallel.For N disks in the mirror, this could improve performanceN-fold.
- Linux MD RAID-1 (mirroring) write performance:
Must wait for the write to occur to all of the disksin the mirror. This is because a copy of the datamust be written to each of the disks in the mirror.Thus, performance will be roughly equal to the writeperformance to a single disk.
- Linux MD RAID-4/5 read performance:
Statistically, a given block can be on any one of a numberof disk drives, and thus RAID-4/5 read performance isa lot like that for RAID-0. It will depend on the data, thestripe size, and the application. It will not be as goodas the read performance of a mirrored array.
- Linux MD RAID-4/5 write performance:
This will in general be considerably slower than that fora single disk. This is because the parity must be written out to one drive as well as the data to another. However,in order to compute the new parity, the old parity and the old data must be read first. The old data, new data andold parity must all be XOR'ed together to determine the newparity: this requires considerable CPU cycles in additionto the numerous disk accesses.
- Q:What RAID configuration should I use for optimal performance?
A:Is the goal to maximize throughput, or to minimize latency?There is no easy answer, as there are many factors thataffect performance:
- operating system - will one process/thread, or manybe performing disk access?
- application - is it accessing data in a sequential fashion, or random access?
- file system - clusters files or spreads them out(the ext2fs clusters together the blocks of a file,and spreads out files)
- disk driver - number of blocks to read ahead(this is a tunable parameter)
- CEC hardware - one drive controller, or many?
- hd controller - able to queue multiple requests or not?Does it provide a cache?
- hard drive - buffer cache memory size -- is it bigenough to handle the write sizes and rate you want?
- physical platters - blocks per cylinder -- accessing blocks on different cylinders will lead to seeks.
- Q:What is the optimal RAID-5 configuration for performance?
A:Since RAID-5 experiences an I/O load that is equally distributed across several drives, the best performance will be obtained when the RAID set is balanced by using identical drives, identical controllers, and thesame (low) number of drives on each controller.Note, however, that using identical components willraise the probability of multiple simultaneous failures,for example due to a sudden jolt or drop, overheating,or a power surge during an electrical storm. Mixingbrands and models helps reduce this risk.
- Q:What is the optimal block size for a RAID-4/5 array?
A:When using the current (November 1997) RAID-4/5implementation, it is strongly recommended thatthe file system be created with
mke2fs -b 4096
instead of the default 1024 byte filesystem block size.This is because the current RAID-5 implementation allocates one 4K memory page per disk block; if a disk block were just 1K in size, then 75% of the memory which RAID-5 is allocating for pending I/O would not be used. If the disk block size matches the memory page size, then the driver can (potentially) use all of the page. Thus, for a filesystem with a 4096 block size asopposed to a 1024 byte block size, the RAID driver will potentially queue 4 times as much pending I/O to the low level drivers without allocating additional memory.
Note: the above remarks do NOT apply to Software RAID-0/1/linear driver.
Note: the statements about 4K memory page size apply to the Intel x86 architecture. The page size on Alpha, Sparc, and other CPUS are different; I believe they're 8K on Alpha/Sparc (????).Adjust the above figures accordingly.
Note: if your file system has a lot of smallfiles (files less than 10KBytes in size), a considerablefraction of the disk space might be wasted. This is because the file system allocates disk space in multiples of the block size. Allocating large blocks for small files clearly results in a waste of disk space: thus, you maywant to stick to small block sizes, get a larger effectivestorage capacity, and not worry about the 'wasted' memorydue to the block-size/page-size mismatch.
Note: most 'typical' systems do not have that manysmall files. That is, although there might be thousandsof small files, this would lead to only some 10 to 100MBwasted space, which is probably an acceptable tradeoff forperformance on a multi-gigabyte disk.
However, for news servers, there might be tens or hundredsof thousands of small files. In such cases, the smallerblock size, and thus the improved storage capacity, may be more important than the more efficient I/O scheduling.
Note: there exists an experimental file system for Linuxwhich packs small files and file chunks onto a single block.It apparently has some very positive performanceimplications when the average file size is much smaller thanthe block size.
Note: Future versions may implement schemes that obsoletethe above discussion. However, this is difficult toimplement, since dynamic run-time allocation can lead todead-locks; the current implementation performs a staticpre-allocation.
- Q:How does the chunk size (stripe size) influence the speed of my RAID-0, RAID-4 or RAID-5 device?
A:The chunk size is the amount of data contiguous on the virtual device that is also contiguous on the physical device. In this HOWTO, 'chunk' and 'stripe' refer to the same thing: what is commonly called the 'stripe' in other RAID documentation is called the 'chunk' in the MD man pages. Stripes or chunks apply only toRAID 0, 4 and 5, since stripes are not used in mirroring (RAID-1) and simple concatenation (RAID-linear).The stripe size affects both read and write latency (delay),throughput (bandwidth), and contention between independent operations (ability to simultaneously service overlapping I/Orequests).
Assuming the use of the ext2fs file system, and the currentkernel policies about read-ahead, large stripe sizes are almost always better than small stripe sizes, and stripe sizesfrom about a fourth to a full disk cylinder in sizemay be best. To understand this claim, let us consider the effects of large stripes on small files, and small stripeson large files. The stripe size does not affect the read performance of small files: For anarray of N drives, the file has a 1/N probability of being entirely within one stripe on any one of the drives. Thus, both the read latency and bandwidth will be comparableto that of a single drive. Assuming that the small filesare statistically well distributed around the filesystem,(and, with the ext2fs file system, they should be), roughlyN times more overlapping, concurrent reads should be possiblewithout significant collision between them. Conversely, if very small stripes are used, and a large file is read sequentially, then a read will issued to all of the disks in the array.For a the read of a single large file, the latency will almost double, as the probability of a block being 3/4'ths of a revolution or farther away will increase. Note, however,the trade-off: the bandwidth could improve almost N-fold for reading a single, large file, as N drives can be readingsimultaneously (that is, if read-ahead is used so that allof the disks are kept active). But there is another, counter-acting trade-off: if all of the drives are already busyreading one file, then attempting to read a second or thirdfile at the same time will cause significant contention,ruining performance as the disk ladder algorithms lead toseeks all over the platter. Thus, large stripes will almostalways lead to the best performance. The sole exception isthe case where one is streaming a single, large file at a time, and one requires the top possible bandwidth, and one is also using a good read-ahead algorithm, in which case smallstripes are desired.
Note that this HOWTO previously recommended small stripesizes for news spools or other systems with lots of smallfiles. This was bad advice, and here's why: news spoolscontain not only many small files, but also large summaryfiles, as well as large directories. If the summary fileis larger than the stripe size, reading it will cause many disks to be accessed, slowing things down as eachdisk performs a seek. Similarly, the current ext2fsfile system searches directories in a linear, sequentialfashion. Thus, to find a given file or inode, on average half of the directory will be read. If this directory is spread across several stripes (several disks), the directory read (e.g. due to the ls command) could get very slow. Thanks to Steven A. Reisman <sar@pressenter.com> for this correction.Steve also adds:
I found that using a 256k stripe gives much better performance. I suspect that the optimum size would be the size of a disk cylinder (or maybe the size of the disk drive's sector cache). However, disks nowadays have recording zones with different sector counts (and sector caches vary among different disk models). There's no way to guarantee stripes won't cross a cylinder boundary.
The tools accept the stripe size specified in KBytes.You'll want to specify a multiple of if the page size for your CPU (4KB on the x86).
- Q:What is the correct stride factor to use when creating the ext2fs file system on the RAID partition? By stride, I meanthe -R flag on the
mke2fs
command:What should the value of nnn be?A:The
-R stride
flag is used to tell the file systemabout the size of the RAID stripes. Since only RAID-0,4 and 5use stripes, and RAID-1 (mirroring) and RAID-linear do not, this flag is applicable only for RAID-0,4,5.Knowledge of the size of a stripe allowsmke2fs
to allocate the block and inode bitmaps so that they don't all end up on the same physical drive. An unknown contributorwrote:I noticed last spring that one drive in a pair always had alarger I/O count, and tracked it down to the these meta-data blocks. Ted added the
For a 4KB block file system, with stripe size 256KB, one would use-R stride=
option in response to my explanation and request for a workaround.-R stride=64
.If you don't trust the
-R
flag, you can get a similareffect in a different way. Steven A. Reisman <sar@pressenter.com> writes:Another consideration is the filesystem used on the RAID-0 device.The ext2 filesystem allocates 8192 blocks per group. Each group has its own set of inodes. If there are 2, 4 or 8 drives, these inodes cluster on the first disk. I've distributed the inodes across all drives by telling mke2fs to allocate only 7932 blocks per group.
Some mke2fs pages do not describe the[-g blocks-per-group]
flag used in this operation. - Q:Where can I put the
md
commands in the startup scripts,so that everything will start automatically at boot time?A:Rod Wilkens<rwilkens@border.net>writes:
What I did is put ``
For raid-5, you will want to look at the return codeformdadd -ar
' inthe ``/etc/rc.d/rc.sysinit
' right after the kernelloads the modules, and before the ``fsck
' disk check.This way, you can put the ``/dev/md?
' device in the ``/etc/fstab
'. Then I put the ``mdstop -a
'right after the ``umount -a
' unmounting the disks,in the ``/etc/rc.d/init.d/halt
' file.mdadd
, and if it failed, do a to repair any damage. - Q:I was wondering if it's possible to setup striping with more than 2 devices in
md0
? This is for a news server,and I have 9 drives... Needless to say I need much more than two.Is this possible?A:Yes. (describe how to do this)
- Q:When is Software RAID superior to Hardware RAID?
A:Normally, Hardware RAID is considered superior to Software RAID, because hardware controllers often have a large cache,and can do a better job of scheduling operations in parallel.However, integrated Software RAID can (and does) gain certain advantages from being close to the operating system.
For example, ... ummm. Opaque description of caching of reconstructed blocks in buffer cache elided ...
On a dual PPro SMP system, it has been reported thatSoftware-RAID performance exceeds the performance of awell-known hardware-RAID board vendor by a factor of 2 to 5.
Software RAID is also a very interesting option forhigh-availability redundant server systems. In sucha configuration, two CPU's are attached to one setor SCSI disks. If one server crashes or fails to respond, then the other server can
mdadd
,mdrun
andmount
the software RAIDarray, and take over operations. This sort of dual-endedoperation is not always possible with many hardwareRAID controllers, because of the state configuration thatthe hardware controllers maintain. - Q:If I upgrade my version of raidtools, will it have trouble manipulating older raid arrays? In short, should I recreate my RAID arrays when upgrading the raid utilities?
A:No, not unless the major version number changes.An MD version x.y.z consists of three sub-versions:Version x1.y1.z1 of the RAID driver supports a RAID array withversion x2.y2.z2 in case (x1 x2) and (y1 >= y2). Different patchlevel (z) versions for the same (x.y) version aredesigned to be mostly compatible.
The minor version number is increased whenever the RAID array layoutis changed in a way which is incompatible with older versions of thedriver. New versions of the driver will maintain compatibility witholder RAID arrays.
The major version number will be increased if it will no longer makesense to support old RAID arrays in the new kernel code.
For RAID-1, it's not likely that the disk layout nor thesuperblock structure will change anytime soon. Most all Any optimization and new features (reconstruction, multithreaded tools, hot-plug, etc.) doesn't affect the physical layout.
- Q:The command
mdstop /dev/md0
says that the device is busy.A:There's a process that has a file open on
/dev/md0
, or/dev/md0
is still mounted. Terminate the process orumount /dev/md0
. - Q:Are there performance tools?
A:There is also a new utility called
iotrace
in thelinux/iotrace
directory. It reads/proc/io-trace
and analyses/plots it'soutput. If you feel your system's block IO performance is too low, just look at the iotrace output. - Q:I was reading the RAID source, and saw the value
SPEED_LIMIT
defined as 1024K/sec. What does this mean?Does this limit performance?A:
SPEED_LIMIT
is used to limit RAID reconstruction speed during automatic reconstruction. Basically, automaticreconstruction allows you toe2fsck
andmount
immediately after an unclean shutdown,without first runningckraid
. Automaticreconstruction is also used after a failed hard drivehas been replaced.In order to avoid overwhelming the system whilereconstruction is occurring, the reconstruction threadmonitors the reconstruction speed and slows it down if its too fast. The 1M/sec limit was arbitrarily chosen as a reasonable rate which allows the reconstruction tofinish reasonably rapidly, while creating only a light loadon the system so that other processes are not interfered with.
- Q:What about 'spindle synchronization' or 'disksynchronization'?
A:Spindle synchronization is used to keep multiple hard drives spinning at exactly the same speed, so that their diskplatters are always perfectly aligned. This is used by somehardware controllers to better organize disk writes. However, for software RAID, this information is not used,and spindle synchronization might even hurt performance.
- Q:How can I set up swap spaces using raid 0?Wouldn't striped swap ares over 4+ drives be really fast?
A:Leonard N. Zubkoff replies:It is really fast, but you don't need to use MD to get striped swap. The kernel automatically stripes across equal priority swap spaces. For example, the following entries from
/etc/fstab
stripe swap space across five drives inthree groups: - Q:I want to maximize performance. Should I use multiple controllers?
A:In many cases, the answer is yes. Using several controllers to perform disk access in parallel will improve performance. However, the actual improvementdepends on your actual configuration. For example,it has been reported (Vaughan Pratt, January 98) thata single 4.3GB Cheetah attached to an Adaptec 2940UW can achieve a rate of 14MB/sec (without using RAID). Installing two disks on one controller, and using a RAID-0 configuration results in a measured performance of 27 MB/sec.
Note that the 2940UW controller is an 'Ultra-Wide'SCSI controller, capable of a theoretical burst rateof 40MB/sec, and so the above measurements are not surprising. However, a slower controller attachedto two fast disks would be the bottleneck. Note also,that most out-board SCSI enclosures (e.g. the kindwith hot-pluggable trays) cannot be run at the 40MB/sec rate, due to cabling and electrical noise problems.
If you are designing a multiple controller system,remember that most disks and controllers typicallyrun at 70-85% of their rated max speeds.
Note also that using one controller per diskcan reduce the likelihood of system outagedue to a controller or cable failure (In theory --only if the device driver for the controller cangracefully handle a broken controller. Not all SCSI device drivers seem to be able to handle sucha situation without panicking or otherwise locking up).