Recent from talks
Nothing was collected or created yet.
Non-RAID drive architectures
View on WikipediaThe most widespread standard for configuring multiple hard disk drives is RAID (redundant array of inexpensive/independent disks), which comes in a number of standard configurations and non-standard configurations. Non-RAID drive architectures also exist, and are referred to by acronyms with tongue-in-cheek similarity to RAID:
- JBOD (just a bunch of disks): described multiple hard disk drives operated as individual independent hard disk drives.
- SPAN or BIG: A method of combining the free space on multiple hard disk drives from "JBoD" to create a spanned volume. Such a concatenation is sometimes also called BIG/SPAN. A SPAN or BIG is generally a spanned volume only, as it often contains mismatched types and sizes of hard disk drives.[1]
- MAID (massive array of idle drives): an architecture using hundreds to thousands of hard disk drives for providing nearline storage of data, primarily designed for write once, read occasionally (WORO) applications, in which increased storage density and decreased cost are traded for increased latency and decreased redundancy.
JBOD
[edit]JBOD (just a bunch of disks or just a bunch of drives) is an architecture using multiple hard drives exposed as individual devices. Hard drives may be treated independently or may be combined into one or more logical volumes using a volume manager like LVM or mdadm, or a device-spanning filesystem like btrfs; such volumes are usually called "spanned" or "linear | SPAN | BIG".[2][3][4] A spanned volume provides no redundancy, so failure of a single hard drive amounts to failure of the whole logical volume. Unlike a RAID 0 (striped) volume, the capacity of a linear volume is not limited by the smallest member drive multiplied by the total number of member drives, but the capacity simply adds up, however, the speed does not multiply like it does on a RAID 0.[5][6] Redundancy for resilience and/or bandwidth improvement may be provided, in software, at a higher level.
Concatenation (SPAN, BIG)
[edit]This section's factual accuracy is disputed. (December 2012) |

Concatenation or spanning of drives is not one of the numbered RAID levels, but it is a popular method for combining multiple physical disk drives into a single logical disk. It provides no data redundancy. Drives are merely concatenated together, end to beginning, so they appear to be a single large disk, known as SPAN or BIG.
In the adjacent diagram, data are concatenated from the end of disk 0 (block A63) to the beginning of disk 1 (block A64); end of disk 1 (block A91) to the beginning of disk 2 (block A92). If RAID 0 were used, then disk 0 and disk 2 would be truncated to 28 blocks, the size of the smallest disk in the array (disk 1) for a total size of 84 blocks.
What makes a SPAN or BIG different from RAID configurations is the possibility for the selection of drives. While RAID usually requires all drives to be of similar capacity[a] and it is preferred that the same or similar drive models are used for performance reasons, a spanned volume does not have such requirements.[1][7]
Implementations
[edit]The initial release of Microsoft's Windows Home Server employs drive extender technology, whereby an array of independent drives is combined by the OS to form a single pool of available storage. This storage is presented to the user as a single set of network shares. Drive extender technology expands on the normal features of concatenation by providing data redundancy through software – a shared folder can be marked for duplication, which signals to the OS that a copy of the data should be kept on multiple physical drives, whilst the user will only ever see a single instance of their data.[8] This feature was removed from Windows Home Server in its subsequent major release.[9]
The btrfs filesystem can span multiple devices of different sizes, including RAID 0/1/10 configurations, storing 1 to 4 redundant copies of both data and metadata.[10] (A flawed RAID 5/6 also exists, but can result in data loss.)[10] For RAID 1, the devices must have complementary sizes. For example, a filesystem spanning two 500 GB devices and one 1 TB device could provide RAID1 for all data, while a filesystem spanning a 1 TB device and a single 500 GB device could only provide RAID1 for 500 GB of data.
The mdadm and LVM services likewise allow combining spanning and RAID.
The ZFS filesystem can likewise pool multiple devices of different sizes and implement RAID, though it is less flexible, requiring the creation of virtual devices of fixed size on each device before pooling.[11]
In enterprise environments, enclosures are used to expand a server's data storage by using JBOD[12] devices. This is often a convenient way to scale-up storage when needed by daisy-chaining additional disk shelves.[13]
MAID
[edit]MAID (massive array of idle drives) is an architecture using hundreds to thousands of hard drives for providing nearline storage of data. MAID is designed for write once, read occasionally (WORO) applications. Hard drives are not spun-up until they are needed.[14][15][16]
Compared to RAID technology, MAID has increased storage density, and decreased cost, electrical power, and cooling requirements. However, these advantages are at the cost of much increased latency, significantly lower throughput, and decreased redundancy. Drives designed for multiple spin-up/down cycles (e.g. laptop drives) are significantly more expensive.[17] Latency may be as high as tens of seconds.[18] MAID can supplement or replace tape libraries in hierarchical storage management.[15]
To allow a more gradual tradeoff between access time and power savings, some MAIDs such as Nexsan's AutoMAID incorporate drives capable of spinning down to a lower speed.[19] Large scale disk storage systems based on MAID architectures allow dense packaging of drives and are designed to have only 25% of disks spinning at any one time.[18]
See also
[edit]Explanatory notes
[edit]- ^ Otherwise, in most cases only the drive portions equaling to the size of the smallest RAID set member would be used.
References
[edit]- ^ a b "Explanation of Spanned Volumes".
- ^ Rouse, Margaret (September 2005). "JBOD (just a bunch of disks or just a bunch of drives)". SearchStorage.TechTarget.com. TechTarget. Retrieved 2013-10-31.
- ^ Manage spanned volumes
- ^ Using spanned volumes
- ^ "LVM HOWTO, Section 3.7. mapping modes (linear/striped)". TLDP.org. Retrieved 2013-12-31.
- ^ "Linux RAID setup". Kernel.org. Linux Kernel Organization. 2013-10-05. Retrieved 2013-12-31.
- ^ Kozierok, Charles M. (April 17, 2001). "Drive Selection Criteria". The PC Guide. RAID Hard Disk Drive Requirements. Archived from the original on 2019-01-30.
- ^ "Windows Home Server Drive Extender Technical Brief". Microsoft.com. Retrieved 2009-03-12.
- ^ "Windows Home Server code name "Vail"– Update".
- ^ a b "Using Btrfs with Multiple Devices - btrfs Wiki". btrfs.wiki.kernel.org. Retrieved 2021-01-19.
- ^ "Five Years of Btrfs | MarkMcB". markmcb.com. Retrieved 2021-01-19.
- ^ "JBOD Data Storage Enclosures". ServerPartDeals.com. Retrieved 2021-04-14.
- ^ "Western Digital Ultrastar Data60 Hybrid Storage Platform SE4U60 Gen3 | 4U60 60-Bay Data Center JBOD Enclosure | Up to 1PB". ServerPartDeals.com. Retrieved 2021-04-14.
- ^ "MAID (massive array of idle disks)". TechTarget.com. January 2009. Retrieved 2013-12-31.
- ^ a b Dennis Colarelli; Dirk Grunwald (2002-07-26). "Massive Arrays of Idle Disks For Storage Archives" (PDF). University of Colorado. Retrieved 2013-12-31.
- ^ "What does WORO stand for?". TheFreeDictionary.com. Retrieved 2013-12-31.
- ^ SGI (2012). "Enterprise MAID Quick Reference Guide" (PDF). Archived from the original (PDF) on 2014-07-14. Retrieved 12 July 2014.
- ^ a b Cook, Rick (2004-07-12). "Backup budgets have it MAID with cheap disk" Retrieved on 2008-07-15
- ^ "AutoMAID Energy Saving Technology". 2011. Archived from the original on 2014-09-19. Retrieved 7 April 2011.
Further reading
[edit]- Kozierok, Charles M. (April 17, 2001). ""Just A Bunch Of Disks" (JBOD)". The PC Guide. Archived from the original on 2009-09-16.
Non-RAID drive architectures
View on GrokipediaOverview
Definition and Principles
Non-RAID drive architectures refer to configurations of multiple hard disk drives (HDDs) that operate without the redundancy, parity calculations, or data striping mechanisms characteristic of RAID systems.[6] In these setups, drives are managed either as independent units or combined linearly to extend storage capacity, prioritizing simplicity over performance optimization or fault tolerance.[7] The fundamental principles of non-RAID architectures emphasize treating drives as discrete entities or sequentially appending their spaces to form larger volumes, without distributing data across drives for load balancing or error correction. This approach relies on basic concatenation or independent access, allowing the operating system to view multiple physical drives as separate logical devices or a single extended volume, but it provides no inherent protection against single-drive failures.[6] Management is typically handled through host software, such as volume managers, or simple hardware controllers that expose drives directly to the system.[8] Concepts of non-RAID aggregation, such as concatenation, originated in the late 1980s with early volume management tools in Unix-like systems.[7] These architectures emerged more prominently in the 1990s alongside tools like the Logical Volume Manager (LVM), which was initially developed in 1998 to enable flexible storage pooling without RAID complexities.[8] Key characteristics include the absence of any built-in data redundancy, making backups essential for reliability, and dependence on software or firmware for configuration, with modern implementations extending to solid-state drives (SSDs) for similar capacity-focused applications.[6]Comparison to RAID
Non-RAID drive architectures primarily aggregate storage capacity from multiple drives without incorporating redundancy or performance-enhancing techniques like striping or mirroring, in contrast to RAID systems, which are designed to provide fault tolerance through data duplication or parity (e.g., RAID 1 mirroring or RAID 5/6 parity) and improved throughput via parallel access (e.g., RAID 0 striping).[1][9] In non-RAID setups, JBOD presents drives independently as separate volumes, while spanning concatenates data across drives in a linear fashion, sequentially filling one drive before moving to the next and maximizing usable space. Both treat the storage as vulnerable to data loss from any single drive failure, unlike RAID levels that can sustain one or more drive failures depending on the configuration.[10][11] Reliability in non-RAID architectures lacks the inherent fault tolerance of RAID, as there is no mechanism for data recovery or reconstruction if a drive fails; for instance, in a spanned volume, the failure of any single drive renders the entire logical unit inaccessible until manual intervention or backups are used, whereas independent JBOD limits loss to the failed drive.[12] This contrasts sharply with RAID 1, 5, or 6, where redundancy allows continued operation and data rebuilding post-failure.[13] Performance-wise, non-RAID configurations do not benefit from the parallel I/O operations enabled by RAID striping, resulting in read/write speeds constrained to those of individual drives rather than scaled across the array, making them unsuitable for high-throughput workloads.[11][9] Non-RAID architectures are best suited for applications prioritizing cost and simplicity over data protection, such as archival storage, backups, or cold data tiers where external redundancy (e.g., offsite copies) mitigates risks, whereas RAID is favored in environments requiring high availability and rapid access, like databases or virtualized servers.[1][13] JBOD, as a representative non-RAID example, pools capacity effectively for these low-access scenarios without the overhead of RAID controllers. As of 2025, non-RAID approaches continue to find application in budget NAS systems for home or small office use and certain cloud object storage setups emphasizing capacity over hardware-level redundancy, while RAID maintains dominance in enterprise servers for mission-critical reliability.[9][14]JBOD
Description and Operation
JBOD, or Just a Bunch Of Disks, is a non-RAID storage architecture that exposes multiple physical drives as separate logical devices directly to the operating system, allowing each drive to function independently without any data distribution or redundancy mechanisms.[2][15] In this configuration, drives are connected through a controller such as a SAS expander or a host bus adapter (HBA), which facilitates connectivity but does not perform striping, mirroring, or any other data aggregation; instead, each drive operates autonomously, with the operating system handling all data access and management.[15] During setup, JBOD drives appear as individual volumes within operating system tools, such as Windows Disk Management or Linux utilities like fdisk, enabling users to partition, format, and manage each drive separately without requiring array-level configuration.[6] This independent presentation allows for straightforward integration, as the system recognizes each drive by its unique identifier, such as /dev/sdX in Linux environments. Technically, JBOD supports heterogeneous drive sizes and types, including mixtures of HDDs and SSDs, SAS and SATA interfaces, and varying capacities, providing flexibility in deployment without uniformity constraints.[15][2] The total usable capacity equals the sum of the individual drive capacities, with no allocation for parity, mirroring, or other overhead, resulting in zero metadata footprint for array management.[2] Common hardware implementations include JBOD enclosures equipped with dual SAS modules and power supplies for scalability, as well as server backplanes that connect directly to an HBA, effectively bypassing traditional RAID controllers to pass drives through unaltered. As of 2025, high-density JBOD systems incorporate advanced technologies like Seagate's Mozaic hard drive platform for increased capacity per watt.[15][2][16] Unlike concatenation, which linearly appends data across drives into a single logical volume, JBOD maintains complete separation of drives at the hardware and OS levels.[2]Advantages and Limitations
One key advantage of JBOD is its maximum storage utilization, as it employs the full capacity of all connected drives without the overhead of redundancy schemes that waste space on parity or mirroring.[6] This approach also provides flexibility in mixing drives of varying sizes and speeds within the same system, allowing users to incorporate heterogeneous hardware without compatibility issues.[17] Additionally, JBOD incurs low costs since it requires no specialized RAID controllers or software, relying instead on standard drive interfaces.[6] Expansion is straightforward, enabling the independent addition of drives to increase capacity without disrupting existing volumes.[17] However, JBOD offers no inherent fault tolerance, meaning the failure of a single drive results in the loss of only its data but introduces management complexities in tracking and recovering individual volumes.[6] It provides no performance benefits from data parallelism, such as striping, leading to speeds limited to those of individual drives rather than aggregated throughput.[6] Without proactive balancing, uneven wear can occur across drives due to disparate usage patterns in independent operations.[6] Furthermore, managing multiple separate volumes generates administrative overhead, including separate backups and monitoring for each drive.[6] In practical terms, JBOD suits non-critical data storage where capacity trumps protection, such as archival or temporary files.[18] As of 2025, it remains common in home media servers for aggregating large media libraries, though it remains vulnerable to silent data corruption without regular backups or integrity checks.[19][20] Compared to concatenation within non-RAID setups, JBOD's treatment of drives as independent units simplifies volume management by avoiding single points of failure inherent in spanned configurations.[6] In contrast to RAID's redundancy, which trades capacity for data protection at the cost of added complexity, JBOD prioritizes simplicity and efficiency for low-risk applications.[6]Concatenation
Principles and Mechanism
Concatenation is a non-RAID technique that appends the free space of multiple physical drives sequentially to form one contiguous logical disk, enabling the creation of a larger storage volume than any single drive can provide.[21][22] In its mechanism, data writes proceed linearly by filling the entire capacity of the first drive before proceeding to the second, continuing this process across all included drives until the logical volume is full. Read operations follow a similar linear path, mapping logical offsets to the appropriate physical drive and sector based on the cumulative capacities of preceding drives.[22][23] The core principles of concatenation emphasize simplicity and capacity expansion without any data duplication, parity computation, or performance optimization through striping, making it suitable for scenarios where redundancy is handled externally. It facilitates dynamic resizing, such as adding drives to extend the volume or removing them after data migration, relying on volume management software to adjust the configuration. Metadata structures, such as extent maps in logical volume managers or extended partition tables, record drive boundaries and mappings to maintain address translation integrity across the spanned space.[24][22] Technically, concatenation accommodates unequal drive sizes by sequentially utilizing the full capacity of each drive, yielding a total logical size equal to the sum of all participating drives without wasting space on larger ones in the chain. A drive failure renders the entire logical volume inaccessible, although data on unaffected drives remains physically intact and can potentially be recovered using specialized tools or by reconfiguring the volume.[25][22][24] This approach builds briefly on JBOD concepts by unifying independent drives into a shared addressable space rather than presenting them separately.[21]Implementations
Software implementations of concatenation are prevalent in Unix-like operating systems, particularly Linux, where tools like mdadm and LVM enable linear array formation and volume spanning. The mdadm utility supports a linear mode that concatenates multiple block devices into a single logical device by sequentially appending their storage spaces, without parity or mirroring, allowing data to overflow from one device to the next as needed. Similarly, Logical Volume Manager (LVM) facilitates concatenation by combining extents from multiple physical volumes into a single logical volume, providing flexible resizing and snapshot capabilities for the spanned storage. Filesystems such as Btrfs and ZFS extend this functionality with multi-device support; Btrfs allows spanning across devices in a single profile (non-RAID) configuration, enabling snapshots and subvolumes on the concatenated storage pool. ZFS achieves similar spanning through pools composed of multiple single-disk vdevs, supporting snapshots while treating the aggregate as a unified namespace without redundancy. In Windows environments, concatenation is realized at the filesystem and disk management levels. NTFS on dynamic disks supports spanned volumes, which sequentially combine unallocated space from multiple physical disks into one logical volume, managed via Disk Management or DiskPart utilities. This approach allows transparent extension of storage capacity across heterogeneous drives. On macOS, Core Storage, deprecated since macOS High Sierra, previously supported concatenation through logical volume groups spanning multiple physical volumes, configurable via Disk Utility for creating unified storage spaces with features like snapshots. Hardware realizations of concatenation emerged in the 1990s with controller firmware supporting non-RAID modes. Adaptec's AAC-series RAID controllers featured "Volume Set" and SPAM (Spanning and Mirroring) modes, enabling firmware-level concatenation of drives into a single logical unit for bootable or data volumes. Modern enterprise JBOD enclosures from vendors like Dell and HP, such as the Dell PowerVault ME5 series and HPE D3000, support spanned modes through SAS expander firmware in passthrough (JBOD) configuration, allowing host software to concatenate drives for large-scale storage pools. These enclosures provide high-density connectivity without built-in RAID processing. A notable deprecated implementation was Windows Home Server's Drive Extender, active from 2007 to 2011, which automatically spanned multiple internal drives into a pooled storage space with optional duplication for redundancy, simplifying home NAS setups. By 2013, Microsoft discontinued it due to reliability concerns, shifting focus to ReFS-based storage spaces. As of 2025, concatenation finds adaptations in containerized and cloud environments. In Docker, multi-device volumes can be concatenated using underlying LVM or device-mapper layers for persistent storage across host disks, supporting orchestration in Kubernetes without RAID overhead. Cloud providers like AWS enable concatenation of multiple Elastic Block Store (EBS) volumes via Linux LVM on EC2 instances, offering scalable non-RAID block storage for applications requiring simple capacity extension. Similar approaches apply in Azure Managed Disks and Google Cloud Persistent Disks, where users attach and span volumes at the OS level for cost-effective, high-availability setups.MAID
Concept and Design
Massive Array of Idle Disks (MAID) is a storage architecture comprising hundreds to thousands of hard disk drives (HDDs), where the majority of drives remain powered down and spun down in an idle state until data access is required, thereby prioritizing power efficiency for large-scale, persistent data storage.[26] The design principles of MAID revolve around hierarchical data access, employing a small subset of always-active cache drives to service frequent read and write requests, while deferring less urgent operations to idle drives that spin up only on demand. This approach introduces a spin-up latency of 10-15 seconds for idle drives, balancing energy savings against acceptable delays for infrequent access patterns.[27] MAID systems often incorporate concatenation to aggregate capacity across drives linearly, forming the basis for scalable storage pools. Key architecture components include an intelligent controller that oversees drive power states, spin-up scheduling, and data migration between cache and idle tiers; integration with nearline storage environments for seamless archival handling; and a metadata subsystem to track data locations and optimize access routing across the array. MAID architectures typically scale to 100 or more drives per array, with potential for thousands in enterprise configurations, and are optimized for write-once-read-occasionally (WORO) workloads such as archival data retention.[26] The MAID concept was introduced in 2002 through research at the University of Colorado, proposing it as an energy-efficient alternative to traditional RAID for archival applications. It was subsequently commercialized in the mid-2000s by vendors including Copan Systems, with broader adoption by major providers like EMC, Hitachi Data Systems, Fujitsu, and NEC incorporating MAID principles into their storage offerings.[28][29]Applications and Benefits
MAID finds primary applications in nearline storage within data centers, where it supports the retention of infrequently accessed data such as historical records and compliance archives.[4] It is also employed for media archiving, enabling efficient storage of large volumes of video, images, and audio files that require occasional retrieval.[28] Additionally, MAID serves as a backup repository for enterprise environments, providing a disk-based alternative to tape for secondary copies of data.[30] In hierarchical storage management systems, it supplements tape libraries by offering faster access times for cold data while maintaining lower operational overhead than always-on solutions.[31] The architecture delivers significant benefits, particularly in power efficiency, achieving up to 87% reduction in energy consumption by idling the majority of drives during periods of inactivity.[32] This leads to high storage density, with systems capable of delivering terabytes per rack unit, as exemplified by configurations supporting over 1 petabyte in compact footprints.[4] Consequently, operational costs for cooling and maintenance are lowered, with annual savings estimated at thousands of dollars per large array due to reduced power draw and heat generation.[33] Unlike RAID, which prioritizes higher throughput for active workloads, MAID excels in scenarios emphasizing capacity and efficiency over speed.[34] In context, MAID's high access latency—stemming from drive spin-up times of 3 to 8 seconds—renders it unsuitable for real-time or high-performance workloads.[4] It provides no inherent redundancy, necessitating external backup mechanisms to ensure data integrity against drive failures.[4] While its relevance has somewhat declined by 2025 with the widespread adoption of SSDs for nearline roles, as solid-state drives offer lower latency and power use without mechanical idling, efforts to revive MAID continue, such as using host-managed shingled magnetic recording (HM-SMR) drives for green storage alternatives to tape.[31] Enterprise case studies highlight MAID's role in cost-effective petabyte-scale archival; for instance, Copan Systems' Revolution platform was adopted for high-density backup and archiving in data centers, enabling up to 8 petabytes in a single square meter while minimizing energy costs.[35] Similarly, Fujitsu's ETERNUS systems with MAID ECO mode have been deployed for nearline storage, yielding 20% or greater power reductions in configurations with hundreds of drives.[36] Looking ahead, MAID is evolving through hybrid HDD/SSD variants that combine SSDs for low-latency caching with idled HDDs for bulk storage, enhancing suitability for edge computing environments where space and power constraints are acute, as explored in 2025 research on tiered SSD+MAID models for improved energy and carbon efficiency.[37]References
- https://www.redbooks.[ibm](/page/IBM).com/redpapers/pdfs/redp5234.pdf
