Source: Huawei data storage FusionCube hyper-converged infrastructure 2024 special issue
from virtualization to data center virtualization
so far, the virtualization we are talking about only focuses on the virtualization of computing resources, which can be called computing virtualization or server virtualization. However, virtualization is not only aimed at server resources.
From the perspective of a data center, virtualized objects should include computing, storage, and network resources. Computing virtualization has been discussed in the first section. Storage virtualization can be traced back to the 1980 s. By Professor Patterson of the University of California, Berkeley, RAID(Redundant Array of Independent Disks) was proposed in the paper "A Case of Redundant Array of Redundant Disks". Technology. RAID is originally a technology to improve reliability, that is, combining multiple independent physical hard disks in different ways to form a hard disk Group, in order to provide higher storage performance and data redundancy technology than a single hard disk, however, the RAID technology not only improves the reliability, but also realizes the abstraction of resources, users do not need to pay attention to which hard disk their data is stored, but the disk array provides a series of LUN(Logic Unit) for users to use. From physical hard disks to logical hard disks, RAID technology has blown the wind of virtualization into storage.
In traditional enterprise storage, one of the main functions of the so-called storage controller is to virtualize the disk array and provide LUN for the host. With the increase of data volume, especially the explosion of massive unstructured data, and the higher requirements for the capacity and scalability of storage systems, distributed storage emerges as the times require. Unlike traditional enterprise storage, distributed storage does not have a dedicated controller, but uses distributed software to combine hard disks in server nodes into a storage pool, and then uses virtual volumes to the host. The implementation principle is different, but the final result is that multiple hard disks are virtualized into a storage pool and provided to users as LUNs or volumes. With the development of data centers, higher requirements are put forward for network virtualization: first, how to build a virtual network within servers, between different virtual machines, between containers and containers, to realize exchange, network functions such as isolation and forwarding. The second is that the servers form a cluster. How can the physical network in the cluster be separated from the logical network to meet the characteristics of multi-tenant access and on-demand service, while having high scalability. in response to these challenges, Professor Nick McKeown of Stanford University put forward the concept of OpenFlow in 2008 and published the classic paper OpenFlow:Enabling Innovation in Campus Networks. The Software- Defined Network (SDN) was boldly conceived, opening a new chapter in Network virtualization.
The successful distributed storage includes VMware's vSAN,EMC's Isilon and Huawei's OceanStor Pacific, while the open source has the famous Ceph. Due to its low threshold, it has gradually become the mainstream of storage virtualization, therefore, storage virtualization mentioned in many scenarios is equivalent to distributed storage, especially in hyper-converged scenarios.