This paper describes the performance and manageability of scalable storage systems based on Object Storage Devices (OSD). Object-based storage was invented to provide scalable performance as the storage cluster scales in size. For example, in our large file tests a 10-OSD system provided 325 MB/sec read bandwidth to 5 clients (from disk), and a 299-OSD system provided 10,334 MB/sec read bandwidth to 151 clients. This shows linear scaling of 30x speedup with 30x more client demand and 30x more storage resources. However, the system must not become more difficult to manage as it grows. Otherwise, the performance benefits can be quickly overshadowed by the administrative burden of managing the system. Instead, the storage cluster must feel like a single system image from the management perspective, even though it may be internally composed of 10’s, 100’s or thousands of object storage devices. For the HPC market, which is characterized as having unusually large clusters with usually small IT budgets, it is important that the storage system “just work” with relatively little administrative overhead. 1. Scale Out, not Scale Up The high-performance computing (HPC) sector has often driven the development of new computing architectures, and has given impetus to the development of the Object Storage Architecture. The new architecture driving change today is the Linux cluster system, which is revolutionizing scientific, technical, and business computing. The invention of Beowulf clustering and the development of the Message Passing Interface (MPI) middleware allowed racks of commodity Intel PC-based systems running the Linux operating system to emulate most of the functionality of monolithic Symmetric MultiProcessing (SMP) systems. Since this can be done at less than 10% the cost of the highly-specialized, shared memory systems, the cost of scientific research dropped dramatically. Linux clusters are now the dominant computing architecture for scientific computing, and are quickly gaining traction in technical computing environments as well.
[1]
Frank B. Schmuck,et al.
GPFS: A Shared-Disk File System for Large Computing Clusters
,
2002,
FAST.
[2]
Jim Zelenka,et al.
A cost-effective, high-bandwidth storage architecture
,
1998,
ASPLOS VIII.
[3]
Garth A. Gibson,et al.
Highly concurrent shared storage
,
2000,
Proceedings 20th IEEE International Conference on Distributed Computing Systems.
[4]
Matthew T. O'Keefe,et al.
The Global File System
,
1996
.
[5]
E. L. Miller,et al.
Efficient Metadata Management in Large Distributed File Systems
,
.
[6]
Noam Rinetzky,et al.
Towards an object store
,
2003,
20th IEEE/11th NASA Goddard Conference on Mass Storage Systems and Technologies, 2003. (MSST 2003). Proceedings..
[7]
Lustre : A Scalable , High-Performance File System Cluster
,
2003
.
[8]
Jim Zelenka,et al.
File server scaling with network-attached secure disks
,
1997,
SIGMETRICS '97.
[9]
Garth A. Gibson,et al.
Security for a high performance commodity storage subsystem
,
1999
.
[10]
Mahadev Satyanarayanan,et al.
Scale and performance in a distributed file system
,
1987,
SOSP '87.