wiki:rock/paper/PFS_HPC/PFS-type1
  • GPFS
    • 早期只能使用在 AIX 系統上,後來 IBM 將其移植到 Linux,早期只能使用在 IBM 的機器上,到了2005年非IBM的機器也可使用,目前只有一家 OEM 提供GPFS (linux networx)
    • it’s direct attached storage (DAS) or some type of Storage Area Network (SAN) storage. In some cases, you can combine various types of storage.
    • GPFS -> high-speed, parallel, distributed file system. GPFS achieves high-performance by striping data across multiple disks on multiple storage devices.
    • 使用三種 striping 方式:
      1. Round Robin
      2. Random
      3. Balanced Random
    • To further improve performance, GPFS uses client-side caching and deep prefetching such as read-ahead and write-behind. It recognizes standard access patterns (sequential, reverse sequential, and random) and will optimize I/O accesses for these particular patterns. Furthermore GPFS can read or write large blocks of data in a single I/O operation.
    • Block size 可選擇 16KB, 64KB, 256KB, 512KB, 1MB, and 2MB block sizes are supported with 256K being the most common
      • large block sizes helps improve performance when large data accesses are common
      • Small block sizes are used when small data accesses are common
      • GPFS subdivides the blocks into 32 sub-blocks.
      • Block is a largest chunk of contiguous data that can be accessed. A sub-block is the smallest contiguous data that can be accessed. Sub-blocks are useful for files that are smaller than a block and are stored using the sub-blocks. This can help the performance of applications that use lots of small data files (i.e. life sciences applications).
    • High Availability (HA) -> GPFS uses distributed metadata so that there is no single point of failure, nor a performance bottleneck. GPFS can be configured to use logging and replication. GPFS will log (journal) the metadata of the file system.
      • GPFS can also be configured for fail-over both at a disk level and at a server level
    • GPFS 現今依舊有在使用,in the Linux world, there are GPFS clusters with over 2,400 nodes (clients). One aspect of GPFS that should be mentioned in this context is that the GPFS is priced by the node for both I/O nodes and clients.
    • GPFS 版本:
      • 3 -> only uses TCP as the transport protocol
      • 4 -> have native IB protocols
      • In addition, the I/O nodes of GPFS can act as NFS servers if NFS is required.
    • Feature -> Multi-cluster, This allows two different GPFS file systems to be connected over a network.This is a great feature for groups in disparate locations to share data
    • last feature -> GPFS Open Source Portability Layer. The portability layer allows these GPFS kernel modules to communicate with the Linux kernel. A way to create a bridge from the GPL kernel to a non-GPL set of kernel modules, it actually serves a very useful purpose.
  • IBRIX
  • EMC MPFS
Last modified 15 years ago Last modified on Mar 6, 2009, 11:27:08 AM