TRIO: Burst Buffer Based I/O Orchestration

@article{Wang2015TRIOBB,
  title={TRIO: Burst Buffer Based I/O Orchestration},
  author={Teng Wang and Sarp H. Oral and Michael Pritchard and Bin Wang and Weikuan Yu},
  journal={2015 IEEE International Conference on Cluster Computing},
  year={2015},
  pages={194-203}
}
  • Teng Wang, S. Oral, Weikuan Yu
  • Published 8 September 2015
  • Computer Science
  • 2015 IEEE International Conference on Cluster Computing
The growing computing power on leadership HPC systems is often accompanied by ever-escalating failure rates. Checkpointing is a common defensive mechanism used by scientific applications for failure recovery. However, directly writing the large and bursty checkpointing dataset to parallel file systems can incur significant I/O contention on storage servers. Such contention in turn degrades bandwidth utilization of storage servers and prolongs the average job I/O time of concurrent applications… 
Toward Managing HPC Burst Buffers Effectively: Draining Strategy to Regulate Bursty I/O Behavior
TLDR
A proactive draining scheme to manage the draining process of distributed node-local burst buffers is proposed and an I/O provisioning model is developed to predict the minimized I/o provisioning requirement for permanent storage systems.
CDBB: an NVRAM-based burst buffer coordination system for parallel file systems
TLDR
Collaborative distributed burst buffer coordination system, named CDBB, coordinates all the available burst buffers, based on their priorities and states, to help overburdened burst buffers and maximize resource utilization.
BurstFS: A Distributed Burst Buffer File System for Scientific Applications
TLDR
This study proposes BurstFS, a distributed BB file system, to exploit this architecture and provide scientific applications with high and scalable performance for bursty I/O requests.
An Ephemeral Burst-Buffer File System for Scientific Applications
TLDR
This study has designed an ephemeral Burst Buffer File System (BurstFS) that supports scalable and efficient aggregation of I/O bandwidth from burst buffers while having the same life cycle as a batch-submitted job.
Data Elevator: Low-Contention Data Movement in Hierarchical Storage System
  • Bin Dong, S. Byna, N. Keen
  • Computer Science
    2016 IEEE 23rd International Conference on High Performance Computing (HiPC)
  • 2016
TLDR
This paper proposes a new system, named Data Elevator, for transparently and efficiently moving data in hierarchical storage, which reduces the resource contention on BB servers via offloading the data movement from a fixed number of BB server nodes to compute nodes.
Explorations of Data Swapping on Burst Buffer
  • T. Xu, Kento Sato, S. Matsuoka
  • Computer Science
    2018 IEEE 24th International Conference on Parallel and Distributed Systems (ICPADS)
  • 2018
TLDR
It is found that most HPC applications can still achieve full performance when using a buffer size that is far less than the total access space of the application, which can lead to a huge reduction on the required capacity for burst buffer.
Leveraging burst buffer coordination to prevent I/O interference
TLDR
This study implemented and tested several I/O scheduling policies, including prioritizing specific applications by leveraging burst buffers to defer the conflicting accesses from another application and/or directing the requests to different storage servers inside the parallel file system infrastructure.
Integration of Burst Buffer in High-level Parallel I/O Library for Exa-scale Computing Era
  • Kai-yuan Hou, Reda Al-Bahrani, W. Liao
  • Computer Science
    2018 IEEE/ACM 3rd International Workshop on Parallel Data Storage & Data Intensive Scalable Computing Systems (PDSW-DISCS)
  • 2018
TLDR
An I/O driver in PnetCDF is developed that uses a log-based format to store individual I/o requests on the burst buffer and shows that IO aggregation is a promising role for burst buffers in high-level I/ O libraries.
Harmonia: An Interference-Aware Dynamic I/O Scheduler for Shared Non-volatile Burst Buffers
TLDR
Harmonia is introduced, a new dynamic I/O scheduler that is aware of interference, adapts to the underlying system, implements a new 2-way decision-making process and employs several scheduling policies to maximize the system efficiency and applications' performance.
UniviStor: Integrated Hierarchical and Distributed Storage for HPC
TLDR
UniviStor is introduced, a data management service offering a unified view of storage layers that provides performance optimizations and data structures tailored for distributed and hierarchical data placement, interferenceaware data movement scheduling, adaptive data striping, and lightweight workflow management.
...
1
2
3
4
...

References

SHOWING 1-10 OF 42 REFERENCES
On the role of burst buffers in leadership-class storage systems
TLDR
It is shown that burst buffers can accelerate the application perceived throughput to the external storage system and can reduce the amount of external storage bandwidth required to meet a desired application perceived bottleneck goal.
IOrchestrator: Improving the Performance of Multi-node I/O Systems via Inter-Server Coordination
  • Xuechen Zhang, K. Davis, Song Jiang
  • Computer Science
    2010 ACM/IEEE International Conference for High Performance Computing, Networking, Storage and Analysis
  • 2010
TLDR
This paper proposes a scheme, IOrchestrator, to improve I/O performance of multi-node storage systems by orchestratingI/O services among programs when such inter-data-server coordination is dynamically determined to be cost effective.
Scaling parallel I/O performance through I/O delegate and caching system
  • Arifa Nisar, W. Liao, A. Choudhary
  • Computer Science
    2008 SC - International Conference for High Performance Computing, Networking, Storage and Analysis
  • 2008
TLDR
A portable MPI-IO layer is proposed where certain tasks, such as file caching, consistency control, and collective I/O optimization are delegated to a small set of compute nodes, collectively termed asI/O Delegate nodes, which alleviates the lock contention at I/o servers.
Two-Choice Randomized Dynamic I/O Scheduler for Object Storage Systems
TLDR
This paper proposes a two-choice randomized dynamic I/O scheduler that schedules the concurrent burst-write operations in a balanced way to avoid stragglers and hence achieve high throughput and designs and implements a redirect table and metadata maintainer to address the metadata management challenge introduced by dynamic I-O scheduling.
BurstMem: A high-performance burst buffer system for scientific applications
TLDR
The design of BurstMem is introduced, a high-performance burst buffer system that provides a storage framework with efficient storage and communication management strategies and is able to speed up the I/O performance of scientific applications by up to 8.5× on leadership computer systems.
Server-side I/O coordination for parallel file systems
TLDR
Experimental results demonstrate that the proposed server-side I/O coordination scheme can reduce average completion time by 8% to 46%, and provide higher I/W bandwidth than that of default data access strategies adopted by PVFS2 for heavy I/ O workloads.
Delegation-Based I/O Mechanism for High Performance Computing Systems
TLDR
A static file domain partitioning method is developed to align the I/O requests and produce a client-server mapping that minimizes the file lock acquisition costs and eliminates the lock contention.
CALCioM: Mitigating I/O Interference in HPC Systems through Cross-Application Coordination
TLDR
Experiments show how CALCioM can be used to efficiently and transparently improve the scheduling strategy between two otherwise interfering applications, given specified metrics of machine wide efficiency.
LADS: Optimizing Data Transfers Using Layout-Aware Data Scheduling
TLDR
This paper identifies the issues that lead to congestion on the path of an end-to-end data transfer in the terabit network environment, and presents a new bulk data movement framework called LADS for terabit networks.
Enhancing I/O throughput via efficient routing and placement for large-scale parallel file systems
TLDR
A method of placing I/O close to the storage nodes to minimize contention on Cray's SeaStar2+ network is described, and it is extended to a routed Lustre configuration to gain the same benefits when running against a center-wide file system.
...
1
2
3
4
5
...