CLC number: TP311
On-line Access: 2022-04-22
Received: 2018-10-09
Revision Accepted: 2018-10-15
Crosschecked: 2018-10-15
Cited: 0
Clicked: 2970
Dhabaleswar Panda, Xiao-yi Lu, Hari Subramon. Networking and communication challenges for post-exascale systems[J]. Frontiers of Information Technology & Electronic Engineering,in press.https://doi.org/10.1631/FITEE.1800631 @article{title="Networking and communication challenges for post-exascale systems", %0 Journal Article TY - JOUR
超百亿亿级系统面临的网络和通信挑战关键词组: Darkslateblue:Affiliate; Royal Blue:Author; Turquoise:Article
Reference[1]ASCAC Subcommittee on Exascale Computing, 2010. The Opportunities and Challenges of Exascale Computing. https://science.energy.gov/media/ascr/ascac/pdf/ reports/Exascale_subcommittee_report.pdf [2]Biswas R, Lu XY, Panda DK, 2018. Accelerating tensorFlow with adaptive RDMA-based gRPC. 25th IEEE Int Conf on High Performance Computing, Data, and Analytic. [3]Cui YF, Moore R, Olsen K, et al., 2007. Enabling very-large scale earthquake simulations on parallel machines. In: Shi Y, van Albada GD, Dongarra J, et al. (Eds.), Computational Science. Springer Berlin Heidelberg, p.46-53. [4]Energy Government, 2011. Workshop on Terabits Networks for Extreme Scale Science. https://science.energy.gov/ /media/ascr/pdf/program-linebreak documents/docs/Terabit_networks_workshop_report.linebreak pdf [5]Graham RL, Bureddy D, Lui P, et al., 2016. Scalable Hierarchical Aggregation Protocol (SHArP): a hardware architecture for efficient data reduction. Proc 1st Workshop on Optimization of Communication in HPC, p.1-10. [6]Intel, 2016. Intel Omni-Path Architecture Driving Exascale Computing and HPC. https://www.intel.com/content/www/us/en/high-linebreak performance-computing-fabrics/omni-path-driving-linebreak exascale-computing.html [7]Li RZ, DeTar C, Gottlieb S, et al., 2017. MILC code performance on high end CPU and GPU supercomputer clusters. http://cn.arxiv.org/abs/1712.00143 [8]Lu XY, Shankar D, Gugnani S, et al., 2016. High-performance design of Apache Spark with RDMA and its benefits on various workloads. Proc IEEE Int Conference on Big Data, p.253-262. [9]Mellanox BlueField, 2017. Multicore System on Chip. http://www.mellanox.com/related-docs/npu-multicore-linebreak processors/PB_Bluefield_SoC.pdf [10]NVMe Express, 2016. NVMe over Fabrics. http://www.nvmexpress.org/wp-content/uploads/linebreak NVMe_Over_Fabrics.pdf [11]ORNL, 2018. Summit: America's Newest and Smartest Supercomputer. https://www.olcf.ornl.gov/summit/ [12]Rahman MWU, Lu XY, Islam NS, et al., 2014. {HOMR: a hybrid approach to exploit maximum overlapping in MapReduce over high performance interconnects}. Proc 28th ACM Int Conf on Supercomputing, p.33-42. [13]Rajachandrasekar R, Jaswani J, Subramoni H, et al., 2012. Minimizing network contention in InfiniBand clusters with a QoS-aware data-staging framework. IEEE Int Conf on Cluster Computing, p.329-336. [14]Sarvestani AMK, Bailey C, Austin J, 2018. Performance analysis of a 3D wireless massively parallel computer. J Sens Actuat Netw, 7(2):18. [15]Shankar D, Lu X, Islam N, et al., 2016. High-performance hybrid key-value store on modern clusters with RDMA interconnects and SSDs: non-blocking extensions, designs, and benefits. IEEE Int Parallel and Distributed Processing Symp, p.393-402. [16]Subramoni H, Lai P, Sur S, et al., 2010. Improving application performance and predictability using multiple virtual lanes in modern multi-core InfiniBand clusters. Int Conf on Parallel Processing. Journal of Zhejiang University-SCIENCE, 38 Zheda Road, Hangzhou
310027, China
Tel: +86-571-87952783; E-mail: cjzhang@zju.edu.cn Copyright © 2000 - 2024 Journal of Zhejiang University-SCIENCE |
Open peer comments: Debate/Discuss/Question/Opinion
<1>