Show Entry in Conferences and Workshops

You may return to the Main Menu when done.

Showing 1 entry

Nicholas Chaimov, Allen D. Malony, Shane Canon, Costin Iancu, Khaled Z. Ibrahim, Jay Srinivasan: Scaling Spark on HPC Systems. HPDC 2016: 97-110


We report our experiences porting Spark to large production HPC systems. While Spark performance in a data center installation (with local disks) is dominated by the network, our results show that file system metadata access latency can dominate in a HPC installation using Lustre: it determines single node performance up to 4⇥ slower than a typical workstation. We evaluate a combination of software techniques and hardware configurations designed to address this problem. For example, on the software side we develop a file pooling layer able to improve per node performance up to 2.8⇥. On the hardware side we evaluate a system with a large NVRAM bu↵er between compute nodes and the backend Lustre file system: this improves scaling at the expense of per-node performance. Overall, our results indicate that scalability is currently limited to O(102) cores in a HPC installation with Lustre and default Spark. After careful configuration combined with our pooling we can scale up to O(104). As our analysis indicates, it is feasible to observe much higher scalability in the near future.


Modified: Fri Jan 20 13:50:31 2017
Created: Fri Jan 20 11:56:25 2017

Current Collection: Conferences and Workshops
[ Menu | List | Show | About ]

Return to the ParaDucks Research Group Publications page.