User-friendly high-performance computing, thanks to MIT Lincoln Lab
Connecting state and local government leaders
Based on lessons learned from the development of the MIT SuperCloud, the Lincoln Laboratory Supercomputing Center developed tools and training to bring HPC capabilities to desktops and laptops.
High-performance computing systems might seem like the perfect tool for big data jobs, but they're not readily available and are notoriously difficult to use – especially for who haven't learned to write batch programming scripts for HPC systems the way engineers had been taught to do. As a result, most researchers have depended on high-end desktop systems, but today's data-intensive research means those systems can no longer deliver adequate performance.
To make it easier for researchers to access HPC resources, the Lincoln Laboratory Supercomputing Center at MIT has developed tools and training to bring HPC capabilities to desktops and laptops. The effort draws on lessons learned from the development of the MIT SuperCloud, a unified platform of four large computing ecosystems: supercomputing, enterprise computing, big data and traditional databases.
In "Lessons Learned from a Decade of Providing Interactive, On-Demand High Performance Computing to Scientists and Engineers," the authors explain how the lab addressed its goal of accommodating more and varied users who needed both performance and usability.
To meet its goal, the lab broadened the definition and architecture of an "interactive" HPC system so that users had a desktop experience in which they could essentially hit a key to run their jobs, rather than having to work with batch processing workloads.
The HPC environment itself included a uniquely configured scheduler that enabled on-demand computing services. Users were limited by the number of cores their jobs required, which ensured compute resources were always available. The scheduler and central file system were set up so that parallel-computing jobs launched in less than 20 seconds on hundreds of cores, "thereby providing the interactivity with job launches that users were used to on their desktop," the authors wrote.
A middleware layer with new user-friendly tools for launching jobs was created to support new data analysis and machine learning applications. New users were onboarded with individualized tutorials that included detailed explanations of parallelization strategies for their application.
Although "the challenges to deploying interactive on-demand HPC environments are both technical and institutional," the authors wrote, "the strategies presented here are general and easily adapted for any center where productivity and the rapid prototyping and testing of algorithms and analytics are key concerns."
Read the full paper here.
NEXT STORY: Fewer People Using Transit in the D.C. Area