LOS ALAMOS, N.M., June 7, 2017 -- At Los Alamos National Laboratory, home to more than 100 supercomputers since the dawn of the computing era, elegance and simplicity of programming are highly valued but not always achieved. In the case of a new product, dubbed "Charliecloud," a crisp 800-line code helps supercomputer users operate in the high-performance world of Big Data without burdening computer center staff with the peculiarities of their particular software needs.
"Charliecloud lets users easily run crazy new things on our supercomputers," said lead developer Reid Priedhorsky of the High Performance Computing Division at Los Alamos. "Los Alamos has lots of supercomputing power, and we do lots of simulations that are well supported here. But we've found that Big Data analysis projects need to use different frameworks, which often have dependencies that differ from what we have already on the supercomputer. So, we've developed a lightweight 'container' approach that lets users package their own user defined software stack in isolation from the host operating system."
https://eurekalert.org/pub_releases/2017-06/danl-sb060717.php
Comments
I read the paper and I think it's very nice work. The authors have a good understanding of current practices and put together a nice solution for their users, using a standard file format, built with Docker tools, but running on the nodes without using the Docker runtime. They are correct in pointing out that the Docker runtime comes with limitations that are worth avoiding.
Nice work all around, it's good to see LANL continuing to try new ideas. I met Reid when he (and I) were both much younger, and it's nice to see that he's done so well.
I also appreciate whimsical names, so I liked the name "Charlie Cloud."