Thanks Dustin,
Finally getting some progress on evaluating this-- wanted to reach back out to quickly update you and see if I can get more advice.
As a bit of a change in strategy, it works out better for access that I simply copy over from NERSC the ls cutout data using GLOBUS then stand up your container on my local server. I'm now at the stage where I can use many-cutouts.py on a table of objects and stream files to a scratch space (cool!). Since many-cutouts.py is a different command than you suggested above, I wanted to check with the author to make sure it was the suggested script for doing something like building a cutout database for ~16Million galaxies.
I wanted to know if there were any performance consideration in that script I should enable. Taking a read through it, there is some caching of files that occur. My plan was to take my all-sky table of spectroscopically confirmed galaxy locations, mask down to objects with RAs that should all be in the 000 brick directory, then sort my DB by declination so that hopefully I'm getting overlaps on bricks and take advantage of the cache.
Typing it all out now, I think a better strategy would be to find a way to mask and sort my DB so that I'm getting all objects that exist on a brick, which should ensure I get the cache.
Regardless, wanted to reach back out and ask if there were other performance/stability considerations? I found that I could generate about 300 cutouts in maybe 30 sec locally.