I was recently interested in doing what most research groups do, setting up a computational environment that would contain version controlled software, and easy ways for users in a group to load it. There are several strategies you can take. Let’s first talk about those.
Strategies for running software on HCP
Use the system default
Yes, your home folder is located on some kind of server with an OS, and whether RHEL, CentOS, Ubuntu, or something else, it likely comes with, for example, standard python. However, you probably don’t have any kind of root access, so a standard install (let’s say we are installing the module pokemon
) like any of the following won’t work:
# If you have the module source code, with a setup.py
python setup.py install
# Install from package manager, pip
pip install pokemon
# use easy_install
easy_install pokemon
Each of the commands above would attempt to install to the system python (something like /usr/local/lib/pythonX.X/site-packages/
) and then you would get a permission denied error.
OSError: [Errno 13] Permission denied: '/usr/local/lib/python2.7/dist-packages/pokemon-0.1-py2.7.egg/EGG-INFO/entry_points.txt'
Yes, each of the commands above needs a sudo, and you aren’t sudo, so you can go home and cry about it. Or you can install to a local library with something like this:
# Install from package manager, pip, but specify as user
pip install pokemon --user
I won’t go into details, but you could also specify a –prefix to be some folder you can write to, and then add that folder to your PYTHONPATH
. This works, but it’s not ideal for a few reasons:
- if you need to capture or package your environment for sharing, you would have a hard time.
- on your
$HOME
folder, it’s likely not accessible by your labmates. This is redundant, and you can’t be sure that if they run something, they will be using the same versions of software.
Thus, what are some other options?
Use a virtual environment
Python has a fantastic thing called virtual environments, or more commonly seen as venv
. It’s actually a package that you install, create an environment for your project, and activate it:
# Install the package
pip install virtualenv --user
virtualenv myvenv
There are also ones that come prepackaged with scientific software that (normally) are quite annoying to compile like anaconda and miniconda (he’s a MINI conda! :D). And then you would install and do stuff, and your dependencies would be captured in that environment. More details and instructions can be found here. What are problems with this approach?
- It’s still REALLY redundant for each user to maintain different virtual environments
- Personally, I just forget which one is active, and then do stupid things.
For all of the above, you could use pip freeze
to generate a list of packages and versions for some requirements.txt file, or to save with your analysis for documentation sake:
pip freeze >> requirements.txt
# Inside looks like this
adium-theme-ubuntu==0.3.4
altgraph==0.12
amqp==1.4.7
aniso8601==1.1.0
anyjson==0.3.3
apptools==4.4.0
apt-xapian-index==0.45
arrow==0.7.0
artbrain==1.0.0
...
xtermcolor==1.3
zonemap==0.0.5
zope.interface==4.0.5
Use a module
Most clusters now use modules to manage versions of software and environments. What it comes down to is running a command like this:
# What versions of python are available?
module spider python
Rebuilding cache, please wait ... (written to file) done.
----------------------------------------------------------------------------
python:
----------------------------------------------------------------------------
Versions:
python/2.7.5
python/3.3.2
----------------------------------------------------------------------------
For detailed information about a specific "python" module (including how to load the modules) use the module's full name.
For example:
$ module spider python/3.3.2
----------------------------------------------------------------------------
Nice! Let’s load 2.7.5. I’m old school.
module load python/2.7.5
What basically happens, behind the scenes, is that there is a file written in a language called lua that adds folders to the beginning of your path with the particular path to the software, and possibly maps the locations as well. We can use the module software to show us this code:
# Show me the lua!
module show python/2.7.5
----------------------------------------------------------------------------
/share/sw/modules/Core/python/2.7.5:
----------------------------------------------------------------------------
conflict("python")
whatis("Provides Python 2.7.5 ")
prepend_path("version","2.7.5")
prepend_path("MANPATH","/opt/rh/python27/root/usr/share/man")
prepend_path("X_SCLS","python27")
prepend_path("LD_LIBRARY_PATH","/opt/rh/python27/root/usr/lib64")
prepend_path("LIBRARY_PATH","/opt/rh/python27/root/usr/lib64")
prepend_path("PATH","/opt/rh/python27/root/usr/bin")
prepend_path("XDG_DATA_DIRS","/opt/rh/python27/root/usr/share")
prepend_path("PKG_CONFIG_PATH","/opt/rh/python27/root/usr/lib64/pkgconfig")
prepend_path("PYTHON_INCLUDE_PATH","/opt/rh/python27/root/usr/include/python2.7")
prepend_path("CPATH","/opt/rh/python27/root/usr/include/python2.7/")
prepend_path("CPATH","/opt/rh/python27/root/usr/lib64/python2.7/site-packages/numpy/core/include/")
help([[ This module provides support for the
Python 2.7.5 via Redhat Software Collections.
]])
I won’t get into the hairy details, but this basically shows that we are adding paths (managed by an administrator) to give us access to a different version of python. This helps with versioning, but what problems do we run into?
- We still have to install additional packages using –user
- We don’t have control over any of the software configuration, we have to ask the admin
- This is specific to one research cluster, who knows if the python/2.7.5 is the same on another one. Or if it exists at all.
Again, it would work, but it’s not great. What else can we do? Well, we could try to use some kind of virtual machine… oh wait we are on a login node with no root access, nevermind. Let’s think through what we would want.
An ideal software environment
Ideally, I want all my group members to have access to it. My pokemon module version should be the same as yours. I also want total control of it. I want to be able to install whatever packages I want, and configure however I want. The first logical thing we know is that whatever we come up with, it probably is going to live in a group shared space. It also then might be handy to have equivalent lua files to load our environments, although I’ll tell you off the bat I haven’t done this yet. When I was contemplating this for my lab, I decided to try something new.
Singularity for contained software environments
A little about Singularity
We will be using Singularity containers that don’t require root privileges to run on the cluster for our environments. Further, we are going to “bootstrap” Docker images so we don’t have to start from nothing! You can think of this like packaging an entire software suite (for example, python) into a container that you can then run as an executable:
$ ./python3
Python 3.5.2 (default, Aug 31 2016, 03:01:41)
[GCC 4.9.2] on linux
Type "help", "copyright", "credits" or "license" for more information.
>>>
Even the environment gets carried through! Try this:
import os
os.environ["HOME"]
We are soon to release a new version of Singularity, and one of the simple features that I’ve been developing is an ability to immediately convert a Docker image into a Singularity image. The first iteration relied upon using the Docker Engine, but the new bootstrap does not. Because… I (finally) figured out the Docker API after many struggles, and the bootstrapping (basically starting with a Docker image as base for a Singularity image) is done using the API, sans need for the Docker engine.
As I was thinking about making a miniconda environment in a shared space for my lab, I realized - why am I not using Singularity? This is one of the main use cases, but no one seems to be doing it yet (at least as determined by the Google Group and Slack). This was my goal - to make contained environments for software (like Python) that my lab can add to their path, and use the image as an executable equivalently to calling python. The software itself, and all of the dependencies and installed modules are included inside, so if I want a truly reproducible analysis, I can just share the image. If I can’t handle about ~1GB to share, I can minimally share the file to create it, called the definition file. Let’s walk through the steps to do this. Or if you want, skip this entirely and just look at the example repo.
Singularity Environments
The basic idea is that we can generate “base” software environments for labs to use on research clusters. The general workflow is as follows:
- On your local machine (or an environment with sudo) build the contained environment
- Transfer the contained environment to your cluster
- Add the executable to your path, or create an alias.
We will first be reviewing the basic steps for building and deploying the environments.
Step 0. Setup and customize one or more environments
You will first want to clone the repository, or if you want to modify and save your definitions, fork and then clone the fork first. Here is the basic clone:
git clone https://www.github.com/radinformatics/singularity-environments
cd singularity-environments
You can then move right into building one or more containers, or optionally customize environments first.
Step 1. Build the Contained Environment
First, you should use the provided build script to generate an executable for your environment:
./build.sh python3.def
The build script is really simple - it just grabs the size (if provided), checks the number of arguments, and then creates and image and runs bootstrap (note in the future this operation will likely be one step):
#!/bin/bash
# Check that the user has supplied at least one argument
if (( "$#" < 1 )); then
echo "Usage: build.sh [image].def [options]\n"
echo "Example:\n"
echo " build.sh python.def --size 786"
exit 1
fi
def=$1
# Pop off the image name
shift
# If there are more args
if [ "$#" -eq 0 ]; then
args="--size 1024*1024B"
else
args="$@"
fi
# Continue if the image is found
if [ -f "$def" ]; then
# The name of the image is the definition file minus extension
imagefile=`echo "${def%%.*}"`
echo "Creating $imagefile using $def..."
sudo singularity create $args $imagefile
sudo singularity bootstrap $imagefile $def
fi
Note that the only two commands you really need are:
sudo singularity create $args $imagefile
sudo singularity bootstrap $imagefile $def
I mostly made the build script because I was lazy. This will generate a python3 executable in the present working directory. If you want to change the size of the container, or add any custom arguments to the Singularity bootstrap command, you can add them after your image name:
./build.sh python3.def --size 786
Note that the maximum size, if not specified, is 1024*1024BMiB. The python3.def file will need the default size to work, otherwise you run out of room and get an error. This is also true for R (r-base), which I used --size 4096
to work. That R, it’s a honkin’ package!
Step 2. Transfer the contained environment to your cluster
You are likely familiar with FTP, or hopefully your cluster uses a secure file transfer (sFTP). You can also use a command line tool scp. For the Sherlock cluster at Stanford, since I use Linux (Ubuntu), my preference is for gftp.
Step 3. Add the executable to your path
Let’s say we are working with a python3
image, and we want this executable to be called before the python3
that is installed on our cluster. We need to either add this python3
to our path (BEFORE the old one) or create an alias.
Add to your path
You likely want to add this to your .bash_profile
, .profile
, or .bashrc
:
mkdir $HOME/env
cd $HOME/env
# (and you would transfer or move your python3 here)
Now add to your .bashrc:
echo "PATH=$HOME/env:$PATH; export PATH;" >> $HOME/.bashrc
Create an alias
This will vary for different kinds of shells, but for bash you can typically do:
alias aliasname='commands'
# Here is for our python3 image
alias python3='/home/vsochat/env/python3'
For both of the above, you should test to make sure you are getting the right one when you type python3
:
which python3
/home/vsochat/env/python3
The definition files in this base directory are for base (not hugey modified) environments. But wait, what if you want to customize your environments?
I want to customize my environments (before build)!
The definition files can be modified before you create the environments! First, let’s talk a little about this Singularity definition file that we use to bootstrap.
A little about the definition file
Okay, so this folder is filled with *.def files, and they are used to create these “executable environments.” What gives? Let’s take a look quickly at a definition file:
Bootstrap: docker
From: python:3.5
%runscript
/usr/local/bin/python
%post
apt-get update
apt-get install -y vim
mkdir -p /scratch
mkdir -p /local-scratch
The first two lines might look (sort of) familiar, because “From” is a Dockerfile spec. Let’s talk about each:
- Bootstrap: is telling Singularity what kind of Build it wants to use. You could actually put some other kind of operating system here, and then you would need to provide a Mirror URL to download it. The “docker” argument tells Singularity we want to use the guts of a particular Docker image. Which one?
- From: is the argument that tells Singularity bootstrap “from this image.”
- runscript: is the one (or more) commands that are run when someone uses the container as an executable. In this case, since we want to use the python 3.5 that is installed in the Docker container, we have the executable call that path.
- post: is a bunch of commands that you want run once (“post” bootstrap), and thus this is where we do things like install additional software or packages.
Making changes
It follows logically that if you want to install additional software, do it in post! For example, you could add a pip install [something]
, and since the container is already bootstrapped from the Docker image, pip should be on the path. For example, here is how I would look around the container via python:
./python3
Python 3.5.2 (default, Aug 31 2016, 03:01:41)
[GCC 4.9.2] on linux
Type "help", "copyright", "credits" or "license" for more information.
>>> import os
>>> os.system('pip --version')
pip 8.1.2 from /usr/local/lib/python3.5/site-packages (python 3.5)
0
>>>
or using the Singularity shell command to bypass the runscript (/usr/local/bin/python) and just poke around the guts of the container:
$ singularity shell python3
Singularity: Invoking an interactive shell within container...
Singularity.python3> which pip
/usr/local/bin/pip
If you would like any additional docs on how to do things, please post an issue or just comment on this post. I’m still in the process of thinking about how to best build and leverage these environments.
I want to customize my environments! (after build)
Let’s say you have an environment (node6, for example), and you want to install a package with npm (which is located at /usr/local/bin/npm), but then when you run the image:
./node6
it takes you right into the node terminal. What gives? How do you do it? You use the Singularity shell, with write mode, and we first want to move the image back to our local machine, because we don’t have sudo on our cluster. We then want to use the writable option:
sudo singularity shell --writable node6
Singularity: Invoking an interactive shell within container...
Singularity.node6>
Then we can make our changes, and move the image back onto the cluster.
A Cool Example
The coolest example I’ve gotten working so far is using Google’s TensorFlow (the basic version without GPU - testing that next!) via a container. Here is the basic workflow:
./build tensorflow.def --size 4096
....
# building... building...
./tensorflow
Python 2.7.6 (default, Jun 22 2015, 17:58:13)
[GCC 4.8.2] on linux2
Type "help", "copyright", "credits" or "license" for more information.
>>> import tensorflow
>>>
Ok, cool! That takes us into the python installed in the image (with tensorflow), and I could run stuff interactively here. What I first tried was the “test” example, to see if it worked:
singularity shell tensorflow
python -m tensorflow.models.image.mnist.convolutional
Note that you can achieve this functionality without shelling into the image if you specify that the image should take command line arguments, something like this in the definition file:
exec /usr/local/bin/python "$@"
and then run like this!
./tensorflow -m tensorflow.models.image.mnist.convolutional
Extracting data/train-images-idx3-ubyte.gz
Extracting data/train-labels-idx1-ubyte.gz
Extracting data/t10k-images-idx3-ubyte.gz
Extracting data/t10k-labels-idx1-ubyte.gz
Initialized!
Step 0 (epoch 0.00), 6.6 ms
Minibatch loss: 12.054, learning rate: 0.010000
Minibatch error: 90.6%
Validation error: 84.6%
...
Another added feature, done specifically when I realized that there are different Docker registries, is an ability to specify the Registry and to use a Token (or not):
Bootstrap: docker
From: tensorflow/tensorflow:latest
IncludeCmd: yes
Registry: gcr.io
Token: no
Final Words
Note that this software is under development, I think the trendy way to say that is “bleeding edge,” and heck, I came up with this idea and wrote all this code most of yesterday, and so this is just an initial example to encourage others to give this a try. We don’t (yet) have a hub to store all these images, so in the meantime if you make environments, or learn something interesting, please share! I’ll definitely be adding more soon, and customizing the ones I’ve started for my lab.
Suggested Citation:
Sochat, Vanessa. "Contained Environments for Software for HPC." @vsoch (blog), 29 Sep 2016, https://vsoch.github.io/2016/singularity-environments/ (accessed 18 Nov 24).