This page introduces some of the key software provided.
Currently there are two Jupyter kernels:
To see the full list of packages and their version numbers, open a terminal and enter
conda list. The usual scientific Python packages including numpy, scipy, and matplotlib, are available. We also include some additional packages relevant to working with Swarm data. These include:
This is a programmatic interface to the VirES server, providing flexible access to Swarm data and model evaluations. Custom data sets can be built as CDF/CSV files, and directly as a pandas
DataFrameor an xarray
Swarm community software¶
- ibp (Ionospheric bubble probability)
General Python software¶
Installing other packages¶
The default environment comes from a Docker image which is based on the “Jupyter Notebook Scientific Python Stack” (jupyter/scipy-notebook) - with our own extra software and configuration layered on top. This image is loaded when when your container starts (i.e. when you log in). It will spin down when you are not using it and restart when you access it again. If you try to install other software, it may not survive the container restart, depending on how you do it. In any case, you will only be affecting your environment. Probably you will run into problems trying to install and use other software if you don’t fully understand what you are doing. In this case, if you are interested in having access to some particular software, please contact us and we can consider adding the software to the default environment.
If you install some packages with
conda install ... or
pip install ..., this will indeed install them and make them available to the default kernel. However, they will disappear the next time your container shuts down. When you have not been active for some period of time (<~hours), the container within which your environment is running will shut down. When you log in again, the environment is created anew from the image that everybody shares, so your modifications will no longer be present. This means that installing packages (& extensions) in this way is only sensible for briefly trying them out.
If you need access to a different collection of packages (e.g. you need a certain unsupported package, or different versions of what is available), then the way to achieve this is through creating an additional kernel, which you can then optionally use to execute scripts or notebooks, in place of the default kernel. This procedure is not covered in full here, but in short:
conda create --prefix <kernelname> <packages>
will create the a new conda “environment” (see the existing ones with
conda env list). For example:
conda create --prefix ~/envs/my_env ipykernel numpy
will create an environment stored within ~/envs/ and called ‘my_env’, with the packages ipykernel (this is required), and numpy. This kernel then needs to be registered with Jupyter:
~/envs/my_env/bin/python -m ipykernel install --user --name my_env --display-name "my_env"
The kernel should now be available to use from within notebooks, and you can also access the environment in a terminal with
conda activate ~/envs/my_env. Since we have stored it in the home directory, it will not be lost when the container shuts down. You can see what kernels are installed with
jupyter kernelspec list.