https://github.com/pc2/jhub-hpc-interface
JupyterHub + High-Performance Computing
Science Score: 10.0%
This score indicates how likely this project is to be science-related based on various indicators:
-
○CITATION.cff file
-
○codemeta.json file
-
○.zenodo.json file
-
○DOI references
-
○Academic publication links
-
✓Committers with academic emails
3 of 4 committers (75.0%) from academic institutions -
○Institutional organization owner
-
○JOSS paper metadata
-
○Scientific vocabulary similarity
Low similarity (10.6%) to scientific vocabulary
Keywords
Repository
JupyterHub + High-Performance Computing
Basic Info
Statistics
- Stars: 6
- Watchers: 7
- Forks: 1
- Open Issues: 0
- Releases: 0
Topics
Metadata Files
README.md
JupyterHub + High-Performance Computing
High performance Jupyter Notebooks
The aim of this project is to connect JupyterHub to a high-performance computer (HPC). By automatically offloading the computations in a Jupyter notebook to the HPC system, even complex calculations are possible. While JupyterHub is deployed on a regular server, the notebooks themselves are spawned and run on the remote HPC system using a workload manager, such as Slurm.
Motivation
The technical core of this project is the transparent integration of digital worksheets (Jupyter notebooks), in which learning content and programs can be displayed, edited and executed on the students' own laptops, with current cloud and high-performance computing (HPC) technologies. This provides the conditions for innovative, digital teaching that encourages independent and interactive development of, for example, data science applications, without imposing the complexity of using a high-performance computer system on the students. Instead, particularly computationally and data-intensive calculations are automatically offloaded to a high-performance computer, enabling even sophisticated analyses to be performed that would otherwise not be feasible on students' laptops.
Features and use cases
- Starting a jupyter notebook server on a remote HPC system in a pre-defined singularity container
- Quick config setup when using the Slurm configuration wizard
- Automatically create a singularity overlay so that user changes are persistent
- Great for managing courses with external participants
- Possibility to include files in the notebook directory using WebDAV
- Suitable for HPC users who have their own JupyterHub instance running and want to use HPC resources
Table of Contents
- JupyterHub + High-Performance Computing
- Table of Contents
- Installation of JupyterHub Server
- JupyterHub and BatchSpawner
- SSH tunnel user
- Node mapping
- Installation on HPC System
- Requirements
- Install using pip
- Singularity Container
- The configuration file
- Slurm configuration wizard
- Examples
- Debug mode
- Shibboleth Integration
- NBGrader Integration
- Installation
- Changing the Student ID to the JupyterHub logged in user name
- Create nbgrader_config.py
- Security Precautions
- Singularity Host Filesystems
- JupyterHub API (HTTPS)
- tunnelbot user
- Troubleshooting
Installation of JupyterHub Server
This section describes the required installations and configurations on the JupyterHub server.
JupyterHub and BatchSpawner
The first thing you should do is install JupyterHub and BatchSpawner. For this purpose we provide an Ansible playbook which can be found in /jupyterhub-deployment/. See the README for details. Alternatively, you can follow the official installation instructions.
If you decide to do the installations yourself, please proceed as follows:
- install JupyterHub
- install BatchSpawner
- install WrapSpawner (make sure to install it in the right environment:
/opt/jupyterhub/bin/pip3 install git+https://github.com/jupyterhub/wrapspawner) - copy the JupyterHub configuration file
/jupyterhub-deployment/config_files/jupyterhub_config.pyto/opt/jupyterhub/etc/jupyterhub/(you will most likely have to edit this file afterwards to make it fit your needs) - restart the JupyterHub service
SSH tunnel user
A user called tunnelbot is needed on the JupyterHub server. This user is responsible for starting an SSH tunnel between the compute node and the JupyterHub server. An SSH key pair for the above mentioned purpose must be generated. See /examples/jupyterhub_config.py for more information.
Node mapping
JupyterHub extracts the execution host name of the HPC system (e.g. node01-002). When a notebook server is started, an SSH tunnel is established using the notebook port.
In order for JupyterHub to be able to resolve the compute nodes host name, the /etc/hosts file must be edited. An example entry might look like the following:
127.0.0.1 node01-001
127.0.0.1 node01-002
127.0.0.1 node01-003
...
127.0.0.1 node12-048
The actual node names depend on your HPC system of course.
Installation on HPC System
This section describes the required installations and configurations of the HPC system to enable the interaction with the JuypterHub server.
Requirements
- You need a user who is allowed to allocate resources on the HPC system
- With a SSH key pair. The public part must be deposited on the JupyterHub serer (
tunnelbotuser) - The public key part of the
tunnelbot-user created on the JupyterHub (-> ~/.ssh/authorizedkeys_)
- With a SSH key pair. The public part must be deposited on the JupyterHub serer (
- Singularity (> v.3.7.0)
- mkfs/e2fsprogs with following option:
- https://git.kernel.org/pub/scm/fs/ext2/e2fsprogs.git/commit/?id=217c0bdf17899c0f79b73f76feeadd6d55863180
Install using pip
You can download and install the required files with pip.
You may want to build a small Python environment, or install the tools with --user.
bash
python3 -m pip install --user jh-hpc-interface
Singularity Container
Singularity recipe examples are in the directory singularity/.
If you do not want to use singularity, then change the value of use_singularity in jh_config.ini to false.
Build Singularity Container
To build the container with the recipe files in singularity/ you have to clone this repository.
The following commands replace USER_ID in the recipes to the output of id -u, create a new hidden file and build the singularity container with the new created file.
Compute
bash
USER_ID=$(id -u) && sed "s/USER_ID/$USER_ID/" < singularity/Singularity > singularity/.recipefile_compute && singularity build --remote singularity/compute_jupyter.sif singularity/.recipefile_compute
GPU (Tensorflow)
bash
USER_ID=$(id -u) && sed "s/USER_ID/$USER_ID/" < singularity/Singularity_Tensorflow > singularity/.recipefile_gpu && singularity build --remote singularity/gpu_jupyter.sif singularity/.recipefile_gpu
singularity build help section:
-r, --remote build image remotely (does not require root)
Please refer to the official docs on how to use the remote build feature: https://sylabs.io/docs/
The configuration file
In the directory bin/ is a script, which is deposited after the installation on the system.
With the following call you can display the location of the configuration file:
bash
$ jh_wrapper getconfig
To learn more about the configuration file, see docs/jh_config.ini.md
Slurm configuration wizard
With the configuration wizard you can prepare your HPC environment.
The script interactively goes through the configuration file and creates a temporary file which can be copied with a simple cp.
To start the wizard type the following:
bash
$ jh_slurm_wizard
Examples
You will find examples for the configuration files jh_config.ini and jupyterhub_config.py in the directory examples/.
Debug mode
By default the logs contain only information such as warnings or error messages. It is also possible to switch on the debug mode, which writes extended information into the log files.
Just set log_level in the configuration file to 'DEBUG'.
Shibboleth Integration
Shibboleth authentication was set up for a JupyterHub server in a test environment. See ./shibboleth/ for an example configuration.
NBGrader Integration
Installation
Installation instructions: https://nbgrader.readthedocs.io/en/latest/configuration/jupyterhub_config.html
To create an exchange directory for every user, just create an empty directory in $scratch_dir and mount it into the container with $singularity_bind_extra.
Changing the Student ID to the JupyterHub logged in user name
Since the containers run as user jovyan, the value from the $JUPYTERHUB_USER variable is automatically used.
See here for more information: https://jupyter.readthedocs.io/en/latest/community/content-community.html#what-is-a-jovyan
Create nbgrader_config.py
See here: https://nbgrader.readthedocs.io/en/stable/configuration/nbgrader_config.html#use-case-3-nbgrader-and-jupyterhub
To make nbgraderconfig.py_ available in the container, just append the file in $singularity_bind_extra.
Security Precautions
Singularity Host Filesystems
In case you are using Singularity, the host file system may be automatically mounted into the container when you start a Singularity Container.
A possible cause is the option mount hostfs in singularity.conf
See here: https://sylabs.io/guides/3.5/admin-guide/configfiles.html#singularity-conf
JupyterHub API (HTTPS)
HTTPS
See here for more information: https://jupyterhub.readthedocs.io/en/stable/reference/websecurity.html
tunnelbot user
You can increase the security by deactivating shell access for this user.
Just type:
bash
usermod -s /bin/false tunnelbot
Troubleshooting
When problems occur with the JupyterHub, some information can be obtained from the logs when debug mode is enabled:
https://github.com/jupyterhub/jupyterhub/wiki/Debug-Jupyterhub
Owner
- Name: Paderborn Center for Parallel Computing
- Login: pc2
- Kind: organization
- Location: Paderborn, Germany
- Website: http://pc2.uni-paderborn.de
- Repositories: 72
- Profile: https://github.com/pc2
GitHub Events
Total
- Watch event: 1
Last Year
- Watch event: 1
Committers
Last synced: 7 months ago
Top Committers
| Name | Commits | |
|---|---|---|
| Marcel-Brian Wilkowsky | m****y@u****e | 33 |
| feberts | 6****s | 16 |
| Fabian Eberts | e****n@f****e | 4 |
| Christian Plessl | c****l@u****e | 1 |
Committer Domains (Top 20 + Academic)
Issues and Pull Requests
Last synced: 7 months ago
All Time
- Total issues: 0
- Total pull requests: 1
- Average time to close issues: N/A
- Average time to close pull requests: less than a minute
- Total issue authors: 0
- Total pull request authors: 1
- Average comments per issue: 0
- Average comments per pull request: 0.0
- Merged pull requests: 1
- Bot issues: 0
- Bot pull requests: 0
Past Year
- Issues: 0
- Pull requests: 0
- Average time to close issues: N/A
- Average time to close pull requests: N/A
- Issue authors: 0
- Pull request authors: 0
- Average comments per issue: 0
- Average comments per pull request: 0
- Merged pull requests: 0
- Bot issues: 0
- Bot pull requests: 0
Top Authors
Issue Authors
Pull Request Authors
- mawigh (1)
Top Labels
Issue Labels
Pull Request Labels
Packages
- Total packages: 1
-
Total downloads:
- pypi 12 last-month
- Total dependent packages: 0
- Total dependent repositories: 1
- Total versions: 4
- Total maintainers: 1
pypi.org: jh-hpc-interface
JupyterHub + High-Performace Computing
- Homepage: https://github.com/pc2/JHub-HPC-Interface
- Documentation: https://jh-hpc-interface.readthedocs.io/
- License: MIT License
-
Latest release: 1.3
published almost 5 years ago
Rankings
Maintainers (1)
Dependencies
- batchspawner *
- notebook *