Setup
Before start
Suppose you already have access to BIH cluster, read the BIH HPC Docs and tried to log into the cluster successfully.
To be able to run applications with graphic user interface on the cluster, check in your terminal if echo $DISPLAY returns empty. If so, add this line export DISPLAY=:0 to the end of your shell rc file (.bashrc or .zshrc depending on which one you use).
For macOS, run brew install xquartz in your terminal to enable X11 forwarding.
Log into the cluster
ssh -Y YOUR_CHARITE_USERNAME_c@hpc-login-1.cubi.bihealth.org
Log into a compute node
srun -I60 --pty --cpus-per-task=8 --mem=32G --part=medium --time=1-0 bash
Everything below should be done in a compute node.
Create folders and soft links
ln -sfn /fast/work/groups/ag_horn group
mkdir -p $HOME/work/home/.cache
mkdir -p $HOME/work/home/.ccache
mkdir -p $HOME/work/home/.conda
mkdir -p $HOME/work/home/.config
mkdir -p $HOME/work/home/.local
mkdir -p $HOME/work/home/.singularity
ln -sfn $(realpath $HOME/work/home/.cache) $HOME/.cache
ln -sfn $(realpath $HOME/work/home/.ccache) $HOME/.ccache
ln -sfn $(realpath $HOME/work/home/.conda) $HOME/.conda
ln -sfn $(realpath $HOME/work/home/.config) $HOME/.config
ln -sfn $(realpath $HOME/work/home/.local) $HOME/.local
ln -sfn $(realpath $HOME/work/home/.singularity) $HOME/.singularityCopy preset files
cp -rT $HOME/group/setup/config $HOME
Things included:
.bash_profile,.bashrc,.zshrc.tmux.conf.condarcys-cluster.zsh-themestartup.m(MATLAB startup script)
Setup Miniforge
curl -L -O "https://github.com/conda-forge/miniforge/releases/latest/download/Miniforge3-Linux-x86_64.sh"
sh Miniforge3-Linux-x86_64.sh -b -p $HOME/work/miniforge3
rm Miniforge3-Linux-x86_64.shInstall and configure zsh and .oh-my-zsh
mamba install zsh -y
ZSH="$HOME/work/home/.oh-my-zsh" sh -c "$(curl -fsSL https://raw.githubusercontent.com/ohmyzsh/ohmyzsh/master/tools/install.sh)"
mv $HOME/.zshrc.pre-oh-my-zsh $HOME/.zshrc
mv $HOME/ys-cluster.zsh-theme $HOME/work/home/.oh-my-zsh/custom/themes/Initialize conda for bash and zsh
conda init bash zshNow open another terminal and try to log into the cluster
ssh -Y YOUR_CHARITE_USERNAME_c@hpc-login-1.cubi.bihealth.org
If you see error during login, fix it in the terminal in which you already logged into the cluster.
If everything works, you should see that you are using zsh now. From now on, you can log into a compute node and start a interactive session using:
srun -I60 --x11 --pty --cpus-per-task=8 --mem=32G --part=medium --time=1-0 --nodelist=hpc-cpu-50 zsh
where --x11, --cpus-per-task, --mem and --nodelist are optional parameters. Check out BIH HPC Docs for details.
You can also check out the BIH HPC Docs on how to use GPU nodes (for example, run CUDA version of eddy from FSL). eddy available in $FSLDIR/bin uses CUDA 11 by default. OpenMPI version is also available via eddy_openmp.
Copy MATLAB Singularity image
mkdir -p $HOME/work/container
cp $HOME/group/setup/container/matlab-ossdbs.sif $HOME/work/containerSet up Lead-DBS
getspm -P $HOME/work/MATLAB/toolbox
unzip -d $HOME/work/MATLAB/toolbox $HOME/work/MATLAB/toolbox/spm12.zip
rm $HOME/work/MATLAB/toolbox/spm12.zip
git clone https://github.com/netstim/leaddbs.git $HOME/work/MATLAB/toolbox/leaddbs
git -C $HOME/work/MATLAB/toolbox/leaddbs checkout develop
cp -r $HOME/group/setup/leaddbs_data/* $HOME/work/MATLAB/toolbox/leaddbsUse MATLAB
Log into a compute node with X11 forwarding enabled (--x11)
actmatlab# This activates MATLAB on the compute noderunmatlab# This starts MATLAB without GUIrunmatlabx# This starts MATLAB with GUI
To add SPM12 and LeadDBS to MATLAB search path:
savepath('~/work/MATLAB/pathdef.m');
addpath('~/work/MATLAB/toolbox/spm12');
addpath('~/work/MATLAB/toolbox/leaddbs');
ea_setpath;After you start LeadDBS, edit the preference file from the menu and set the two keys below:
prefs.lc.datadir = [ea_gethome, 'group', filesep, 'connectomes', filesep];
prefs.env.dev = 1;Last updated
Was this helpful?