Line: 1 to 1 | ||||||||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
NWZPHI the cluster of the IVV 4 | ||||||||||||||||
Line: 105 to 105 | ||||||||||||||||
| ||||||||||||||||
Changed: | ||||||||||||||||
< < |
| |||||||||||||||
> > |
| |||||||||||||||
To choose the node type that you want to use, you have to use the correct queue. So if you want to run a large computation which needs more than 128 GB of RAM for a single process, the dl560 is right for you. In this case, you have to use the bigsmp queue: |
Line: 1 to 1 | ||||||||
---|---|---|---|---|---|---|---|---|
NWZPHI the cluster of the IVV 4 | ||||||||
Line: 142 to 142 | ||||||||
If you want to use the accelerators, you always have to reserver three times more CPU cores than accelerators like: | ||||||||
Changed: | ||||||||
< < | qsub -I -q test -l nodes=1:ppn=9:mics=3 | |||||||
> > | qsub -I -q default -l nodes=1:ppn=9:mics=3 | |||||||
You have to recompile your code (with the Intel compiler) with the "-mmic" Flag, so you have to create a separate version for the host and the accelerator. |
Line: 1 to 1 | ||||||||
---|---|---|---|---|---|---|---|---|
NWZPHI the cluster of the IVV 4 | ||||||||
Line: 19 to 19 | ||||||||
NWZPHI for the impatient reader | ||||||||
Changed: | ||||||||
< < | The name of the login-server is NWZPHI. Allowed are all users that are members of the group u0clustr and at least one of the groups starting with p0, q0 or r0. In addition, every user allowed for PALMA may use NWZPHI. You can register yourself for u0clstr at ZIV.MeinZIV![]() | |||||||
> > | The name of the login-server is NWZPHI. Allowed are all users that are members of the group u0clustr and at least one of the groups starting with p0, q0 or r0. In addition, every user allowed for PALMA may use NWZPHI. You can register yourself for u0clstr at MeinZIV![]() | |||||||
The batch and module system are working very similar to PALMA. | ||||||||
Line: 49 to 49 | ||||||||
| ||||||||
Changed: | ||||||||
< < | To use the same modules at every login, put the commands in your $HOME/.bashrc. Recommended modules are | |||||||
> > | To use the same modules at every login, put the commands in your $HOME/.bashrc. The recommended default module is | |||||||
Changed: | ||||||||
< < | module add compiler/intel/14
module add mpi/intel/4.1.0.024
module add tools/mic
Example: Compile a program that uses the FFTW:
module add compiler/intel/14 module add mpi/intel/4.1.0.024 module add lib/fftw/intel/3.3.3 ${MPIICC} -I ${FFTW_INCLUDE_DIR} -o program program.c -g ${FLAGS_FAST} -L${FFTW_LIB_DIR} -lfftw_mpi -lfftw -lm | |||||||
> > | module add intel/2016a This is a toolchain that loads other modules like (Intel-) MPI and the MKL. | |||||||
Batch system | ||||||||
Line: 147 to 138 | ||||||||
StorageThere is a 88 TB partition for /home and /scratch using the BeeGFS![]() | ||||||||
Changed: | ||||||||
< < | Using Xeon Phi Accelerators | |||||||
> > | Using the Xeon Phi accelerators | |||||||
Changed: | ||||||||
< < | to be done... | |||||||
> > | If you want to use the accelerators, you always have to reserver three times more CPU cores than accelerators like:
qsub -I -q test -l nodes=1:ppn=9:mics=3 | |||||||
Changed: | ||||||||
< < | Most of the computational power of NWZPHI stems from the accelerators. To have some numbers: The 24 CPU cores of one node deliver some hundred GFLOPS, a single Xeon PHI accelerator has a peak performance of 1 TFLOP. | |||||||
> > | You have to recompile your code (with the Intel compiler) with the "-mmic" Flag, so you have to create a separate version for the host and the accelerator. | |||||||
Changed: | ||||||||
< < | You can reserve the Xeon Phi Cards via the batch system. This would be done the following way:
#PBS -l nodes=1:ppn=24:mics=8 | |||||||
> > | mpiicpc code.c -mmic -o program.mic | |||||||
Changed: | ||||||||
< < | Since the operating system cannot see, which cards are already in use, it is a good idea to reserve complete nodes with all of its cards and distribute the jobs by yourself. | |||||||
> > | To use the accelerators that have been reserved for you, you can use the script "allocated-mics.pl" which is in your PATH. The host names of the cards are no longer mic0, mic1..., but have the name of their hosts in it so this would be sl270-01-mic0, sl270-01-mic1 and so on. This is necessary to set up the communication between the accelerators. | |||||||
Changed: | ||||||||
< < | When you submit a job, there will be a variable PBS_MICFILE inside your job. In addition to the PBS_NODEFILE, in PBS_MICFILE there will be a list of the Xeon PHIs reserved for you. | |||||||
> > | An example how to use the accelerators with MPI could look like this:
allocated-mics.pl > ${HOME}/mics.listEach card has 60 cores and can have up to four threads per core, so all in all 240 threads per card can be created. But from my experience it might be better to use only 120 threads per card. | |||||||
Deleted: | ||||||||
< < | To use the Infiniband interfaces, append a "-ib" to the hostnames:
cat $PBS_MICFILE | awk '{print $0"-ib"}' > micfile-ib.$PBS_JOBID | |||||||
Some useful links | ||||||||
Line: 175 to 173 | ||||||||
For temporary problems please read the login messages. Other problems are: | ||||||||
Changed: | ||||||||
< < |
| |||||||
> > |
| |||||||
SupportIn case of questions, please ask Holger Angenent or Martin Leweling via hpc@uni-muenster.de. |
Line: 1 to 1 | ||||||||
---|---|---|---|---|---|---|---|---|
NWZPHI the cluster of the IVV 4NWZPHI is a cluster equipped with 98 Xeon Phi cards. These are PCIe based accelerators similar to GPUs, but can be used with regular programming languages. | ||||||||
Added: | ||||||||
> > | Update: New Centos 7 Installation | |||||||
Content of this page
Hard- and Software overview |
Line: 1 to 1 | ||||||||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
NWZPHI the cluster of the IVV 4 | ||||||||||||||||
Line: 17 to 17 | ||||||||||||||||
NWZPHI for the impatient reader | ||||||||||||||||
Changed: | ||||||||||||||||
< < | The name of the login-server is NWZPHI. Allowed are all users that are members of the group u0clustr and at least one of the groups starting with p0, q0 or r0. In addition, every user allowed for PALMA may use NWZPHI. You can register yourself for u0clstr at ZIV.MeinZIV![]() | |||||||||||||||
> > | The name of the login-server is NWZPHI. Allowed are all users that are members of the group u0clustr and at least one of the groups starting with p0, q0 or r0. In addition, every user allowed for PALMA may use NWZPHI. You can register yourself for u0clstr at ZIV.MeinZIV![]() | |||||||||||||||
The batch and module system are working very similar to PALMA. | ||||||||||||||||
Line: 52 to 51 | ||||||||||||||||
module add compiler/intel/14 module add mpi/intel/4.1.0.024 | ||||||||||||||||
Changed: | ||||||||||||||||
< < | module add tools/mic | |||||||||||||||
> > | module add tools/mic | |||||||||||||||
Example: Compile a program that uses the FFTW: | ||||||||||||||||
Deleted: | ||||||||||||||||
< < | ||||||||||||||||
module add compiler/intel/14 module add mpi/intel/4.1.0.024 | ||||||||||||||||
Line: 62 to 59 | ||||||||||||||||
module add mpi/intel/4.1.0.024 module add lib/fftw/intel/3.3.3 | ||||||||||||||||
Changed: | ||||||||||||||||
< < | ${MPIICC} -I ${FFTW_INCLUDE_DIR} -o program program.c -g ${FLAGS_FAST} -L${FFTW_LIB_DIR} -lfftw_mpi -lfftw -lm | |||||||||||||||
> > | ${MPIICC} -I ${FFTW_INCLUDE_DIR} -o program program.c -g ${FLAGS_FAST} -L${FFTW_LIB_DIR} -lfftw_mpi -lfftw -lm | |||||||||||||||
Batch system | ||||||||||||||||
Line: 116 to 110 | ||||||||||||||||
| ||||||||||||||||
Deleted: | ||||||||||||||||
< < | ||||||||||||||||
To choose the node type that you want to use, you have to use the correct queue. So if you want to run a large computation which needs more than 128 GB of RAM for a single process, the dl560 is right for you. In this case, you have to use the bigsmp queue: | ||||||||||||||||
Deleted: | ||||||||||||||||
< < | ||||||||||||||||
Changed: | ||||||||||||||||
< < | #PBS -q bigsmp | |||||||||||||||
> > | #PBS -q bigsmp | |||||||||||||||
Submitting jobs / Managing the queue | ||||||||||||||||
Changed: | ||||||||||||||||
< < | A job is submitted by entering the command qsub submit.cmd, where submit.cmd is the name of the submit-file. | |||||||||||||||
> > | A job is submitted by entering the command
qsub submit.cmd, where submit.cmd is the name of the submit-file. | |||||||||||||||
Further commands:
| ||||||||||||||||
Line: 152 to 146 | ||||||||||||||||
Most of the computational power of NWZPHI stems from the accelerators. To have some numbers: The 24 CPU cores of one node deliver some hundred GFLOPS, a single Xeon PHI accelerator has a peak performance of 1 TFLOP. You can reserve the Xeon Phi Cards via the batch system. This would be done the following way: | ||||||||||||||||
Added: | ||||||||||||||||
> > | ||||||||||||||||
Changed: | ||||||||||||||||
< < | #PBS -l nodes=1:ppn=24:mics=8 | |||||||||||||||
> > | #PBS -l nodes=1:ppn=24:mics=8 | |||||||||||||||
Since the operating system cannot see, which cards are already in use, it is a good idea to reserve complete nodes with all of its cards and distribute the jobs by yourself. When you submit a job, there will be a variable PBS_MICFILE inside your job. In addition to the PBS_NODEFILE, in PBS_MICFILE there will be a list of the Xeon PHIs reserved for you. | ||||||||||||||||
Line: 160 to 155 | ||||||||||||||||
When you submit a job, there will be a variable PBS_MICFILE inside your job. In addition to the PBS_NODEFILE, in PBS_MICFILE there will be a list of the Xeon PHIs reserved for you. To use the Infiniband interfaces, append a "-ib" to the hostnames: | ||||||||||||||||
Added: | ||||||||||||||||
> > | ||||||||||||||||
Changed: | ||||||||||||||||
< < | cat $PBS_MICFILE | awk '{print $0"-ib"}' > micfile-ib.$PBS_JOBID | |||||||||||||||
> > | cat $PBS_MICFILE | awk '{print $0"-ib"}' > micfile-ib.$PBS_JOBID | |||||||||||||||
Some useful links |
Line: 1 to 1 | ||||||||
---|---|---|---|---|---|---|---|---|
NWZPHI the cluster of the IVV 4 | ||||||||
Line: 17 to 17 | ||||||||
NWZPHI for the impatient reader | ||||||||
Changed: | ||||||||
< < | The name of the login-server is NWZPHI. Allowed are all users that are members of the group u0clustr and at least one of the groups starting with p0, q0 or r0. In addition, every user allowed for PALMA may use NWZPHI. You can register yourself for u0clstr at MeinZIV![]() | |||||||
> > | The name of the login-server is NWZPHI. Allowed are all users that are members of the group u0clustr and at least one of the groups starting with p0, q0 or r0. In addition, every user allowed for PALMA may use NWZPHI. You can register yourself for u0clstr at ZIV.MeinZIV![]() | |||||||
The batch and module system are working very similar to PALMA. |
Line: 1 to 1 | ||||||||
---|---|---|---|---|---|---|---|---|
NWZPHI the cluster of the IVV 4 | ||||||||
Line: 157 to 157 | ||||||||
Since the operating system cannot see, which cards are already in use, it is a good idea to reserve complete nodes with all of its cards and distribute the jobs by yourself. | ||||||||
Added: | ||||||||
> > | When you submit a job, there will be a variable PBS_MICFILE inside your job. In addition to the PBS_NODEFILE, in PBS_MICFILE there will be a list of the Xeon PHIs reserved for you.
To use the Infiniband interfaces, append a "-ib" to the hostnames:
cat $PBS_MICFILE | awk '{print $0"-ib"}' > micfile-ib.$PBS_JOBID | |||||||
Some useful links |
Line: 1 to 1 | ||||||||||||||||||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
NWZPHI the cluster of the IVV 4 | ||||||||||||||||||||||||||
Line: 65 to 65 | ||||||||||||||||||||||||||
${MPIICC} -I ${FFTW_INCLUDE_DIR} -o program program.c -g ${FLAGS_FAST} -L${FFTW_LIB_DIR} -lfftw_mpi -lfftw -lm | ||||||||||||||||||||||||||
Changed: | ||||||||||||||||||||||||||
< < | Batch system | |||||||||||||||||||||||||
> > | Batch system | |||||||||||||||||||||||||
Added: | ||||||||||||||||||||||||||
> > | The batch system Torque and the scheduler Moab are used to submit jobs. It is not allowed, to start jobs manually. Batch jobs should only be submitted from the server mn02.
Creating submit-filesExample of a submit-file of a MPI-job:#PBS -o output.dat #PBS -l walltime=01:00:00,nodes=2:ppn=24 #PBS -M username@uni-muenster.de #PBS -m ae #PBS -q default #PBS -N job_name #PBS -j oe cd $PBS_O_WORKDIR mpdboot -n 2 -f $PBS_NODEFILE -v mpirun -machinefile $PBS_NODEFILE -np 48 ./executableAn MPI-job with 48 processes is started. Further Information:
#PBS -o output.dat #PBS -l walltime=01:00:00,nodes=1:ppn=24 #PBS -M username@uni-muenster.de #PBS -m ae #PBS -q default #PBS -N job_name #PBS -j oe cd $PBS_O_WORKDIR export OMP_NUM_THREADS=12 ./executable Choosing the nodesThe cluster consists of the following nodes:
#PBS -q bigsmp Submitting jobs / Managing the queueA job is submitted by entering the commandqsub submit.cmd, where submit.cmd is the name of the submit-file. Further commands:
Monitoring jobsThere are different tools for monitoring
StorageThere is a 88 TB partition for /home and /scratch using the BeeGFS![]() | |||||||||||||||||||||||||
Using Xeon Phi Accelerators | ||||||||||||||||||||||||||
Added: | ||||||||||||||||||||||||||
> > | to be done...
Most of the computational power of NWZPHI stems from the accelerators. To have some numbers: The 24 CPU cores of one node deliver some hundred GFLOPS, a single Xeon PHI accelerator has a peak performance of 1 TFLOP.
You can reserve the Xeon Phi Cards via the batch system. This would be done the following way:
#PBS -l nodes=1:ppn=24:mics=8Since the operating system cannot see, which cards are already in use, it is a good idea to reserve complete nodes with all of its cards and distribute the jobs by yourself. Some useful links
| |||||||||||||||||||||||||
Known IssuesFor temporary problems please read the login messages. |
Line: 1 to 1 | |||||||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
NWZPHI the cluster of the IVV 4 | |||||||||||||||
Line: 37 to 37 | |||||||||||||||
Environment Modules | |||||||||||||||
Added: | |||||||||||||||
> > | Environment variables (like PATH, LD_LIBRARY_PATH) for compilers and libraries can be set by modules:
module add compiler/intel/14 module add mpi/intel/4.1.0.024 module add tools/micExample: Compile a program that uses the FFTW: module add compiler/intel/14 module add mpi/intel/4.1.0.024 module add lib/fftw/intel/3.3.3 ${MPIICC} -I ${FFTW_INCLUDE_DIR} -o program program.c -g ${FLAGS_FAST} -L${FFTW_LIB_DIR} -lfftw_mpi -lfftw -lm | ||||||||||||||
Batch systemUsing Xeon Phi Accelerators |
Line: 1 to 1 | ||||||||
---|---|---|---|---|---|---|---|---|
NWZPHI the cluster of the IVV 4NWZPHI is a cluster equipped with 98 Xeon Phi cards. These are PCIe based accelerators similar to GPUs, but can be used with regular programming languages. | ||||||||
Changed: | ||||||||
< < | Hardware overview | |||||||
> > | Content of this page
Hard- and Software overview | |||||||
| ||||||||
Added: | ||||||||
> > |
| |||||||
NWZPHI for the impatient reader | ||||||||
Changed: | ||||||||
< < | The name of the login-server is NWZPHI. Allowed are all users that are members of the group u0clustr and at least one of the groups starting with p0, q0 or r0. You can register yourself for u0clstr at http://www.uni-muenster.de/MeinZIV/. | |||||||
> > | The name of the login-server is NWZPHI. Allowed are all users that are members of the group u0clustr and at least one of the groups starting with p0, q0 or r0. In addition, every user allowed for PALMA may use NWZPHI. You can register yourself for u0clstr at MeinZIV![]() | |||||||
The batch and module system are working very similar to PALMA. | ||||||||
Added: | ||||||||
> > | Differences to PALMAIf you are familiar with PALMA, starting jobs on NWZPHI is quite easy. There are some differences mentioned here
Starting jobs on NWZPHI
| |||||||
Environment ModulesBatch systemUsing Xeon Phi Accelerators | ||||||||
Added: | ||||||||
> > | Known IssuesFor temporary problems please read the login messages. Other problems are:
| |||||||
Support | ||||||||
Added: | ||||||||
> > | In case of questions, please ask Holger Angenent or Martin Leweling via hpc@uni-muenster.de. | |||||||
-- HolgerAngenent - 2014-07-23 |
Line: 1 to 1 | ||||||||
---|---|---|---|---|---|---|---|---|
Added: | ||||||||
> > |
NWZPHI the cluster of the IVV 4NWZPHI is a cluster equipped with 98 Xeon Phi cards. These are PCIe based accelerators similar to GPUs, but can be used with regular programming languages.Hardware overview
NWZPHI for the impatient readerThe name of the login-server is NWZPHI. Allowed are all users that are members of the group u0clustr and at least one of the groups starting with p0, q0 or r0. You can register yourself for u0clstr at http://www.uni-muenster.de/MeinZIV/. The batch and module system are working very similar to PALMA.Environment ModulesBatch systemUsing Xeon Phi AcceleratorsSupport-- HolgerAngenent - 2014-07-23 |