azurehpc/tutorials/oil_and_gas_opm
Xavier Pillons 9cfbb19853 use sudo for installing PBS 2020-05-12 12:21:28 +02:00
..
scripts add intersect, quick fix for broken tutorial due to missing setup_win.sh 2019-08-02 00:12:31 -07:00
config.json use sudo for installing PBS 2020-05-12 12:21:28 +02:00
readme.md Removed azhpc-view and put links for azurehpc-ui 2019-09-13 15:58:39 +01:00

readme.md

Build a PBS compute cluster with a Windows visualization node to run OPM and ResInsight

Visualisation: config.json

This example will create an HPC cluster with a CentOS 7.6 headnode running PBS Pro 19.1 exporting a 4TB NFS space and multiple CentOS 7.6 HB60rs compute nodes; and a Windows visualization node. This tutorial uses NFS and RDP for simplicity but you can also easily set it up with an alternate storage or visualization solution using the examples here.

NOTE:

  • MAKE SURE you have followed the steps in prerequisite before proceeding here

First initialise a new project. AZHPC provides the azhpc-init command that will help here. Running with the -s parameter will show all the variables that need to be set, e.g.

azhpc-init -c $azhpc_dir/tutorials/oil_and_gas_opm -d oil_and_gas_opm -s

The variables can be set with the -v option where variables are comma separated. The -d option is required and will create a new directory name for you.

azhpc-init -c $azhpc_dir/tutorials/oil_and_gas_opm -d oil_and_gas_opm -v location=southcentralus,resource_group=azhpc-cluster,win_password=[password or secret.azhpc-vault.winadmin-secret]

Create the cluster

cd oil_and_gas_opm
azhpc-build

Allow ~10 minutes for deployment.

To check the status of the VMs run

azhpc-status

Connect to the headnode and check PBS and NFS

$ azhpc-connect -u hpcuser headnode
Fri Jun 28 09:18:04 UTC 2019 : logging in to headnode (via headnode6cfe86.westus2.cloudapp.azure.com)
[hpcuser@headnode ~]$ pbsnodes -avS
vnode           state           OS       hardware host            queue        mem     ncpus   nmics   ngpus  comment
--------------- --------------- -------- -------- --------------- ---------- -------- ------- ------- ------- ---------
compuc407000003 free            --       --       10.2.4.8        --            224gb      60       0       0 --
compuc407000002 free            --       --       10.2.4.7        --            224gb      60       0       0 --
[hpcuser@headnode ~]$ sudo exportfs -v
/share/apps     <world>(sync,wdelay,hide,no_subtree_check,sec=sys,rw,secure,root_squash,no_all_squash)
/share/data     <world>(sync,wdelay,hide,no_subtree_check,sec=sys,rw,secure,root_squash,no_all_squash)
/share/home     <world>(sync,wdelay,hide,no_subtree_check,sec=sys,rw,secure,root_squash,no_all_squash)
/mnt/resource/scratch
                <world>(sync,wdelay,hide,no_subtree_check,sec=sys,rw,secure,root_squash,no_all_squash)
[hpcuser@headnode ~]$

To check the state of the cluster you can run the following commands
azhpc-connect -u hpcuser headnode
qstat -Q
pbsnodes -avS
df -h

Return to the deployment node to install applications

exit

Install applications

First copy the apps directory to the cluster. The azhpc-scp can be used to do this:

azhpc-scp -r $azhpc_dir/apps hpcuser@headnode:.

Alternatively you can checkout the azurehpc repository but you will need to update the paths to apps directory according to where you put it.

Install OPM

Run the install script:

azhpc-run -u hpcuser apps/opm/install_opm.sh

Note: this can be run when on the cluster

OPM requires that lapack be installed on all of the compute nodes. The scripts/app_opm_req.sh script was created and has been added as an install step in the config.json.

The full_install_opm.sh can be used to build everything from source.

ResInsight Installation

Run the install script:

azhpc-run -u hpcuser apps/resinsight/install_resinsight_v2019.04.sh

Connect to the headnode

azhpc-connect -u hpcuser headnode

Run the OPM norne scenario

To run on a single node with 30 cores run

qsub -l select=1:ncpus=30:mpiprocs=30,place=scatter:excl $HOME/apps/opm/flow_norne.sh

To run on two node with 30 cores run

qsub -l select=2:ncpus=15:mpiprocs=15,place=scatter:excl $HOME/apps/opm/flow_norne.sh

Notes:

  • The job outputs files will be stored in \data\opm-data\norne\out_parallel.

Remote Visualization

Connect to the viznode using RDB (get the RDP file for nvnode from the Azure Portal)

  • Username: hpcadmin
  • Password:

Check that Y: and Z: drives are mapped to the NFS server

Note : Y: and Z: drives appears as disconnected while they are not.

Launch ResInsight from the Y: drive (Y:\resinsight\ResInsight-2019.04.0_oct-4.0.0_souring_win64\ResInsight) and then, from ResInsight, click import eclipse file. The result file is located on the Z: drive (Z:\opm-data\norne\out_parallel\NORNE_ATW2013.EGRID)