Skip to content

Warning

This page has not been updated yet. The page does not reflect the transition from PBS to Slurm.

Workbench

Workbench Batch Mode

It is possible to run Workbench scripts in a batch mode. You need to configure solvers of individual components to run in parallel mode. Open your project in Workbench. Then, for example, in Mechanical, go to Tools - Solve Process Settings....

Enable the Distribute Solution checkbox and enter the number of cores (e.g. 72 to run on two Barbora nodes). If you want the job to run on more than 1 node, you must also provide a so called MPI appfile. In the Additional Command Line Arguments input field, enter:

    -mpifile /path/to/my/job/mpifile.txt

Where /path/to/my/job is the directory where your project is saved. We will create the file mpifile.txt programmatically later in the batch script. For more information, refer to *ANSYS Mechanical APDL Parallel Processing* *Guide*.

Now, save the project and close Workbench. We will use this script to launch the job:

    #!/bin/bash
    #PBS -l select=2:ncpus=128
    #PBS -q qprod
    #PBS -N test9_mpi_2
    #PBS -A OPEN-0-0


    # change the working directory
    DIR=/scratch/project/PROJECT_ID/$PBS_JOBID
    mkdir -p "$DIR"
    cd "$DIR" || exit

    echo Running on host `hostname`
    echo Time is `date`
    echo Directory is `pwd`
    echo This jobs runs on the following nodes:
    echo `cat $PBS_NODEFILE`

    ml ANSYS/21.1-intel-2018a

    #### Set number of processors per host listing
    procs_per_host=24
    #### Create MPI appfile
    echo -n "" > mpifile.txt
    for host in `cat $PBS_NODEFILE`
    do
      echo "-h $host -np $procs_per_host $ANSYS160_DIR/bin/ansysdis161 -dis" > mpifile.txt
    done

    #-i input.dat includes the input of analysis in APDL format
    #-o file.out is output file from ansys where all text outputs will be redirected
    #-p the name of license feature (aa_r=ANSYS Academic Research, ane3fl=Multiphysics(commercial), aa_r_dy=Academic AUTODYN)

    # prevent using scsif0 interface on accelerated nodes
    export MPI_IC_ORDER="UDAPL"
    # spawn remote process using SSH (default is RSH)
    export MPI_REMSH="/usr/bin/ssh"

    runwb2 -R jou6.wbjn -B -F test9.wbpj

The solver settings are saved in the solvehandlers.xml file, which is not located in the project directory. Verify your solved settings when uploading a project from your local computer.