ANSYS: Difference between revisions
mNo edit summary |
(Add text interface and batch job script examples) |
||
Line 11: | Line 11: | ||
Typically, researchers will install ANSYS on their own computers to develop models in a graphical user interface and then run simulations that exceed their local hardware capabilities on ARC. | Typically, researchers will install ANSYS on their own computers to develop models in a graphical user interface and then run simulations that exceed their local hardware capabilities on ARC. | ||
The software can be downloaded, upon approval, from the [https://iac01.ucalgary.ca/SDSWeb/IT Software Distribution] web site. | The software can be downloaded, upon approval, from the [https://iac01.ucalgary.ca/SDSWeb/ IT Software Distribution] web site. | ||
ANSYS is available to all U of C researchers with an ARC account, but, with licensing restrictions as outlined in the next section.= <span id="licensing"></span>Licensing considerations = | ANSYS is available to all U of C researchers with an ARC account, but, with licensing restrictions as outlined in the next section.= <span id="licensing"></span>Licensing considerations = | ||
Line 24: | Line 24: | ||
lmutil lmstat -c 1055@ansyslic.ucalgary.ca -a | lmutil lmstat -c 1055@ansyslic.ucalgary.ca -a | ||
</pre> | </pre> | ||
For ANSYS Fluent, each job on ARC will use one token of the software feature "aa_r" in the lmstat output. In addition, one license token per core is used of the "aa_r_hpc" type for cores in excess of 16. So, for example, a job using a 40-core node from the cpu2019 partition will use one aa_r token and 24 aa_r_hpc tokens. | |||
Using the fastest hardware available will provide the most value a given number of license tokens, so, using the 40-core compute nodes, selected by specifying the cpu2019 partition in your batch job (see example scripts below), is preferred. However, if there is a shortage of license tokens, you may use just part of a compute node or compute nodes from the older legacy partitions, such as parallel. | |||
= Running ANSYS Fluent batch jobs = | |||
Researchers using ANSYS on ARC are expected to be generally familiar with ANSYS capabilities, input file format and the use of restart files. | |||
== Creating a Fluent input file == | |||
After preparing your model, at the point where you are ready to run a Fluent solver, you save the case and data files and transfer them to ARC. In addition to those files, to run your model on ARC you need an input file containing Fluent text interface commands to specify such parameters as the solver to use, the number of time steps, the frequency of output and other simulation controls. | |||
Typically the main difficulty in getting started with Fluent on ARC is figuring out what text interface commands correspond to the graphical interface commands with which you may be more familiar from using a desktop version of Fluent. At the Fluent command prompt, if you just hit enter the available commands will be shown, similar to: | |||
<pre> | |||
adapt/ file/ report/ | |||
define/ mesh/ solve/ | |||
display/ parallel/ surface/ | |||
exit plot/ views/ | |||
</pre> | |||
Entering one of those commands and then another enter will give sub-options: | |||
<pre> | |||
> file | |||
/file> | |||
async-optimize? read-case-data start-journal | |||
auto-save/ read-field-functions start-transcript | |||
binary-files? read-journal stop-journal | |||
confirm-overwrite? read-macros stop-macro | |||
define-macro read-profile stop-transcript | |||
execute-macro read-transient-table transient-export/ | |||
export/ set-batch-options write-cleanup-script | |||
import/ show-configuration write-field-functions | |||
read-case solution-files/ write-macros | |||
</pre> | |||
So, for example, one can discover by exploring these menus that the commands to set the frequency with which data and case files can be automatically stored periodically during a long run are of the form: | |||
<pre> | |||
/file/auto-save/data-frequency 1000 | |||
/file/auto-save/case-frequency if-case-is-modified | |||
</pre> | |||
Here is an example of a complete text input file in which case and data files are read in, some parameters are set related to the storing of output, the solver is run and data and case files saved at the end of the run. | |||
<pre> | |||
/file/read-case test.cas | |||
/file/read-data test.dat | |||
/file/confirm-overwrite no | |||
/file/auto-save/data-frequency 1000 | |||
/file/auto-save/case-frequency if-case-is-modified | |||
/file/auto-save/root-name test | |||
/solve/dual-time-iterate | |||
22200 | |||
150 | |||
/file/write-case test.%t.%i.cas | |||
/file/write-data test.%t.%i.dat | |||
</pre> | |||
Note that blank lines are significant for some commands. | |||
== Slurm batch job script examples == | |||
Like other calculations on ARC systems, ANSYS software is run by submitting an appropriate script for batch scheduling using the sbatch command. For more information about submitting jobs, see the [[ARC_Cluster_Guide|ARC Cluster Guide]]. | |||
The scripts below can serve as a template for your own batch job scripts. | |||
=== Full node example - cpu2019 partition === | |||
When running on a full compute node, specify --mem=0 to request all the associated memory on the node. Note that when using the cpu2019 partition (40-core nodes), an n-node ANSYS job will take 40*n-16 license tokens from the aa_r_hpc pool. | |||
See /global/software/ansys/scripts/ansys_2019r2_fluent_cpu2019_node.slurm on ARC | |||
=== Legacy node example - parallel partition === | |||
Use the parallel partition only when the waiting time for cpu2019 nodes is comparable to the run time, as the cpu2019 partition nodes should run Fluent about twice as fast as the parallel partition nodes. | |||
See /global/software/ansys/scripts/ansys_2019r2_fluent_parallel_node.slurm on ARC | |||
== Support == | |||
Please send any questions regarding using ANSYS on ARC to support@hpc.ucalgary.ca. |
Revision as of 10:18, 29 May 2020
Introduction
ANSYS (external link) is a commercial suite of programs for engineering simulation, including fluid dynamics (Fluent and CFX), structural analysis (ANSYS Mechanical) and electromagnetics/electronics software.
Typically, researchers will install ANSYS on their own computers to develop models in a graphical user interface and then run simulations that exceed their local hardware capabilities on ARC.
The software can be downloaded, upon approval, from the IT Software Distribution web site.
ANSYS is available to all U of C researchers with an ARC account, but, with licensing restrictions as outlined in the next section.= Licensing considerations =
For many years, Information Technologies has provided a limited number of license tokens for ANSYS software, sometimes supplemented by contributions from researchers. The software contract is typically renewed annually in August. If you are interested in contributing to the pool of licenses, you can write to the IT Help Desk itsupport@ucalgary.ca and ask that your email be redirected to the IT software librarian.
The discussion that follows relates only to the research version of the software. Note that the conditions of use of the teaching licenses prohibits them from being used for research projects.
At the time of this writing in May 2020, there are 50 basic academic licenses and 512 extended "HPC" license tokens available (with 256 of the latter reserved for a specific research group who purchased their own licenses). The number of tokens available at a given time can be seen by running the following commands on ARC:
module load ansys/2019r2 lmutil lmstat -c 1055@ansyslic.ucalgary.ca -a
For ANSYS Fluent, each job on ARC will use one token of the software feature "aa_r" in the lmstat output. In addition, one license token per core is used of the "aa_r_hpc" type for cores in excess of 16. So, for example, a job using a 40-core node from the cpu2019 partition will use one aa_r token and 24 aa_r_hpc tokens.
Using the fastest hardware available will provide the most value a given number of license tokens, so, using the 40-core compute nodes, selected by specifying the cpu2019 partition in your batch job (see example scripts below), is preferred. However, if there is a shortage of license tokens, you may use just part of a compute node or compute nodes from the older legacy partitions, such as parallel.
Running ANSYS Fluent batch jobs
Researchers using ANSYS on ARC are expected to be generally familiar with ANSYS capabilities, input file format and the use of restart files.
Creating a Fluent input file
After preparing your model, at the point where you are ready to run a Fluent solver, you save the case and data files and transfer them to ARC. In addition to those files, to run your model on ARC you need an input file containing Fluent text interface commands to specify such parameters as the solver to use, the number of time steps, the frequency of output and other simulation controls.
Typically the main difficulty in getting started with Fluent on ARC is figuring out what text interface commands correspond to the graphical interface commands with which you may be more familiar from using a desktop version of Fluent. At the Fluent command prompt, if you just hit enter the available commands will be shown, similar to:
adapt/ file/ report/ define/ mesh/ solve/ display/ parallel/ surface/ exit plot/ views/
Entering one of those commands and then another enter will give sub-options:
> file /file> async-optimize? read-case-data start-journal auto-save/ read-field-functions start-transcript binary-files? read-journal stop-journal confirm-overwrite? read-macros stop-macro define-macro read-profile stop-transcript execute-macro read-transient-table transient-export/ export/ set-batch-options write-cleanup-script import/ show-configuration write-field-functions read-case solution-files/ write-macros
So, for example, one can discover by exploring these menus that the commands to set the frequency with which data and case files can be automatically stored periodically during a long run are of the form:
/file/auto-save/data-frequency 1000 /file/auto-save/case-frequency if-case-is-modified
Here is an example of a complete text input file in which case and data files are read in, some parameters are set related to the storing of output, the solver is run and data and case files saved at the end of the run.
/file/read-case test.cas /file/read-data test.dat /file/confirm-overwrite no /file/auto-save/data-frequency 1000 /file/auto-save/case-frequency if-case-is-modified /file/auto-save/root-name test /solve/dual-time-iterate 22200 150 /file/write-case test.%t.%i.cas /file/write-data test.%t.%i.dat
Note that blank lines are significant for some commands.
Slurm batch job script examples
Like other calculations on ARC systems, ANSYS software is run by submitting an appropriate script for batch scheduling using the sbatch command. For more information about submitting jobs, see the ARC Cluster Guide.
The scripts below can serve as a template for your own batch job scripts.
Full node example - cpu2019 partition
When running on a full compute node, specify --mem=0 to request all the associated memory on the node. Note that when using the cpu2019 partition (40-core nodes), an n-node ANSYS job will take 40*n-16 license tokens from the aa_r_hpc pool.
See /global/software/ansys/scripts/ansys_2019r2_fluent_cpu2019_node.slurm on ARC
Legacy node example - parallel partition
Use the parallel partition only when the waiting time for cpu2019 nodes is comparable to the run time, as the cpu2019 partition nodes should run Fluent about twice as fast as the parallel partition nodes.
See /global/software/ansys/scripts/ansys_2019r2_fluent_parallel_node.slurm on ARC
Support
Please send any questions regarding using ANSYS on ARC to support@hpc.ucalgary.ca.