79 | | ||base_directory ||Working directory from where {{{palmrun}}} or {{{palmbuild}}} is called. This folder must contain the configuration file {{{.palm.config....}}}. || |
80 | | ||base_data ||Directory where the PALM I/O-files are stored. This variable is used in the file configuration file (see [source:palm/trunk/SCRIPTS/.palm.iofiles .palm.iofiles], and the [wiki:doc/app/palm_iofiles description] of this file). Since this variable is also used to determine file locations on remote hosts, it must use the {{{~}}} instead of {{{$HOME}}}. || |
81 | | ||source_path ||Path to PALM's FORTRAN source files. This is the place where the automatic installer has put the download, or which has been defined in the user's {{{svn checkout}}} command. || |
82 | | ||user_source_path ||Path to the [wiki:doc/app/userint user interface routines]. The variable {{{jobname}}} that may be used in the default path is replaced by the argument given with {{{palmrun}}}-option {{{-r}}}. || |
83 | | ||fast_io_catalog ||Path to a file system with fast discs (if available). This folder is used so store the temporary catalog generated by {{{palmrun}}} during each run. It should also be used to store large I/O files (e.g. restart data or 3D-output) in order to reduce I/O time. This variable is used in the default {{{.palm.iofiles}}} for the restart data files. The folder must be accessible from all compute nodes, i.e. it must reside in a global file system. WARNING: {{{/tmp}}} will only work on single node systems! In case of batch jobs on remote hosts, the variable refers to a folder on the remote host. The variable has no default value and must be set by the user. || |
84 | | ||local_jobcatalog ||Folder on the local host to store the batch job protocols. In case of batch jobs running on remote hosts, the job protocol will be created on the {{{remote_jobcatalog}}} and then be transferred via scp to the {{{local_jobcatalog}}}. || |
85 | | ||remote_jobcatalog ||In case of batch jobs running on remote hosts, the job protocol will be put in this folder, and then automatically transferred via scp to the {{{local_jobcatalog}}}. The transfer is done by a separate small batch job, which directives are defined by the {{{BDT:}}} lines. The variable has no default value and must be set by the user. Absolute paths need to be given. Using {{{$HOME}}} is not allowed / does not work. || |
86 | | ||local_ip ||IP-address of your local computer / the computer on which you call the {{{palmrun}}}/{{{palmbuild}}} command. You may use {{{127.0.0.0}}} if you are running PALM in interactive mode or in batch mode on your local computer. The address is only used to identify where to send the output data in case of batch jobs on a remote host. || |
87 | | ||local_username ||Your username on the local computer / the computer on which you call the {{{palmrun}}}/{{{palmbuild}}} command. The local username is required for running batch jobs on a remote host in order to allow the batch job to access your local system (e.g. for sending back output data or for automatically starting restart runs). || |
88 | | ||remote_ip ||IP-address of the remote system where the batch job shall be started. On large cluster systems this will usually be the address of a login node. Setting this variable in the configuration file will cause {{{palmrun}}} to run in remote batch job mode, i.e. a batch job will be created and send to the remote system automatically without giving {{{palmrun}}}-option {{{-b}}}. || |
89 | | ||remote_username ||Your username on the remote computer that is given by {{{remote_ip}}}. || |
90 | | ||ssh_key ||Name of the file from which the identity (private key) for public key authentication is read. This file is assumed to be in folder {{{$HOME/.ssh}}}. By default (if you omit this variable), file {{{id_dsa}}} or {{{id_rsa}}} is used. || |
91 | | ||remote_loginnode ||Name of the login node of the remote computer. Nodes on big compute clusters are separated into compute nodes and login nodes (and sometimes I/O nodes). Some computer centers only allow the login nodes to establish ssh/scp connections to addresses outside the computing center. In such cases, since {{{palmrun}}} is executed on the compute nodes, it first has to send the output data to the login node, from where it is then forwarded to your local computer. If the compute nodes on your remote host do not allow direct ssh/scp connections to your local computer, you need to provide the name of the login node of the remote host. Typically, this is a mnemonic name like ''loginnode1'' and not an IP-address (like ''111.111.11.11''). Several login nodes often exist. You just have to give one of them. If you do not provide a name, you probably will not receive data on your local host from the PALM run. || |
92 | | ||defaultqueue ||Batch job queue to be used if no queue is explicitly given with {{{palmrun}}} option {{{-q}}}. || |
93 | | ||submit_command ||Full path to the command that has to be used to submit batch jobs on your system (either on the local, or on the remote host), including required option. See documentation of your batch system / computing center to find out which command has to be used. An example for a {{{moab}}} batch system could be {{{/opt/moab/default/bin/msub -E}}}. If you only know the command name (e.g. ''msub''), entering {{{which msub}}} on the local/remote host will give you the full path. || |
94 | | ||compiler_name ||Name of the FORTRAN compiler to be used to create the PALM executable. Typically, this is the name of a wrapper script like ''mpif90'' or e.g. ''ftn'' on Cray machines, which automatically invokes the required MPI library and MPI include file. If you don't have a wrapper script, you may need to explicitly give compiler options (see {{{compiler_options}}}) to provide paths to the library / include file. If you like to run PALM without MPI (serial mode, or OpenMP parallelization), you should not use a wrapper script and give the original compiler name instead. || |
95 | | ||compiler_name_ser ||FORTRAN compiler name to create non-MPI executables. This name is required, because {{{palmbuild}}} generates several helper programs for pre-/post-processing, which run in serial mode on just one code. Here you give the original compiler name, like ''ifort'', ''pgfortran'', ''gfortran'', or ''xlf95''. || |
96 | | ||cpp_options ||Preprocessor directives to be used for compiling the PALM code. They allow for conditional compilation using the {{{-D}}} compiler option. Compiling PALM with MPI support requires options {{{-D__parallel -DMPI_REAL=MPI_DOUBLE_PRECISION -DMPI_2REAL=MPI_2DOUBLE_PRECISION}}}. Many compilers require to set an additional option to run the FORTRAN preprocessor on source files before compilation (e.g. ''-fpp'' for the Intel compiler). This option has to be given here too. Alternatively, you can provide it as part of the {{{compiler_options}}}. See [wiki:doc/app/cpp_options cpp_options] for a complete list of preprocessor define strings that are used in the PALM code. || |
97 | | ||make_options ||Options for the UNIX {{{make}}}-command, which is used by {{{palmbuild}}} to compile the PALM code. In order to speed up compilation, you may use the {{{-j}}} option, which specifies the number of jobs to run simultaneously. If you have e.g. 4 cores on your local computer system, then {{{-j 4}}} starts 4 instances of the FORTRAN compiler, i.e. 4 FORTRAN-files are compiled simultaneously (if the dependencies allow for that). Do not try to start more instances than the number of available cores, because this will decrease the compiler performance significantly. || |
98 | | ||compiler_options ||Options to be used by the compiler that has been specified by {{{compiler_name}}} / {{{compiler_name_ser}}} in order to compile the PALM and utilities source code. Please see [wiki:doc/app/recommended_compiler_options] for recommended compiler options for specific compilers. Library paths do not have to be given here (although you can do that), but paths to INCLUDE files may need to be specified. || |
99 | | ||linker_options ||Compiler options to be used to link the PALM executable. Typically, these are paths to libraries used by PALM, e.g. NetCDF, FFTW, MPI, etc. You may repeat the options that you have given with {{{compiler_options}}} here. See your local system documentation / software manuals for required path settings. Requirements differ from system to system and also depend on the respective libraries that you are using. See [wiki:doc/app/recommended_compiler_options] for specific path settings that we, the PALM group, are using on our computers. Be aware, that these settings probably will not work on your computer system. || |
100 | | ||'''hostfile''' ||'''Name of the hostfile that is used by MPI to determine the nodes on which the MPI processes are started.'''\\\\ {{{palmrun}}} automatically generates the hostfile if you set {{{auto}}}. All MPI processes will then be started on the node on which {{{palmrun}}} is executed. The real name of the hostfile will then be set to {{{hostfile}}} (instead of {{{auto}}}) and, depending on your local MPI implementation, you may have to give this name in the {{{execute_command}}}. MPI implementations on large computer centers often do not require to explicitly specify a hostfile (in such a case you can remove this line from the configuration file), or the batch systems provides a hostfile which name you may access via environment variables (e.g. {{{$PBS_NODEFILE}}}) and which needs to be given in the {{{execute_command}}}. Please see your local system / batch system documentation about the hostfile policy on your system. || |
101 | | ||execute_command ||MPI command to start the PALM executable. \\ Please see your local MPI documentation about which command needs to be used on your system. The name of the PALM executable, usually the last argument of the execute command, must be {{{palm}}}. Typically, the command requires to give several further options like the number of MPI processes to be started, or the number of compute nodes to be used. Values of these options may change from run to run. Don't give specific values here and use variables (written in double curly brackets) instead which will be automatically replaced by {{{palmrun}}} with values that you have specified with respective {{{palmrun}}} options. As an example {{{aprun -n {{mpi_tasks}} -N {{tasks_per_node}} palm}}} will be interpreted as {{{aprun -n 240 -N 24 palm}}} if you call {{{palmrun ... -X240 -T24 ...}}}. See the batch job section below about further variables that are recognized by {{{palmrun}}}. || |
102 | | ||memory ||Memory request per MPI process (or CPU core) in MByte. \\ {{{palmrun}}} option{{{-m}}} overwrites this setting. || |
103 | | ||module_commands ||Module command(s) for loading required software / libraries. \\ In case that you have a {{{modules}}} package on your system, you can specify here the command(s) to load the specific software / libraries that your PALM run requires, e.g. the compiler, the NetCDF software, the MPI library, etc. Alternatively, you can load the modules from your shell profile (e.g. {{{.bashrc}}}), but then all your PALM runs will use the same settings. An example for a Cray system to use fftw and parallel NetCDF is {{{module load fftw cray-hdf5-parallel cray-netcdf-hdf5parallel}}}. The commands are carried out at the beginning of a batch job, or before PALM is compiled with {{{palmbuild}}}. || |
104 | | ||login_init_cmd ||Special commands to be carried out at login or start of batch jobs on the remote host. \\ You may specify here a command, e.g. for setting up special system environments in batch jobs. It is carried out as first command in the batch job. || |
| 79 | || base_data || Directory where the PALM I/O-files are stored. This variable is used in the file configuration file (see [source:palm/trunk/SCRIPTS/.palm.iofiles .palm.iofiles], and the [wiki:doc/app/palm_iofiles description] of this file). Since this variable is also used to determine file locations on remote hosts, it must use the {{{~}}} instead of {{{$HOME}}}. |
| 80 | || base_directory || Working directory from where {{{palmrun}}} or {{{palmbuild}}} is called. This folder must contain the configuration file {{{.palm.config....}}}. |
| 81 | || compiler_name || Name of the FORTRAN compiler to be used to create the PALM executable. Typically, this is the name of a wrapper script like ''mpif90'' or e.g. ''ftn'' on Cray machines, which automatically invokes the required MPI library and MPI include file. If you don't have a wrapper script, you may need to explicitly give compiler options (see {{{compiler_options}}}) to provide paths to the library / include file. If you like to run PALM without MPI (serial mode, or OpenMP parallelization), you should not use a wrapper script and give the original compiler name instead. |
| 82 | || compiler_name_ser || FORTRAN compiler name to create non-MPI executables. This name is required, because {{{palmbuild}}} generates several helper programs for pre-/post-processing, which run in serial mode on just one code. Here you give the original compiler name, like ''ifort'', ''pgfortran'', ''gfortran'', or ''xlf95''. |
| 83 | || compiler_options || Options to be used by the compiler that has been specified by {{{compiler_name}}} / {{{compiler_name_ser}}} in order to compile the PALM and utilities source code. Please see [wiki:doc/app/recommended_compiler_options] for recommended compiler options for specific compilers. Library paths do not have to be given here (although you can do that), but paths to INCLUDE files may need to be specified. |
| 84 | || cpp_options || Preprocessor directives to be used for compiling the PALM code. They allow for conditional compilation using the {{{-D}}} compiler option. Compiling PALM with MPI support requires options {{{-D__parallel -DMPI_REAL=MPI_DOUBLE_PRECISION -DMPI_2REAL=MPI_2DOUBLE_PRECISION}}}. Many compilers require to set an additional option to run the FORTRAN preprocessor on source files before compilation (e.g. ''-fpp'' for the Intel compiler). This option has to be given here too. Alternatively, you can provide it as part of the {{{compiler_options}}}. See [wiki:doc/app/cpp_options cpp_options] for a complete list of preprocessor define strings that are used in the PALM code. |
| 85 | || defaultqueue || Batch job queue to be used if no queue is explicitly given with {{{palmrun}}} option {{{-q}}}. |
| 86 | || execute_command || MPI command to start the PALM executable. \\ Please see your local MPI documentation about which command needs to be used on your system. The name of the PALM executable, usually the last argument of the execute command, must be {{{palm}}}. Typically, the command requires to give several further options like the number of MPI processes to be started, or the number of compute nodes to be used. Values of these options may change from run to run. Don't give specific values here and use variables (written in double curly brackets) instead which will be automatically replaced by {{{palmrun}}} with values that you have specified with respective {{{palmrun}}} options. As an example {{{aprun -n {{mpi_tasks}} -N {{tasks_per_node}} palm}}} will be interpreted as {{{aprun -n 240 -N 24 palm}}} if you call {{{palmrun ... -X240 -T24 ...}}}. See the batch job section below about further variables that are recognized by {{{palmrun}}}. |
| 87 | || fast_io_catalog || Path to a file system with fast discs (if available). This folder is used so store the temporary catalog generated by {{{palmrun}}} during each run. It should also be used to store large I/O files (e.g. restart data or 3D-output) in order to reduce I/O time. This variable is used in the default {{{.palm.iofiles}}} for the restart data files. The folder must be accessible from all compute nodes, i.e. it must reside in a global file system. WARNING: {{{/tmp}}} will only work on single node systems! In case of batch jobs on remote hosts, the variable refers to a folder on the remote host. The variable has no default value and must be set by the user. |
| 88 | || hostfile || Name of the hostfile that is used by MPI to determine the nodes on which the MPI processes are started. \\\\ {{{palmrun}}} automatically generates the hostfile if you set {{{auto}}}. All MPI processes will then be started on the node on which {{{palmrun}}} is executed. The real name of the hostfile will then be set to {{{hostfile}}} (instead of {{{auto}}}) and, depending on your local MPI implementation, you may have to give this name in the {{{execute_command}}}. MPI implementations on large computer centers often do not require to explicitly specify a hostfile (in such a case you can remove this line from the configuration file), or the batch systems provides a hostfile which name you may access via environment variables (e.g. {{{$PBS_NODEFILE}}}) and which needs to be given in the {{{execute_command}}}. Please see your local system / batch system documentation about the hostfile policy on your system. |
| 89 | || linker_options || Compiler options to be used to link the PALM executable. Typically, these are paths to libraries used by PALM, e.g. NetCDF, FFTW, MPI, etc. You may repeat the options that you have given with {{{compiler_options}}} here. See your local system documentation / software manuals for required path settings. Requirements differ from system to system and also depend on the respective libraries that you are using. See [wiki:doc/app/recommended_compiler_options] for specific path settings that we, the PALM group, are using on our computers. Be aware, that these settings probably will not work on your computer system. |
| 90 | || local_ip || IP-address of your local computer / the computer on which you call the {{{palmrun}}}/{{{palmbuild}}} command. You may use {{{127.0.0.0}}} if you are running PALM in interactive mode or in batch mode on your local computer. The address is only used to identify where to send the output data in case of batch jobs on a remote host. |
| 91 | || local_jobcatalog || Folder on the local host to store the batch job protocols. In case of batch jobs running on remote hosts, the job protocol will be created on the {{{remote_jobcatalog}}} and then be transferred via scp to the {{{local_jobcatalog}}}. |
| 92 | || local_username || Your username on the local computer / the computer on which you call the {{{palmrun}}}/{{{palmbuild}}} command. The local username is required for running batch jobs on a remote host in order to allow the batch job to access your local system (e.g. for sending back output data or for automatically starting restart runs). |
| 93 | || login_init_cmd || Special commands to be carried out at login or start of batch jobs on the remote host. \\ You may specify here a command, e.g. for setting up special system environments in batch jobs. It is carried out as first command in the batch job. |
| 94 | || make_options || Options for the UNIX {{{make}}}-command, which is used by {{{palmbuild}}} to compile the PALM code. In order to speed up compilation, you may use the {{{-j}}} option, which specifies the number of jobs to run simultaneously. If you have e.g. 4 cores on your local computer system, then {{{-j 4}}} starts 4 instances of the FORTRAN compiler, i.e. 4 FORTRAN-files are compiled simultaneously (if the dependencies allow for that). Do not try to start more instances than the number of available cores, because this will decrease the compiler performance significantly. |
| 95 | || memory || Memory request per MPI process (or CPU core) in MByte. \\ {{{palmrun}}} option{{{-m}}} overwrites this setting. |
| 96 | || module_commands || Module command(s) for loading required software / libraries. \\ In case that you have a {{{modules}}} package on your system, you can specify here the command(s) to load the specific software / libraries that your PALM run requires, e.g. the compiler, the NetCDF software, the MPI library, etc. Alternatively, you can load the modules from your shell profile (e.g. {{{.bashrc}}}), but then all your PALM runs will use the same settings. An example for a Cray system to use fftw and parallel NetCDF is {{{module load fftw cray-hdf5-parallel cray-netcdf-hdf5parallel}}}. The commands are carried out at the beginning of a batch job, or before PALM is compiled with {{{palmbuild}}}. |
| 97 | || remote_ip || IP-address of the remote system where the batch job shall be started. On large cluster systems this will usually be the address of a login node. Setting this variable in the configuration file will cause {{{palmrun}}} to run in remote batch job mode, i.e. a batch job will be created and send to the remote system automatically without giving {{{palmrun}}}-option {{{-b}}}. |
| 98 | || remote_jobcatalog || In case of batch jobs running on remote hosts, the job protocol will be put in this folder, and then automatically transferred via scp to the {{{local_jobcatalog}}}. The transfer is done by a separate small batch job, which directives are defined by the {{{BDT:}}} lines. The variable has no default value and must be set by the user. Absolute paths need to be given. Using {{{$HOME}}} is not allowed / does not work. |
| 99 | || remote_loginnode || Name of the login node of the remote computer. Nodes on big compute clusters are separated into compute nodes and login nodes (and sometimes I/O nodes). Some computer centers only allow the login nodes to establish ssh/scp connections to addresses outside the computing center. In such cases, since {{{palmrun}}} is executed on the compute nodes, it first has to send the output data to the login node, from where it is then forwarded to your local computer. If the compute nodes on your remote host do not allow direct ssh/scp connections to your local computer, you need to provide the name of the login node of the remote host. Typically, this is a mnemonic name like ''loginnode1'' and not an IP-address (like ''111.111.11.11''). Several login nodes often exist. You just have to give one of them. If you do not provide a name, you probably will not receive data on your local host from the PALM run. |
| 100 | || remote_username || Your username on the remote computer that is given by {{{remote_ip}}}. |
| 101 | || source_path || Path to PALM's FORTRAN source files. This is the place where the automatic installer has put the download, or which has been defined in the user's {{{svn checkout}}} command. |
| 102 | || ssh_key || Name of the file from which the identity (private key) for public key authentication is read. This file is assumed to be in folder {{{$HOME/.ssh}}}. By default (if you omit this variable), file {{{id_dsa}}} or {{{id_rsa}}} is used. |
| 103 | || submit_command || Full path to the command that has to be used to submit batch jobs on your system (either on the local, or on the remote host), including required option. See documentation of your batch system / computing center to find out which command has to be used. An example for a {{{moab}}} batch system could be {{{/opt/moab/default/bin/msub -E}}}. If you only know the command name (e.g. ''msub''), entering {{{which msub}}} on the local/remote host will give you the full path. |
| 104 | || user_source_path || Path to the [wiki:doc/app/userint user interface routines]. The variable {{{jobname}}} that may be used in the default path is replaced by the argument given with {{{palmrun}}}-option {{{-r}}}. |
| 105 | |
| 106 | |
| 107 | |
| 108 | |