Changes between Version 80 and Version 81 of doc/install
- Timestamp:
- Oct 6, 2016 6:11:02 PM (9 years ago)
Legend:
- Unmodified
- Added
- Removed
- Modified
-
doc/install
v80 v81 25 25 7. ssh/scp-connections to and from the remote host must not be blocked by a firewall. 26 26 27 Regarding batch support PALM is currently configured to be used on a limited number of selected machines. These are SGI-ICE systems at computing center HLRN in Hannover (''lc sgih''), Berlin (''lcsgib''), IBM-Regatta systems at Yonsei University (''ibmy''), and at DKRZ, Hamburg (''ibmh''), an NEC-SX8 system at RIAM, Kyushu University, Fukuoka (''necriam''), as well as on the Linux cluster of IMUK (''lcmuk''), Tokyo Institute of Technology (''lctit''), Kyoto computing center (''lckyoto''), and Cray-XE6 systems at the Bergen Center for Computational Science (''lcxe6'') and at the Finish Meteorological Institute (''lcxt5m''). The strings given in brackets are the systems names (host identifiers) under which '''mrun''' identifies the different hosts.\\\\27 Regarding batch support PALM is currently configured to be used on a limited number of selected machines. These are SGI-ICE systems at computing center HLRN in Hannover (''lccrayh''), Berlin (''lccrayb''), IBM-Regatta systems at Yonsei University (''ibmy''), and at DKRZ, Hamburg (''ibmh''), an NEC-SX8 system at RIAM, Kyushu University, Fukuoka (''necriam''), as well as on the Linux cluster of IMUK (''lcmuk''), Tokyo Institute of Technology (''lctit''), Kyoto computing center (''lckyoto''), and Cray-XE6 systems at the Bergen Center for Computational Science (''lcxe6'') and at the Finish Meteorological Institute (''lcxt5m''). The strings given in brackets are the systems names (host identifiers) under which '''mrun''' identifies the different hosts.\\\\ 28 28 You can also use '''mrun'''/PALM on other Linux-Cluster, IBM-AIX, or NEC-SX machines. See [#other_machines below] on how to configure mrun for other machines. However, these configurations currently (version 3.7a) allow to run PALM in interactive mode only. Batch mode requires manual adjustments for the respective queing system and MPI installation in scripts '''mrun''', '''mbuild''', and '''subjob'''.\\\\ 29 29 The examples given in this chapter refer to an installation of PALM on an IMUK Linux workstation and (for the [#advanced advanced method]) the SGI-ICE system of HLRN, used as remote host. They are just referred to as local and remote host from now on.\\\\ … … 165 165 {{{ 166 166 mbuild -u -h lcmuk 167 mbuild -u -h lc sgih167 mbuild -u -h lccrayh 168 168 }}} 169 169 The second call also copies the PALM scripts (like '''mrun''' and '''mbuild''') to the remote host.\\\\ … … 171 171 === [=#pre_compilation]Pre-compilation of PALM code === 172 172 173 To avoid the re-compilation of the complete source code for each model run, PALM willl be pre-compiled once on the remote host by again using the script '''mbuild'''. Due to the use of FORTRAN modules in the source code, the subroutines must be compiled in a certain order. Therefore the so-called ''make'' mechanism is used (see the respective man-page of the Unix operating system), requiring a {{{Makefile}}}, in which the dependencies are described. This file is found in subdirectory {{{trunk/SOURCE}}}, where also the PALM code is stored. The compiled sources (object files) are stored on the remote computer in the default directory {{{~/palm/current_version/MAKE_DEPOSITORY_<block_descriptor>}}}, where {{{<block_descriptor>}}} is composed of the third (and fourth, if existing) column of the respective block in the configuration file (e.g. {{{lc sgih_parallel}}} for HLRN).173 To avoid the re-compilation of the complete source code for each model run, PALM willl be pre-compiled once on the remote host by again using the script '''mbuild'''. Due to the use of FORTRAN modules in the source code, the subroutines must be compiled in a certain order. Therefore the so-called ''make'' mechanism is used (see the respective man-page of the Unix operating system), requiring a {{{Makefile}}}, in which the dependencies are described. This file is found in subdirectory {{{trunk/SOURCE}}}, where also the PALM code is stored. The compiled sources (object files) are stored on the remote computer in the default directory {{{~/palm/current_version/MAKE_DEPOSITORY_<block_descriptor>}}}, where {{{<block_descriptor>}}} is composed of the third (and fourth, if existing) column of the respective block in the configuration file (e.g. {{{lccrayh_parallel}}} for HLRN). 174 174 175 175 The pre-compilation for the remote host (here the SGI-ICE system of HLRN) is done by 176 176 {{{ 177 mbuild -h lc sgih177 mbuild -h lccrayh 178 178 }}} 179 179 'mbuild' will prompt some queries, which must all be answered "y" by the user. The compiling process will take some time. '''mbuild''' transfers the respective compiler calls to the remote host where they are carried out interactively. You can follow the progress at the terminal window, where also error messages are displayed (hopefully not for this standard installation). By just entering … … 196 196 The test run can now be started by executing the command 197 197 {{{ 198 mrun -d example_cbl -h lc sgih -K parallel -X 8 -T 8 -t 500 -q testq -r "d3# pr#"199 }}} 200 This specific run will be carried out on 8 PEs and is allowed to use up to 500 seconds CPU time. After pressing <return>, the most important settings of the job are displayed at the terminal window and the user is prompted for o.k. ("{{{y}}}"). Next, a message of the queuing system like "''Request … Submitted to queue… by…''" should be displayed. Now the job is queued and either started immediately or at a later time, depending on the current workload of the remote host. Provided that it is executed immediately and that all things work as designed, the job protocol of this run will appear under the file name {{{~/job_queue/lc sgih_example}}} no more than a few minutes later. The content of this file should be carefully examined for any error messages.\\\\198 mrun -d example_cbl -h lccrayh -K parallel -X 8 -T 8 -t 500 -q testq -r "d3# pr#" 199 }}} 200 This specific run will be carried out on 8 PEs and is allowed to use up to 500 seconds CPU time. After pressing <return>, the most important settings of the job are displayed at the terminal window and the user is prompted for o.k. ("{{{y}}}"). Next, a message of the queuing system like "''Request … Submitted to queue… by…''" should be displayed. Now the job is queued and either started immediately or at a later time, depending on the current workload of the remote host. Provided that it is executed immediately and that all things work as designed, the job protocol of this run will appear under the file name {{{~/job_queue/lccrayh_example}}} no more than a few minutes later. The content of this file should be carefully examined for any error messages.\\\\ 201 201 Beside the job protocol and according to the configuration file and arguments given for 'mrun' options {{{-d}}} and {{{-r}}}, further files should be found in the directories 202 202 {{{ … … 209 209 Please compare the contents of file 210 210 {{{ 211 ~/palm/current_version/JOBS/example_cbl/MONITORING/lc sgih_example_rc211 ~/palm/current_version/JOBS/example_cbl/MONITORING/lccrayh_example_rc 212 212 }}} 213 213 with those of the example result file which can be found under {{{trunk/INSTALL/example_cbl_rc}}}, e.g. by using the standard {{{diff}}} command 214 214 {{{ 215 diff JOBS/example_cbl/MONITORING/lc sgih_example_cbl_rc trunk/INSTALL/example_cbl_rc215 diff JOBS/example_cbl/MONITORING/lccrayh_example_cbl_rc trunk/INSTALL/example_cbl_rc 216 216 }}} 217 217 where it is assumed that your working directory is {{{~/palm/current_version}}}.\\\\