Changes between Version 80 and Version 81 of doc/install


Ignore:
Timestamp:
Oct 6, 2016 6:11:02 PM (9 years ago)
Author:
Giersch
Comment:

--

Legend:

Unmodified
Added
Removed
Modified
  • doc/install

    v80 v81  
    2525 7. ssh/scp-connections to and from the remote host must not be blocked by a firewall.
    2626
    27 Regarding batch support PALM is currently configured to be used on a limited number of selected machines. These are SGI-ICE systems at computing center HLRN in Hannover (''lcsgih''), Berlin (''lcsgib''), IBM-Regatta systems at Yonsei University (''ibmy''), and at DKRZ, Hamburg (''ibmh''), an NEC-SX8 system at RIAM, Kyushu University, Fukuoka (''necriam''), as well as on the Linux cluster of IMUK (''lcmuk''), Tokyo Institute of Technology (''lctit''), Kyoto computing center (''lckyoto''), and Cray-XE6 systems at the Bergen Center for Computational Science (''lcxe6'') and at the Finish Meteorological Institute (''lcxt5m''). The strings given in brackets are the systems names (host identifiers) under which '''mrun''' identifies the different hosts.\\\\
     27Regarding batch support PALM is currently configured to be used on a limited number of selected machines. These are SGI-ICE systems at computing center HLRN in Hannover (''lccrayh''), Berlin (''lccrayb''), IBM-Regatta systems at Yonsei University (''ibmy''), and at DKRZ, Hamburg (''ibmh''), an NEC-SX8 system at RIAM, Kyushu University, Fukuoka (''necriam''), as well as on the Linux cluster of IMUK (''lcmuk''), Tokyo Institute of Technology (''lctit''), Kyoto computing center (''lckyoto''), and Cray-XE6 systems at the Bergen Center for Computational Science (''lcxe6'') and at the Finish Meteorological Institute (''lcxt5m''). The strings given in brackets are the systems names (host identifiers) under which '''mrun''' identifies the different hosts.\\\\
    2828You can also use '''mrun'''/PALM on other Linux-Cluster, IBM-AIX, or NEC-SX machines. See [#other_machines below] on how to configure mrun for other machines. However, these configurations currently (version 3.7a) allow to run PALM in interactive mode only. Batch mode requires manual adjustments for the respective queing system and MPI installation in scripts '''mrun''', '''mbuild''', and '''subjob'''.\\\\
    2929The examples given in this chapter refer to an installation of PALM on an IMUK Linux workstation and (for the [#advanced advanced method]) the SGI-ICE system of HLRN, used as remote host. They are just referred to as local and remote host from now on.\\\\
     
    165165{{{
    166166  mbuild -u -h lcmuk
    167   mbuild -u -h lcsgih
     167  mbuild -u -h lccrayh
    168168}}}
    169169The second call also copies the PALM scripts (like '''mrun''' and '''mbuild''') to the remote host.\\\\
     
    171171=== [=#pre_compilation]Pre-compilation of PALM code ===
    172172
    173 To avoid the re-compilation of the complete source code for each model run, PALM willl be pre-compiled once on the remote host by again using the script '''mbuild'''. Due to the use of FORTRAN modules in the source code, the subroutines must be compiled in a certain order. Therefore the so-called ''make'' mechanism is used (see the respective man-page of the Unix operating system), requiring a {{{Makefile}}}, in which the dependencies are described. This file is found in subdirectory {{{trunk/SOURCE}}}, where also the PALM code is stored. The compiled sources (object files) are stored on the remote computer in the default directory {{{~/palm/current_version/MAKE_DEPOSITORY_<block_descriptor>}}}, where {{{<block_descriptor>}}} is composed of the third (and fourth, if existing) column of the respective block in the configuration file (e.g. {{{lcsgih_parallel}}} for HLRN).
     173To avoid the re-compilation of the complete source code for each model run, PALM willl be pre-compiled once on the remote host by again using the script '''mbuild'''. Due to the use of FORTRAN modules in the source code, the subroutines must be compiled in a certain order. Therefore the so-called ''make'' mechanism is used (see the respective man-page of the Unix operating system), requiring a {{{Makefile}}}, in which the dependencies are described. This file is found in subdirectory {{{trunk/SOURCE}}}, where also the PALM code is stored. The compiled sources (object files) are stored on the remote computer in the default directory {{{~/palm/current_version/MAKE_DEPOSITORY_<block_descriptor>}}}, where {{{<block_descriptor>}}} is composed of the third (and fourth, if existing) column of the respective block in the configuration file (e.g. {{{lccrayh_parallel}}} for HLRN).
    174174
    175175The pre-compilation for the remote host (here the SGI-ICE system of HLRN) is done by
    176176{{{
    177   mbuild -h lcsgi
     177  mbuild -h lccray
    178178}}}
    179179'mbuild' will prompt some queries, which must all be answered "y" by the user. The compiling process will take some time. '''mbuild''' transfers the respective compiler calls to the remote host where they are carried out interactively. You can follow the progress at the terminal window, where also error messages are displayed (hopefully not for this standard installation). By just entering
     
    196196The test run can now be started by executing the command
    197197{{{
    198   mrun -d example_cbl -h lcsgih -K parallel -X 8 -T 8 -t 500 -q testq -r "d3# pr#"
    199 }}}
    200 This specific run will be carried out on 8 PEs and is allowed to use up to 500 seconds CPU time. After pressing <return>, the most important settings of the job are displayed at the terminal window and the user is prompted for o.k. ("{{{y}}}"). Next, a message of the queuing system like "''Request … Submitted to queue… by…''" should be displayed. Now the job is queued and either started immediately or at a later time, depending on the current workload of the remote host. Provided that it is executed immediately and that all things work as designed, the job protocol of this run will appear under the file name {{{~/job_queue/lcsgih_example}}} no more than a few minutes later. The content of this file should be carefully examined for any error messages.\\\\
     198  mrun -d example_cbl -h lccrayh -K parallel -X 8 -T 8 -t 500 -q testq -r "d3# pr#"
     199}}}
     200This specific run will be carried out on 8 PEs and is allowed to use up to 500 seconds CPU time. After pressing <return>, the most important settings of the job are displayed at the terminal window and the user is prompted for o.k. ("{{{y}}}"). Next, a message of the queuing system like "''Request … Submitted to queue… by…''" should be displayed. Now the job is queued and either started immediately or at a later time, depending on the current workload of the remote host. Provided that it is executed immediately and that all things work as designed, the job protocol of this run will appear under the file name {{{~/job_queue/lccrayh_example}}} no more than a few minutes later. The content of this file should be carefully examined for any error messages.\\\\
    201201Beside the job protocol and according to the configuration file and arguments given for 'mrun' options {{{-d}}} and {{{-r}}}, further files should be found in the directories
    202202{{{
     
    209209Please compare the contents of file
    210210{{{
    211   ~/palm/current_version/JOBS/example_cbl/MONITORING/lcsgih_example_rc
     211  ~/palm/current_version/JOBS/example_cbl/MONITORING/lccrayh_example_rc
    212212}}}
    213213with those of the example result file which can be found under {{{trunk/INSTALL/example_cbl_rc}}}, e.g. by using the standard {{{diff}}} command
    214214{{{
    215 diff  JOBS/example_cbl/MONITORING/lcsgih_example_cbl_rc trunk/INSTALL/example_cbl_rc
     215diff  JOBS/example_cbl/MONITORING/lccrayh_example_cbl_rc trunk/INSTALL/example_cbl_rc
    216216}}}
    217217where it is assumed that your working directory is {{{~/palm/current_version}}}.\\\\