source: palm/trunk/SOURCE/pmc_handle_communicator_mod.f90 @ 2794

Last change on this file since 2794 was 2718, checked in by maronga, 7 years ago

deleting of deprecated files; headers updated where needed

  • Property svn:keywords set to Id
File size: 19.2 KB
RevLine 
[2696]1!> @file pmc_handle_communicator_mod.f90
[2000]2!------------------------------------------------------------------------------!
[2696]3! This file is part of the PALM model system.
[1762]4!
[2000]5! PALM is free software: you can redistribute it and/or modify it under the
6! terms of the GNU General Public License as published by the Free Software
7! Foundation, either version 3 of the License, or (at your option) any later
8! version.
[1762]9!
10! PALM is distributed in the hope that it will be useful, but WITHOUT ANY
11! WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS FOR
12! A PARTICULAR PURPOSE.  See the GNU General Public License for more details.
13!
14! You should have received a copy of the GNU General Public License along with
15! PALM. If not, see <http://www.gnu.org/licenses/>.
16!
[2718]17! Copyright 1997-2018 Leibniz Universitaet Hannover
[2000]18!------------------------------------------------------------------------------!
[1762]19!
20! Current revisions:
21! ------------------
[1851]22!
[2516]23!
[1792]24! Former revisions:
25! -----------------
26! $Id: pmc_handle_communicator_mod.f90 2718 2018-01-02 08:49:38Z knoop $
[2716]27! Corrected "Former revisions" section
28!
29! 2696 2017-12-14 17:12:51Z kanani
30! Change in file header (GPL part)
[2696]31! Bugfix, give Intent(inout) attributes in routine reading nestpar-namelist (MS)
32!
33! 2599 2017-11-01 13:18:45Z hellstea
[2599]34! Separate peer communicator peer_comm introduced for MPI_INTERCOMM_CREATE.
35! Some cleanup and commenting improvements.
36!
37! 2516 2017-10-04 11:03:04Z suehring
[2516]38! Remove tabs
39!
40! 2514 2017-10-04 09:52:37Z suehring
[2280]41! Bugfix, set filepointer to the beginning of the file after namelist read,
42! in order to assure that further namelists are also found.
43!
44! 2279 2017-06-12 15:23:44Z suehring
[2271]45! Error message text changed
46!
47! 2101 2017-01-05 16:42:31Z suehring
[1792]48!
[2014]49! 2013 2016-09-21 13:07:56Z suehring
50! Bugfix in format descriptor
51!
[2001]52! 2000 2016-08-20 18:09:15Z knoop
53! Forced header and separation lines into 80 columns
54!
[1939]55! 1938 2016-06-13 15:26:05Z hellstea
56! Minor clean-up.
57!
[1933]58! 1901 2016-05-04 15:39:38Z raasch
59! Initial version of purely vertical nesting introduced.
60! Code clean up. The words server/client changed to parent/child.
61!
[1901]62! 1900 2016-05-04 15:27:53Z raasch
63! re-formatting to match PALM style
64!
[1883]65! 1882 2016-04-20 15:24:46Z hellstea
66! MPI_BCAST-calls to broadcast nesting_mode and nesting_datatransfer_mode
67! are moved out from the DO i = 1, m_ncpl loop.
68!
[1851]69! 1850 2016-04-08 13:29:27Z maronga
70! Module renamed
71!
[1809]72! 1808 2016-04-05 19:44:00Z raasch
73! MPI module used by default on all machines
74!
[1798]75! 1797 2016-03-21 16:50:28Z raasch
76! introduction of different datatransfer modes,
77! export of comm_world_nesting
78!
[1792]79! 1791 2016-03-11 10:41:25Z raasch
[1791]80! m_nrofcpl renamed m_ncpl,
81! pmc_get_local_model_info renamed pmc_get_model_info, some keywords also
82! renamed and some added,
83! debug write-statements commented out
[1765]84!
[1787]85! 1786 2016-03-08 05:49:27Z raasch
86! Bugfix: nesting_mode is broadcast now
87!
[1780]88! 1779 2016-03-03 08:01:28Z raasch
89! only the total number of PEs is given in the nestpar-NAMELIST,
90! additional comments included
91!
[1765]92! 1764 2016-02-28 12:45:19Z raasch
[1764]93! pmc_layout type: comm_cpl and comm_parent removed, character "name" moved at
94! the beginning of the variable list,
95! domain layout is read with new NAMELIST nestpar from standard file PARIN,
96! MPI-datatype REAL8 replaced by REAL, kind=8 replaced by wp,
97! variable domain_layouts instead of m_couplers introduced for this NAMELIST,
98! general format changed to PALM style
[1762]99!
[1763]100! 1762 2016-02-25 12:31:13Z hellstea
101! Initial revision by K. Ketelsen
102!
[1762]103! Description:
104! ------------
[1764]105! Handle MPI communicator in PALM model coupler
[1933]106!-------------------------------------------------------------------------------!
[2696]107 MODULE PMC_handle_communicator
[1764]108#if defined( __parallel )
109    USE kinds
[1762]110
[1808]111#if defined( __mpifh )
112    INCLUDE "mpif.h"
113#else
[1764]114    USE MPI
115#endif
[1762]116
[1933]117    USE pmc_general,                                                            &
[1900]118        ONLY: pmc_status_ok, pmc_status_error, pmc_max_models
[1933]119    USE control_parameters,                                                     &
120        ONLY: message_string
[1762]121
[1900]122    IMPLICIT NONE
[1762]123
[1900]124    TYPE pmc_layout
[1762]125
[1900]126       CHARACTER(LEN=32) ::  name
[1762]127
[1900]128       INTEGER  ::  id            !<
129       INTEGER  ::  parent_id     !<
130       INTEGER  ::  npe_total     !<
[1762]131
[1900]132       REAL(wp) ::  lower_left_x  !<
133       REAL(wp) ::  lower_left_y  !<
[1762]134
[1900]135    END TYPE pmc_layout
[1764]136
[1900]137    PUBLIC  pmc_status_ok, pmc_status_error
[1764]138
[2599]139    INTEGER, PARAMETER, PUBLIC ::  pmc_error_npes        = 1  !< illegal number of processes
[1900]140    INTEGER, PARAMETER, PUBLIC ::  pmc_namelist_error    = 2  !< error(s) in nestpar namelist
141    INTEGER, PARAMETER, PUBLIC ::  pmc_no_namelist_found = 3  !< no couple layout namelist found
[1764]142
[1900]143    INTEGER ::  m_world_comm  !< global nesting communicator
[2279]144    INTEGER ::  m_my_cpl_id   !< coupler id of this modelfortran return
[1900]145    INTEGER ::  m_parent_id   !< coupler id of parent of this model
146    INTEGER ::  m_ncpl        !< number of couplers given in nestpar namelist
[1762]147
[1900]148    TYPE(pmc_layout), DIMENSION(pmc_max_models) ::  m_couplers  !< information of all couplers
[1762]149
[1900]150    INTEGER, PUBLIC ::  m_model_comm          !< communicator of this model
[1933]151    INTEGER, PUBLIC ::  m_to_parent_comm      !< communicator to the parent
[1900]152    INTEGER, PUBLIC ::  m_world_rank          !<
153    INTEGER         ::  m_world_npes          !<
154    INTEGER, PUBLIC ::  m_model_rank          !<
155    INTEGER, PUBLIC ::  m_model_npes          !<
[2599]156    INTEGER         ::  m_parent_remote_size  !< number of processes in the parent model
157    INTEGER         ::  peer_comm             !< peer_communicator for inter communicators
[1791]158
[1933]159    INTEGER, DIMENSION(pmc_max_models), PUBLIC ::  m_to_child_comm    !< communicator to the child(ren)
160    INTEGER, DIMENSION(:), POINTER, PUBLIC ::  pmc_parent_for_child   !<
[1762]161
162
[1900]163    INTERFACE pmc_is_rootmodel
164       MODULE PROCEDURE pmc_is_rootmodel
165    END INTERFACE pmc_is_rootmodel
[1762]166
[1900]167    INTERFACE pmc_get_model_info
168       MODULE PROCEDURE pmc_get_model_info
169    END INTERFACE pmc_get_model_info
[1762]170
[1900]171    PUBLIC pmc_get_model_info, pmc_init_model, pmc_is_rootmodel
[1762]172
[1764]173 CONTAINS
[1762]174
[1933]175 SUBROUTINE pmc_init_model( comm, nesting_datatransfer_mode, nesting_mode,      &
[2696]176                            pmc_status )
[1762]177
[1933]178    USE control_parameters,                                                     &
[1900]179        ONLY:  message_string
[1764]180
[1933]181    USE pegrid,                                                                 &
[1900]182        ONLY:  myid
[1764]183
184      IMPLICIT NONE
185
[2696]186    CHARACTER(LEN=8), INTENT(INOUT) ::  nesting_mode               !<
187    CHARACTER(LEN=7), INTENT(INOUT) ::  nesting_datatransfer_mode  !<
[1764]188
[2696]189    INTEGER, INTENT(INOUT) ::  comm        !<
190    INTEGER, INTENT(INOUT) ::  pmc_status  !<
[1764]191
[1933]192    INTEGER ::  childcount     !<
[1900]193    INTEGER ::  i              !<
194    INTEGER ::  ierr           !<
195    INTEGER ::  istat          !<
196    INTEGER ::  m_my_cpl_rank  !<
197    INTEGER ::  tag            !<
[1764]198
[1933]199    INTEGER, DIMENSION(pmc_max_models)   ::  activeparent  ! I am active parent for this child ID
[1900]200    INTEGER, DIMENSION(pmc_max_models+1) ::  start_pe
[1762]201
[1900]202    pmc_status   = pmc_status_ok
203    comm         = -1
204    m_world_comm = MPI_COMM_WORLD
205    m_my_cpl_id  = -1
[1933]206    childcount   =  0
207    activeparent = -1
[1900]208    start_pe(:)  =  0
209
210    CALL MPI_COMM_RANK( MPI_COMM_WORLD, m_world_rank, istat )
211    CALL MPI_COMM_SIZE( MPI_COMM_WORLD, m_world_npes, istat )
[1764]212!
[2599]213!-- Only process 0 of root model reads
[1900]214    IF ( m_world_rank == 0 )  THEN
[1762]215
[1933]216       CALL read_coupling_layout( nesting_datatransfer_mode, nesting_mode,      &
[1900]217                                  pmc_status )
[1762]218
[1933]219       IF ( pmc_status /= pmc_no_namelist_found  .AND.                          &
220            pmc_status /= pmc_namelist_error )                                  &
[1900]221       THEN
[1764]222!
[2599]223!--       Determine the first process id of each model
[1900]224          start_pe(1) = 0
225          DO  i = 2, m_ncpl+1
226             start_pe(i) = start_pe(i-1) + m_couplers(i-1)%npe_total
227          ENDDO
[1762]228
[1764]229!
[2599]230!--       The sum of numbers of processes requested by all the domains
231!--       must be equal to the total number of processes of the run
[1900]232          IF ( start_pe(m_ncpl+1) /= m_world_npes )  THEN
[2013]233             WRITE ( message_string, '(2A,I6,2A,I6,A)' )                        &
[1933]234                             'nesting-setup requires different number of ',     &
235                             'MPI procs (', start_pe(m_ncpl+1), ') than ',      &
236                             'provided (', m_world_npes,')'
[1900]237             CALL message( 'pmc_init_model', 'PA0229', 3, 2, 0, 6, 0 )
238          ENDIF
[1762]239
[1900]240       ENDIF
[1762]241
[1900]242    ENDIF
[1764]243!
[2599]244!-- Broadcast the read status. This synchronises all other processes with
245!-- process 0 of the root model. Without synchronisation, they would not
246!-- behave in the correct way (e.g. they would not return in case of a
247!-- missing NAMELIST).
[1900]248    CALL MPI_BCAST( pmc_status, 1, MPI_INTEGER, 0, MPI_COMM_WORLD, istat )
[1762]249
[1900]250    IF ( pmc_status == pmc_no_namelist_found )  THEN
[1764]251!
[1900]252!--    Not a nested run; return the MPI_WORLD communicator
253       comm = MPI_COMM_WORLD
254       RETURN
[1762]255
[1900]256    ELSEIF ( pmc_status == pmc_namelist_error )  THEN
[1764]257!
[1900]258!--    Only the root model gives the error message. Others are aborted by the
259!--    message-routine with MPI_ABORT. Must be done this way since myid and
260!--    comm2d have not yet been assigned at this point.
261       IF ( m_world_rank == 0 )  THEN
262          message_string = 'errors in \$nestpar'
263          CALL message( 'pmc_init_model', 'PA0223', 3, 2, 0, 6, 0 )
264       ENDIF
[1762]265
[1900]266    ENDIF
[1762]267
[2599]268    CALL MPI_BCAST( m_ncpl,          1, MPI_INTEGER, 0, MPI_COMM_WORLD, istat )
269    CALL MPI_BCAST( start_pe, m_ncpl+1, MPI_INTEGER, 0, MPI_COMM_WORLD, istat )
[1764]270!
[1900]271!-- Broadcast coupling layout
272    DO  i = 1, m_ncpl
[1933]273       CALL MPI_BCAST( m_couplers(i)%name, LEN( m_couplers(i)%name ),           &
[1900]274                       MPI_CHARACTER, 0, MPI_COMM_WORLD, istat )
[1933]275       CALL MPI_BCAST( m_couplers(i)%id,           1, MPI_INTEGER, 0,           &
[1900]276                       MPI_COMM_WORLD, istat )
[1933]277       CALL MPI_BCAST( m_couplers(i)%Parent_id,    1, MPI_INTEGER, 0,           &
[1900]278                       MPI_COMM_WORLD, istat )
[1933]279       CALL MPI_BCAST( m_couplers(i)%npe_total,    1, MPI_INTEGER, 0,           &
[1900]280                       MPI_COMM_WORLD, istat )
[1933]281       CALL MPI_BCAST( m_couplers(i)%lower_left_x, 1, MPI_REAL,    0,           &
[1900]282                       MPI_COMM_WORLD, istat )
[1933]283       CALL MPI_BCAST( m_couplers(i)%lower_left_y, 1, MPI_REAL,    0,           &
[1900]284                       MPI_COMM_WORLD, istat )
285    ENDDO
[1933]286    CALL MPI_BCAST( nesting_mode, LEN( nesting_mode ), MPI_CHARACTER, 0,        &
[1900]287                    MPI_COMM_WORLD, istat )
[1933]288    CALL MPI_BCAST( nesting_datatransfer_mode, LEN(nesting_datatransfer_mode),  &
[1900]289                    MPI_CHARACTER, 0, MPI_COMM_WORLD, istat )
[1764]290!
[1900]291!-- Assign global MPI processes to individual models by setting the couple id
292    DO  i = 1, m_ncpl
[1933]293       IF ( m_world_rank >= start_pe(i)  .AND.  m_world_rank < start_pe(i+1) )  &
[1900]294       THEN
295          m_my_cpl_id = i
296          EXIT
297       ENDIF
298    ENDDO
299    m_my_cpl_rank = m_world_rank - start_pe(i)
[1764]300!
[1900]301!-- MPI_COMM_WORLD is the communicator for ALL models (MPI-1 approach).
302!-- The communictors for the individual models as created by MPI_COMM_SPLIT.
303!-- The color of the model is represented by the coupler id
[1933]304    CALL MPI_COMM_SPLIT( MPI_COMM_WORLD, m_my_cpl_id, m_my_cpl_rank, comm,      &
[1900]305                         istat )
[1764]306!
[2599]307!-- Get size and rank of the model running on this process
[1900]308    CALL  MPI_COMM_RANK( comm, m_model_rank, istat )
309    CALL  MPI_COMM_SIZE( comm, m_model_npes, istat )
[1764]310!
[2599]311!-- Broadcast (from process 0) the parent id and id of every model
[1900]312    DO  i = 1, m_ncpl
[1933]313       CALL MPI_BCAST( m_couplers(i)%parent_id, 1, MPI_INTEGER, 0,              &
[1900]314                       MPI_COMM_WORLD, istat )
[1933]315       CALL MPI_BCAST( m_couplers(i)%id,        1, MPI_INTEGER, 0,              &
[1900]316                       MPI_COMM_WORLD, istat )
317    ENDDO
[1764]318!
[1900]319!-- Save the current model communicator for pmc internal use
320    m_model_comm = comm
[1762]321
[1764]322!
[2599]323!-- Create intercommunicator between the parent and children.
[1900]324!-- MPI_INTERCOMM_CREATE creates an intercommunicator between 2 groups of
325!-- different colors.
[2599]326!-- The grouping was done above with MPI_COMM_SPLIT.
327!-- A duplicate of MPI_COMM_WORLD is created and used as peer communicator
328!-- (peer_comm) for MPI_INTERCOMM_CREATE.
329    CALL MPI_COMM_DUP( MPI_COMM_WORLD, peer_comm, ierr ) 
[1900]330    DO  i = 2, m_ncpl
331       IF ( m_couplers(i)%parent_id == m_my_cpl_id )  THEN
[1764]332!
[2599]333!--       Identify all children models of the current model and create
334!--       inter-communicators to connect between the current model and its
335!--       children models.
[1900]336          tag = 500 + i
[2599]337          CALL MPI_INTERCOMM_CREATE( comm, 0, peer_comm, start_pe(i),           &
[1933]338                                     tag, m_to_child_comm(i), istat)
339          childcount = childcount + 1
340          activeparent(i) = 1
[1900]341       ELSEIF ( i == m_my_cpl_id)  THEN
[1764]342!
[2599]343!--       Create an inter-communicator to connect between the current
344!--       model and its parent model.   
[1900]345          tag = 500 + i
[2599]346          CALL MPI_INTERCOMM_CREATE( comm, 0, peer_comm,                        &
[1933]347                                     start_pe(m_couplers(i)%parent_id),         &
348                                     tag, m_to_parent_comm, istat )
[1900]349       ENDIF
350    ENDDO
[1764]351!
[2599]352!-- If I am a parent, count the number of children I have.
[1933]353!-- Although this loop is symmetric on all processes, the "activeparent" flag
[2599]354!-- is true (==1) on the respective individual process only.
[1933]355    ALLOCATE( pmc_parent_for_child(childcount+1) )
[1762]356
[1933]357    childcount = 0
[1900]358    DO  i = 2, m_ncpl
[1933]359       IF ( activeparent(i) == 1 )  THEN
360          childcount = childcount + 1
361          pmc_parent_for_child(childcount) = i
[1900]362       ENDIF
363    ENDDO
[1764]364!
[1933]365!-- Get the size of the parent model
[1900]366    IF ( m_my_cpl_id > 1 )  THEN
[1933]367       CALL MPI_COMM_REMOTE_SIZE( m_to_parent_comm, m_parent_remote_size,       &
[2599]368                                  istat )
[1900]369    ELSE
[1764]370!
[1933]371!--    The root model does not have a parent
372       m_parent_remote_size = -1
[1900]373    ENDIF
[1764]374!
[2599]375!-- Set myid to non-zero value except for the root domain. This is a setting
[1900]376!-- for the message routine which is called at the end of pmci_init. That
377!-- routine outputs messages for myid = 0, only. However, myid has not been
[2599]378!-- assigened so far, so that all processes of the root model would output a
379!-- message. To avoid this, set myid to some other value except for process 0
380!-- of the root domain.
[1900]381    IF ( m_world_rank /= 0 )  myid = 1
[1762]382
[1900]383 END SUBROUTINE PMC_init_model
[1762]384
385
[1900]386
[1933]387 SUBROUTINE pmc_get_model_info( comm_world_nesting, cpl_id, cpl_name,           &
388                                cpl_parent_id, lower_left_x, lower_left_y,      &
[1900]389                                ncpl, npe_total, request_for_cpl_id )
[1764]390!
[1791]391!-- Provide module private variables of the pmc for PALM
[1762]392
[1900]393    USE kinds
[1762]394
[1900]395    IMPLICIT NONE
[1762]396
[1933]397    CHARACTER(LEN=*), INTENT(OUT), OPTIONAL ::  cpl_name   !<
[1762]398
[1933]399    INTEGER, INTENT(IN), OPTIONAL ::  request_for_cpl_id   !<
[1762]400
[1900]401    INTEGER, INTENT(OUT), OPTIONAL ::  comm_world_nesting  !<
402    INTEGER, INTENT(OUT), OPTIONAL ::  cpl_id              !<
403    INTEGER, INTENT(OUT), OPTIONAL ::  cpl_parent_id       !<
404    INTEGER, INTENT(OUT), OPTIONAL ::  ncpl                !<
405    INTEGER, INTENT(OUT), OPTIONAL ::  npe_total           !<
[1762]406
[2599]407    INTEGER ::  requested_cpl_id                           !<
[1764]408
[2599]409    REAL(wp), INTENT(OUT), OPTIONAL ::  lower_left_x       !<
410    REAL(wp), INTENT(OUT), OPTIONAL ::  lower_left_y       !<
[1764]411
[1791]412!
[1900]413!-- Set the requested coupler id
414    IF ( PRESENT( request_for_cpl_id ) )  THEN
415       requested_cpl_id = request_for_cpl_id
[1791]416!
[1900]417!--    Check for allowed range of values
418       IF ( requested_cpl_id < 1  .OR.  requested_cpl_id > m_ncpl )  RETURN
419    ELSE
420       requested_cpl_id = m_my_cpl_id
421    ENDIF
[1791]422!
[1900]423!-- Return the requested information
424    IF ( PRESENT( comm_world_nesting )  )  THEN
425       comm_world_nesting = m_world_comm
426    ENDIF
427    IF ( PRESENT( cpl_id )        )  THEN
428       cpl_id = requested_cpl_id
429    ENDIF
430    IF ( PRESENT( cpl_parent_id ) )  THEN
431       cpl_parent_id = m_couplers(requested_cpl_id)%parent_id
432    ENDIF
433    IF ( PRESENT( cpl_name )      )  THEN
434       cpl_name = m_couplers(requested_cpl_id)%name
435    ENDIF
436    IF ( PRESENT( ncpl )          )  THEN
437       ncpl = m_ncpl
438    ENDIF
439    IF ( PRESENT( npe_total )     )  THEN
440       npe_total = m_couplers(requested_cpl_id)%npe_total
441    ENDIF
442    IF ( PRESENT( lower_left_x )  )  THEN
443       lower_left_x = m_couplers(requested_cpl_id)%lower_left_x
444    ENDIF
445    IF ( PRESENT( lower_left_y )  )  THEN
446       lower_left_y = m_couplers(requested_cpl_id)%lower_left_y
447    ENDIF
[1791]448
[1900]449 END SUBROUTINE pmc_get_model_info
[1791]450
451
452
[1900]453 LOGICAL function pmc_is_rootmodel( )
[1764]454
[1900]455    IMPLICIT NONE
[1764]456
[1900]457    pmc_is_rootmodel = ( m_my_cpl_id == 1 )
[1764]458
[1900]459 END FUNCTION pmc_is_rootmodel
[1764]460
461
462
[1933]463 SUBROUTINE read_coupling_layout( nesting_datatransfer_mode, nesting_mode,      &
[2599]464      pmc_status )
[1762]465
[1764]466    IMPLICIT NONE
[1762]467
[1933]468    CHARACTER(LEN=8), INTENT(INOUT) ::  nesting_mode
[1797]469    CHARACTER(LEN=7), INTENT(INOUT) ::  nesting_datatransfer_mode
[1764]470
[1933]471    INTEGER(iwp), INTENT(INOUT) ::  pmc_status
472    INTEGER(iwp)                ::  bad_llcorner
473    INTEGER(iwp)                ::  i
474    INTEGER(iwp)                ::  istat
[1764]475
[1900]476    TYPE(pmc_layout), DIMENSION(pmc_max_models) ::  domain_layouts
[1764]477
[1797]478    NAMELIST /nestpar/  domain_layouts, nesting_datatransfer_mode, nesting_mode
[1764]479
480!
481!-- Initialize some coupling variables
[1900]482    domain_layouts(1:pmc_max_models)%id = -1
[1791]483    m_ncpl =   0
[1762]484
[1764]485    pmc_status = pmc_status_ok
486!
487!-- Open the NAMELIST-file and read the nesting layout
488    CALL check_open( 11 )
489    READ ( 11, nestpar, IOSTAT=istat )
[2279]490!
[2599]491!-- Set filepointer to the beginning of the file. Otherwise process 0 will later
[2279]492!-- be unable to read the inipar-NAMELIST
493    REWIND ( 11 )
[1764]494
495    IF ( istat < 0 )  THEN
496!
497!--    No nestpar-NAMELIST found
498       pmc_status = pmc_no_namelist_found
[1762]499       RETURN
[1764]500    ELSEIF ( istat > 0 )  THEN
501!
502!--    Errors in reading nestpar-NAMELIST
503       pmc_status = pmc_namelist_error
504       RETURN
505    ENDIF
506!
507!-- Output location message
508    CALL location_message( 'initialize communicators for nesting', .FALSE. )
509!
[2599]510!-- Assign the layout to the corresponding internally used variable m_couplers
[1764]511    m_couplers = domain_layouts
512!
513!-- Get the number of nested models given in the nestpar-NAMELIST
[1900]514    DO  i = 1, pmc_max_models
[1764]515!
[1791]516!--    When id=-1 is found for the first time, the list of domains is finished
[2599]517       IF ( m_couplers(i)%id == -1  .OR.  i == pmc_max_models )  THEN
[1791]518          IF ( m_couplers(i)%id == -1 )  THEN
519             m_ncpl = i - 1
520             EXIT
521          ELSE
[1900]522             m_ncpl = pmc_max_models
[1791]523          ENDIF
[1764]524       ENDIF
525    ENDDO
[1933]526!
527!-- Make sure that all domains have equal lower left corner in case of vertical
528!-- nesting
529    IF ( nesting_mode == 'vertical' )  THEN
530       bad_llcorner = 0
531       DO  i = 1, m_ncpl
532          IF ( domain_layouts(i)%lower_left_x /= 0.0_wp .OR.                    &
533               domain_layouts(i)%lower_left_y /= 0.0_wp )  THEN
534             bad_llcorner = bad_llcorner + 1
535             domain_layouts(i)%lower_left_x = 0.0_wp
536             domain_layouts(i)%lower_left_y = 0.0_wp
537          ENDIF
538       ENDDO
539       IF ( bad_llcorner /= 0)  THEN
[2271]540          WRITE ( message_string, *)  'at least one dimension of lower ',       &
[2514]541                                      'left corner of one domain is not 0. ',   &
[2271]542                                      'All lower left corners were set to (0, 0)'
[1933]543          CALL message( 'read_coupling_layout', 'PA0427', 0, 0, 0, 6, 0 )
544       ENDIF
545    ENDIF
546
[1764]547 END SUBROUTINE read_coupling_layout
548
549#endif
550 END MODULE pmc_handle_communicator
Note: See TracBrowser for help on using the repository browser.