source: palm/trunk/SOURCE/palm.f90 @ 3759

Last change on this file since 3759 was 3719, checked in by kanani, 6 years ago

Correct and clean-up cpu_logs, some overlapping counts (chemistry_model_mod, disturb_heatflux, large_scale_forcing_nudging_mod, ocean_mod, palm, prognostic_equations, synthetic_turbulence_generator_mod, time_integration, time_integration_spinup, turbulence_closure_mod)

  • Property svn:keywords set to Id
File size: 21.1 KB
Line 
1!> @file palm.f90
2!------------------------------------------------------------------------------!
3! This file is part of the PALM model system.
4!
5! PALM is free software: you can redistribute it and/or modify it under the
6! terms of the GNU General Public License as published by the Free Software
7! Foundation, either version 3 of the License, or (at your option) any later
8! version.
9!
10! PALM is distributed in the hope that it will be useful, but WITHOUT ANY
11! WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS FOR
12! A PARTICULAR PURPOSE.  See the GNU General Public License for more details.
13!
14! You should have received a copy of the GNU General Public License along with
15! PALM. If not, see <http://www.gnu.org/licenses/>.
16!
17! Copyright 1997-2019 Leibniz Universitaet Hannover
18!------------------------------------------------------------------------------!
19!
20! Current revisions:
21! -----------------
22!
23!
24! Former revisions:
25! -----------------
26! $Id: palm.f90 3719 2019-02-06 13:10:18Z suehring $
27! Included cpu measurement for wall/soil spinup
28!
29! 3703 2019-01-29 16:43:53Z knoop
30! Some interface calls moved to module_interface + cleanup
31!
32! 3648 2019-01-02 16:35:46Z suehring
33! Rename subroutines for surface-data output
34!
35! 3524 2018-11-14 13:36:44Z raasch
36! unused variable removed
37!
38! 3494 2018-11-06 14:51:27Z suehring
39! Last actions for surface output added
40!
41! 3487 2018-11-05 07:18:02Z maronga
42! Updates version to 6.0
43!
44! 3484 2018-11-02 14:41:25Z hellstea
45! pmci_ensure_nest_mass_conservation removed permanently
46!
47! 3458 2018-10-30 14:51:23Z kanani
48! from chemistry branch r3443, forkel:
49! removed double do_emis check around CALL chem_init
50! replaced call to calc_date_and_time to init_date_and_time
51!
52! 3337 2018-10-12 15:17:09Z kanani
53! (from branch resler)
54! Fix chemistry call
55!
56! 3298 2018-10-02 12:21:11Z kanani
57! - Minor formatting (kanani)
58! - Added Call of date_and_time_init (Russo)
59! - Added Call of calc_date_and_time before call of init_3d where emissions
60!   are initialized:
61!   we have to know the time indices to initialize emission values (Russo)
62! - Added Call of netcdf_data_input_chemistry_data (Russo)
63!
64! 3274 2018-09-24 15:42:55Z knoop
65! Modularization of all bulk cloud physics code components
66!
67! 3258 2018-09-18 07:23:31Z Giersch
68! current revision for user interface has been changed to 3240
69!
70! 3241 2018-09-12 15:02:00Z raasch
71! unused variables removed
72!
73! 3235 2018-09-07 14:06:15Z sward
74! Added mas_last_actions call and multi_agent_system_mod dependency
75!
76! 3232 2018-09-07 12:21:44Z raasch
77! references to mrun replaced by palmrun, and updated
78!
79! 3182 2018-07-27 13:36:03Z suehring
80! Deduct spinup_time from RUN_CONTROL output of main 3d run
81! (use time_since_reference_point instead of simulated_time)
82!
83! 2951 2018-04-06 09:05:08Z kanani
84! Add log_point_s for pmci_init
85!
86! 2903 2018-03-16 08:17:06Z hellstea
87! Nesting-related calls to pmci_ensure_nest_mass_conservation and pres after
88! the nest initialization are removed as they may create unwanted initial
89! perturbation in some cases.
90!
91! 2894 2018-03-15 09:17:58Z Giersch
92! Modified todo list, _write_restart_data/_last_actions has been renamed to
93! _wrd_local, unit 14 will be opened now for each io_group
94! write_3d_binary is called wrd_local now, wrd_global moved from wrd_local to
95! palm.f90, unit 14 is closed directly after the wrd_local call, Module related
96! routines for writing restart data have been moved to wrd_local
97!
98! 2801 2018-02-14 16:01:55Z suehring
99! Changed lpm from subroutine to module.
100! Introduce particle transfer in nested models.
101!
102! 2766 2018-01-22 17:17:47Z kanani
103! Removed preprocessor directive __chem
104!
105! 2720 2018-01-02 16:27:15Z kanani
106! Version update to 5.0
107!
108! 2718 2018-01-02 08:49:38Z maronga
109! Corrected "Former revisions" section
110!
111! 2696 2017-12-14 17:12:51Z kanani
112! Change in file header (GPL part)
113! Implementation of chemistry module (FK)
114! Introduce input-data standard
115! Rename lsm_last_actions into lsm_write_restart_data
116! Move usm_write_restart_data into io_blocks loop (MS)
117!
118! 2512 2017-10-04 08:26:59Z raasch
119! user interface required revision updated
120!
121! 2320 2017-07-21 12:47:43Z suehring
122! Modularize large-scale forcing and nudging
123!
124! 2298 2017-06-29 09:28:18Z raasch
125! type of write_binary changed from CHARACTER to LOGICAL,
126! user interface required revision updated, MPI2 related part removed
127!
128! 2296 2017-06-28 07:53:56Z maronga
129! Added call to new spinup routine
130!
131! 2292 2017-06-20 09:51:42Z schwenkel
132! Implementation of new microphysic scheme: cloud_scheme = 'morrison'
133! includes two more prognostic equations for cloud drop concentration (nc) 
134! and cloud water content (qc).
135!
136! 2261 2017-06-08 14:25:57Z raasch
137! output of run number for mrun to create unified cycle numbers
138!
139! 2233 2017-05-30 18:08:54Z suehring
140!
141! 2232 2017-05-30 17:47:52Z suehring
142! Renamed wall_flags_0 and wall_flags_00 into advc_flags_1 and advc_flags_2,
143! respectively, within copyin statement. Moreover, introduced further flag
144! array wall_flags_0.
145! Remove unused variables from ONLY list.
146!
147! 2178 2017-03-17 11:07:39Z hellstea
148! Calls for pmci_ensure_nest_mass_conservation and pres are added after
149! the nest initialization
150!
151! 2118 2017-01-17 16:38:49Z raasch
152! OpenACC directives and related code removed
153!
154! 2011 2016-09-19 17:29:57Z kanani
155! Flag urban_surface is now defined in module control_parameters.
156!
157! 2007 2016-08-24 15:47:17Z kanani
158! Temporarily added CALL for writing of restart data for urban surface model
159!
160! 2000 2016-08-20 18:09:15Z knoop
161! Forced header and separation lines into 80 columns
162!
163! 1976 2016-07-27 13:28:04Z maronga
164! Added call to radiation_last_actions for binary output of land surface model
165! data
166!
167! 1972 2016-07-26 07:52:02Z maronga
168! Added call to lsm_last_actions for binary output of land surface model data
169!
170! 1960 2016-07-12 16:34:24Z suehring
171! Separate humidity and passive scalar
172!
173! 1834 2016-04-07 14:34:20Z raasch
174! Initial version of purely vertical nesting introduced.
175!
176! 1833 2016-04-07 14:23:03Z raasch
177! required user interface version changed
178!
179! 1808 2016-04-05 19:44:00Z raasch
180! routine local_flush replaced by FORTRAN statement
181!
182! 1783 2016-03-06 18:36:17Z raasch
183! required user interface version changed
184!
185! 1781 2016-03-03 15:12:23Z raasch
186! pmc initialization moved from time_integration to here
187!
188! 1779 2016-03-03 08:01:28Z raasch
189! setting of nest_domain and coupling_char moved to the pmci
190!
191! 1764 2016-02-28 12:45:19Z raasch
192! cpp-statements for nesting removed, communicator settings cleaned up
193!
194! 1762 2016-02-25 12:31:13Z hellstea
195! Introduction of nested domain feature
196!
197! 1747 2016-02-08 12:25:53Z raasch
198! OpenACC-adjustment for new surface layer parameterization
199!
200! 1682 2015-10-07 23:56:08Z knoop
201! Code annotations made doxygen readable
202!
203! 1668 2015-09-23 13:45:36Z raasch
204! warning replaced by abort in case of failed user interface check
205!
206! 1666 2015-09-23 07:31:10Z raasch
207! check for user's interface version added
208!
209! 1482 2014-10-18 12:34:45Z raasch
210! adjustments for using CUDA-aware OpenMPI
211!
212! 1468 2014-09-24 14:06:57Z maronga
213! Adapted for use on up to 6-digit processor cores
214!
215! 1402 2014-05-09 14:25:13Z raasch
216! location messages added
217!
218! 1374 2014-04-25 12:55:07Z raasch
219! bugfix: various modules added
220!
221! 1320 2014-03-20 08:40:49Z raasch
222! ONLY-attribute added to USE-statements,
223! kind-parameters added to all INTEGER and REAL declaration statements,
224! kinds are defined in new module kinds,
225! old module precision_kind is removed,
226! revision history before 2012 removed,
227! comment fields (!:) to be used for variable explanations added to
228! all variable declaration statements
229!
230! 1318 2014-03-17 13:35:16Z raasch
231! module interfaces removed
232!
233! 1241 2013-10-30 11:36:58Z heinze
234! initialization of nuding and large scale forcing from external file
235!
236! 1221 2013-09-10 08:59:13Z raasch
237! +wall_flags_00, rflags_invers, rflags_s_inner in copyin statement
238!
239! 1212 2013-08-15 08:46:27Z raasch
240! +tri in copyin statement
241!
242! 1179 2013-06-14 05:57:58Z raasch
243! ref_state added to copyin-list
244!
245! 1113 2013-03-10 02:48:14Z raasch
246! openACC statements modified
247!
248! 1111 2013-03-08 23:54:10Z raasch
249! openACC statements updated
250!
251! 1092 2013-02-02 11:24:22Z raasch
252! unused variables removed
253!
254! 1036 2012-10-22 13:43:42Z raasch
255! code put under GPL (PALM 3.9)
256!
257! 1015 2012-09-27 09:23:24Z raasch
258! Version number changed from 3.8 to 3.8a.
259! OpenACC statements added + code changes required for GPU optimization
260!
261! 849 2012-03-15 10:35:09Z raasch
262! write_particles renamed lpm_write_restart_file
263!
264! Revision 1.1  1997/07/24 11:23:35  raasch
265! Initial revision
266!
267!
268! Description:
269! ------------
270!> Large-Eddy Simulation (LES) model for atmospheric and oceanic boundary-layer
271!> flows
272!> see the PALM homepage https://palm-model.org for further information
273!------------------------------------------------------------------------------!
274 PROGRAM palm
275 
276
277    USE arrays_3d
278
279    USE bulk_cloud_model_mod,                                                  &
280        ONLY: bulk_cloud_model, microphysics_morrison, microphysics_seifert
281
282    USE control_parameters,                                                    &
283        ONLY:  air_chemistry, constant_diffusion, child_domain,                &
284               coupling_char, do2d_at_begin, do3d_at_begin, humidity,          &
285               initializing_actions, io_blocks, io_group, message_string,      &
286               neutral, passive_scalar, runnr, simulated_time_chr, spinup,     &
287               time_since_reference_point, user_interface_current_revision,    &
288               user_interface_required_revision, version, write_binary
289
290    USE cpulog,                                                                &
291        ONLY:  cpu_log, log_point, log_point_s, cpu_statistics
292
293    USE date_and_time_mod,                                                     &
294        ONLY:  calc_date_and_time, init_date_and_time
295
296    USE indices,                                                               &
297        ONLY:  nbgp
298
299    USE kinds
300
301    USE module_interface,                                                      &
302        ONLY:  module_interface_last_actions
303
304    USE multi_agent_system_mod,                                                &
305        ONLY:  agents_active, mas_last_actions
306
307    USE netcdf_data_input_mod,                                                 &
308        ONLY:  netcdf_data_input_inquire_file, netcdf_data_input_init,         &
309               netcdf_data_input_surface_data, netcdf_data_input_topo
310
311    USE particle_attributes,                                                   &
312        ONLY:  particle_advection
313
314    USE pegrid
315
316    USE pmc_particle_interface,                                                &
317        ONLY: pmcp_g_alloc_win
318
319    USE pmc_interface,                                                         &
320        ONLY:  nested_run, pmci_child_initialize, pmci_init,                   &
321               pmci_modelconfiguration, pmci_parent_initialize
322               
323    USE surface_data_output_mod,                                               &
324        ONLY:  surface_data_output_last_action
325
326    USE write_restart_data_mod,                                                &
327        ONLY:  wrd_global, wrd_local
328
329#if defined( __parallel) && defined( _OPENACC )
330    USE openacc
331#endif
332
333
334    IMPLICIT NONE
335
336!
337!-- Local variables
338    CHARACTER(LEN=9) ::  time_to_string  !<
339    INTEGER(iwp)     ::  i               !< loop counter for blocked I/O
340#if defined( __parallel) && defined( _OPENACC )
341    INTEGER(iwp)     :: local_comm       !< local communicator (shared memory)
342    INTEGER(iwp)     :: local_num_procs  !< local number of processes
343    INTEGER(iwp)     :: local_id         !< local id
344    INTEGER(acc_device_kind) :: device_type !< device type for OpenACC
345    INTEGER(iwp)     ::  num_devices     !< number of devices visible to OpenACC
346    INTEGER(iwp)     ::  my_device       !< device used by this process
347#endif
348
349    version = 'PALM 6.0'
350    user_interface_required_revision = 'r3703'
351
352#if defined( __parallel )
353!
354!-- MPI initialisation. comm2d is preliminary set, because
355!-- it will be defined in init_pegrid but is used before in cpu_log.
356    CALL MPI_INIT( ierr )
357
358!
359!-- Initialize the coupling for nested-domain runs
360!-- comm_palm is the communicator which includes all PEs (MPI processes)
361!-- available for this (nested) model. If it is not a nested run, comm_palm
362!-- is returned as MPI_COMM_WORLD
363    CALL cpu_log( log_point_s(70), 'pmci_init', 'start' )
364    CALL pmci_init( comm_palm )
365    CALL cpu_log( log_point_s(70), 'pmci_init', 'stop' )
366    comm2d = comm_palm
367!
368!-- Get the (preliminary) number of MPI processes and the local PE-id (in case
369!-- of a further communicator splitting in init_coupling, these numbers will
370!-- be changed in init_pegrid).
371    IF ( nested_run )  THEN
372
373       CALL MPI_COMM_SIZE( comm_palm, numprocs, ierr )
374       CALL MPI_COMM_RANK( comm_palm, myid, ierr )
375
376    ELSE
377
378       CALL MPI_COMM_SIZE( MPI_COMM_WORLD, numprocs, ierr )
379       CALL MPI_COMM_RANK( MPI_COMM_WORLD, myid, ierr )
380!
381!--    Initialize PE topology in case of coupled atmosphere-ocean runs (comm_palm
382!--    will be splitted in init_coupling)
383       CALL init_coupling
384    ENDIF
385
386#ifdef _OPENACC
387!
388!-- Select OpenACC device to use in this process. For this find out how many
389!-- neighbors there are running on the same node and which id this process is.
390    IF ( nested_run )  THEN
391       CALL MPI_COMM_SPLIT_TYPE( comm_palm, MPI_COMM_TYPE_SHARED, 0,           &
392                                 MPI_INFO_NULL, local_comm, ierr )
393    ELSE
394       CALL MPI_COMM_SPLIT_TYPE( MPI_COMM_WORLD, MPI_COMM_TYPE_SHARED, 0,      &
395                                 MPI_INFO_NULL, local_comm, ierr )
396    ENDIF
397    CALL MPI_COMM_SIZE( local_comm, local_num_procs, ierr )
398    CALL MPI_COMM_RANK( local_comm, local_id, ierr )
399
400!
401!-- This loop including the barrier is a workaround for PGI compiler versions
402!-- up to and including 18.4. Later releases are able to select their GPUs in
403!-- parallel, without running into spurious errors.
404    DO i = 0, local_num_procs-1
405       CALL MPI_BARRIER( local_comm, ierr )
406
407       IF ( i == local_id )  THEN
408          device_type = acc_get_device_type()
409          num_devices = acc_get_num_devices( device_type )
410          my_device = MOD( local_id, num_devices )
411          CALL acc_set_device_num( my_device, device_type )
412       ENDIF
413    ENDDO
414
415    CALL MPI_COMM_FREE( local_comm, ierr )
416#endif
417#endif
418
419!
420!-- Initialize measuring of the CPU-time remaining to the run
421    CALL local_tremain_ini
422
423!
424!-- Start of total CPU time measuring.
425    CALL cpu_log( log_point(1), 'total', 'start' )
426    CALL cpu_log( log_point(2), 'initialisation', 'start' )
427
428!
429!-- Open a file for debug output
430    WRITE (myid_char,'(''_'',I6.6)')  myid
431    OPEN( 9, FILE='DEBUG'//TRIM( coupling_char )//myid_char, FORM='FORMATTED' )
432
433!
434!-- Initialize dvrp logging. Also, one PE maybe split from the global
435!-- communicator for doing the dvrp output. In that case, the number of
436!-- PEs available for PALM is reduced by one and communicator comm_palm
437!-- is changed respectively.
438#if defined( __parallel )
439    CALL MPI_COMM_RANK( comm_palm, myid, ierr )
440#endif
441
442    CALL init_dvrp_logging
443
444!
445!-- Read control parameters from NAMELIST files and read environment-variables
446    CALL parin
447
448!
449!-- Check for the user's interface version
450    IF ( user_interface_current_revision /= user_interface_required_revision )  &
451    THEN
452       message_string = 'current user-interface revision "' //                  &
453                        TRIM( user_interface_current_revision ) // '" does ' // &
454                        'not match the required revision ' //                   &
455                        TRIM( user_interface_required_revision )
456        CALL message( 'palm', 'PA0169', 1, 2, 0, 6, 0 )
457    ENDIF
458
459!
460!-- Determine processor topology and local array indices
461    CALL init_pegrid
462!
463!-- Check if input file according to input-data standard exists
464    CALL netcdf_data_input_inquire_file
465!
466!-- Read topography input data if required. This is required before the
467!-- numerical grid is finally created in init_grid
468    CALL netcdf_data_input_topo 
469!
470!-- Generate grid parameters, initialize generic topography and further process
471!-- topography information if required
472    CALL init_grid
473!
474!-- Read global attributes if available. 
475    CALL netcdf_data_input_init 
476!
477!-- Read surface classification data, e.g. vegetation and soil types, water
478!-- surfaces, etc., if available. Some of these data is required before
479!-- check parameters is invoked.     
480    CALL netcdf_data_input_surface_data
481!
482!-- Check control parameters and deduce further quantities
483    CALL check_parameters
484!
485!-- Initial time for chem_emissions_mod
486    CALL init_date_and_time
487
488    CALL init_3d_model
489
490!
491!-- Coupling protocol setup for nested-domain runs
492    IF ( nested_run )  THEN
493       CALL pmci_modelconfiguration
494!
495!--    Receive and interpolate initial data on children.
496!--    Child initialization must be made first if the model is both child and
497!--    parent if necessary
498       IF ( TRIM( initializing_actions ) /= 'read_restart_data' )  THEN
499          CALL pmci_child_initialize
500!
501!--       Send initial condition data from parent to children
502          CALL pmci_parent_initialize
503!
504!--       Exchange_horiz is needed after the nest initialization
505          IF ( child_domain )  THEN
506             CALL exchange_horiz( u, nbgp )
507             CALL exchange_horiz( v, nbgp )
508             CALL exchange_horiz( w, nbgp )
509             IF ( .NOT. neutral )  THEN
510                CALL exchange_horiz( pt, nbgp )
511             ENDIF
512             IF ( .NOT. constant_diffusion )  CALL exchange_horiz( e, nbgp )
513             IF ( humidity )  THEN
514                CALL exchange_horiz( q, nbgp )
515                IF ( bulk_cloud_model  .AND.  microphysics_morrison )  THEN
516                  CALL exchange_horiz( qc, nbgp )
517                  CALL exchange_horiz( nc, nbgp )
518                ENDIF
519                IF ( bulk_cloud_model  .AND.  microphysics_seifert )  THEN
520                   CALL exchange_horiz( qr, nbgp ) 
521                   CALL exchange_horiz( nr, nbgp )
522                ENDIF
523             ENDIF
524             IF ( passive_scalar )  CALL exchange_horiz( s, nbgp )
525          ENDIF
526       ENDIF
527
528       CALL pmcp_g_alloc_win                    ! Must be called after pmci_child_initialize and pmci_parent_initialize
529    ENDIF
530
531!
532!-- Output of program header
533    IF ( myid == 0 )  CALL header
534
535    CALL cpu_log( log_point(2), 'initialisation', 'stop' )
536
537!
538!-- Integration of the non-atmospheric equations (land surface model, urban
539!-- surface model)
540    IF ( spinup )  THEN
541       CALL cpu_log( log_point(41), 'wall/soil spinup', 'start' )
542       CALL time_integration_spinup
543       CALL cpu_log( log_point(41), 'wall/soil spinup', 'stop' )
544    ENDIF
545
546!
547!-- Set start time in format hh:mm:ss
548    simulated_time_chr = time_to_string( time_since_reference_point )
549
550!
551!-- If required, output of initial arrays
552    IF ( do2d_at_begin )  THEN
553       CALL data_output_2d( 'xy', 0 )
554       CALL data_output_2d( 'xz', 0 )
555       CALL data_output_2d( 'yz', 0 )
556    ENDIF
557
558    IF ( do3d_at_begin )  THEN
559       CALL data_output_3d( 0 )
560    ENDIF
561
562!
563!-- Integration of the model equations using timestep-scheme
564    CALL time_integration
565
566!
567!-- If required, write binary data for restart runs
568    IF ( write_binary )  THEN
569
570       CALL cpu_log( log_point(22), 'wrd_local', 'start' )
571
572       CALL location_message( 'writing restart data', .FALSE. )
573
574       DO  i = 0, io_blocks-1
575          IF ( i == io_group )  THEN
576
577!
578!--          Open binary file
579             CALL check_open( 14 )
580!
581!--          Write control parameters and other global variables for restart.
582             IF ( myid == 0 )  CALL wrd_global
583!
584!--          Write processor specific flow field data for restart runs
585             CALL wrd_local
586!
587!--          Close binary file
588             CALL close_file( 14 )
589
590          ENDIF
591#if defined( __parallel )
592          CALL MPI_BARRIER( comm2d, ierr )
593#endif
594       ENDDO
595
596       CALL location_message( 'finished', .TRUE. )
597
598       CALL cpu_log( log_point(22), 'wrd_local', 'stop' )
599
600!
601!--    If required, write particle data in own restart files
602       IF ( particle_advection )  CALL lpm_write_restart_file
603       
604    ENDIF
605!
606!-- Last actions for surface output, for instantaneous and time-averaged data
607    CALL surface_data_output_last_action( 0 )
608    CALL surface_data_output_last_action( 1 )
609
610!
611!-- If required, repeat output of header including the required CPU-time
612    IF ( myid == 0 )  CALL header
613!
614!-- Perform module specific last actions
615    CALL cpu_log( log_point(4), 'last actions', 'start' )
616
617    IF ( myid == 0 .AND. agents_active ) CALL mas_last_actions ! ToDo: move to module_interface
618
619    CALL module_interface_last_actions
620
621    CALL cpu_log( log_point(4), 'last actions', 'stop' )
622
623!
624!-- Close files
625    CALL close_file( 0 )
626    CALL close_dvrp
627
628!
629!-- Write run number to file (used by palmrun to create unified cycle numbers
630!-- for output files
631    IF ( myid == 0  .AND.  runnr > 0 )  THEN
632       OPEN( 90, FILE='RUN_NUMBER', FORM='FORMATTED' )
633       WRITE( 90, '(I4)' )  runnr
634       CLOSE( 90 )
635    ENDIF
636
637!
638!-- Take final CPU-time for CPU-time analysis
639    CALL cpu_log( log_point(1), 'total', 'stop' )
640    CALL cpu_statistics
641
642#if defined( __parallel )
643    CALL MPI_FINALIZE( ierr )
644#endif
645
646 END PROGRAM palm
Note: See TracBrowser for help on using the repository browser.