Ignore:
Timestamp:
2019-07-22T10:32:59+02:00 (17 months ago)
Author:
smasson
Message:

dev_r10984_HPC-13 : improve error handling, see #2307 and #2285

Location:
NEMO/branches/2019/dev_r10984_HPC-13_IRRMANN_BDY_optimization/doc/namelists
Files:
3 edited

Legend:

Unmodified
Added
Removed
  • NEMO/branches/2019/dev_r10984_HPC-13_IRRMANN_BDY_optimization/doc/namelists/nambdy

    r11042 r11317  
    3333   ln_vol        = .false.    !  total volume correction (see nn_volctl parameter) 
    3434   nn_volctl     =  1         !  = 0, the total water flux across open boundaries is zero 
    35    nb_jpk_bdy    = -1         ! number of levels in the bdy data (set < 0 if consistent with planned run) 
    3635/ 
  • NEMO/branches/2019/dev_r10984_HPC-13_IRRMANN_BDY_optimization/doc/namelists/nambdy_dta

    r10075 r11317  
    22&nambdy_dta    !  open boundaries - external data                       (see nam_bdy) 
    33!----------------------------------------------------------------------- 
    4    ln_full_vel = .false.      !  ??? 
    5  
     4   ln_zinterp  = .false.      !  T if a vertical interpolation is required. Variables gdep[tuv] and e3[tuv] must exist in the file 
     5   !                          !  automatically defined to T if the number of vertical levels in bdy dta /= jpk 
     6   ln_full_vel = .false.      !  T if [uv]3d are "full" velocities and not only its baroclinic components 
     7   !                          !  in this case, baroclinic and barotropic velocities will be recomputed -> [uv]2d not needed 
     8   ! 
    69   cn_dir      = 'bdydta/'    !  root directory for the BDY data location 
    710   !___________!_________________________!___________________!___________!_____________!________!___________!__________________!__________!_______________! 
  • NEMO/branches/2019/dev_r10984_HPC-13_IRRMANN_BDY_optimization/doc/namelists/nammpp

    r10445 r11317  
    22&nammpp        !   Massively Parallel Processing                        ("key_mpp_mpi") 
    33!----------------------------------------------------------------------- 
    4    cn_mpi_send =  'I'      !  mpi send/recieve type   ='S', 'B', or 'I' for standard send, 
    5    !                       !  buffer blocking send or immediate non-blocking sends, resp. 
    6    nn_buffer   =   0       !  size in bytes of exported buffer ('B' case), 0 no exportation 
    74   ln_nnogather =  .true.  !  activate code to avoid mpi_allgather use at the northfold 
    85   jpni        =   0       !  jpni   number of processors following i (set automatically if < 1) 
Note: See TracChangeset for help on using the changeset viewer.