wiki:Doc/Intro/News

Very Important News

See also platform-users mailing list archive (for registred users only) here : http://courriel.ipsl.jussieu.fr/mailman/listinfo/platform-users or directly here : http://courriel.ipsl.jussieu.fr/mailman/private/platform-users/

---

2023

2023/04/14 Information about TGCC maintenance

Hello,

following the maintenance on April 11 and 12 on the Irene machines of the TGCC, we have passed a series of control tests on the computing and post-processing running environment of the IPSL.
These tests are now completed and we give you the "green light" to launch your simulations on the new TGCC environment.

Before you start, please take note of the following points:
 - you will need to update your personal environments, the compiling and execution environments in the running workflow and in some cases the model codes. For that, you will find here 
https://forge.ipsl.fr/igcmg_doc/wiki/Doc/ComputingCenters/TGCC/IreneRedHat8 a documentation listing all the modifications and corrections to be done into your working configurations 
so that they can work on the new redhat8 environment.
- we randomly encounter crashes on some simulations launched on the Irene SKL machine. In this case, the model does not advance any more, the Job leaves in time-limit and the current period 
must be restarted "by hand". Investigations are underway to solve this problem.

Please use the platform-users list to ask your questions or to report any problem.

Good luck with this update !

Arnaud, for Platform group ICMC

2023/02/17 - WARNING - major change on TGCC computing - WARNING -

Bonjour,  (french below)

During the maintenance of April 11 and 12, 2023 all the Irene machines of the TGCC (skylake, rome, xlarge,...) 
will change OS (it will go from redhat7 to redhat8). The computing environment currently used by modipsl 
will no longer be available on the new OS. This implies that the executables of the codes compiled with 
the current environment will no longer be usable and that the results of the simulations will not be strictly 
identical to what we currently obtain. The Platform group is currently working on porting the codes and 
validating them on this new environment in order to be ready for the switchover on April 13.

In the meantime, we strongly encourage you to think about the timing of your future simulations in relation 
to this date, and to contact the technical manager of the configuration you are using if you have any questions.

-------------------------------------------------

Lors de la maintenance des 11 et 12 avril 2023 toutes les machines Irene du TGCC (skylake, rome, xlarge, ...) 
vont changer d'OS (cela passera de redhat7 à redhat8). L'environnement de calcul que nous utilisons actuellement 
via modipsl ne sera plus proposé dans ce nouveau système d'exploitation. Cela implique que les exécutables des 
codes compilés avec l'environnement actuel ne pourront plus être utilisés et que les résultats des simulations 
ne seront pas strictement identiques à ce que l'on obtient actuellement. Le groupe Plateforme travaille actuellement 
au portage des codes et à leur validation sur ce nouvel environnement afin d'être prêt pour la bascule le 13 avril.

En attendant nous vous encourageons fortement à penser le calendrier de vos futures simulations en fonction de cette 
date, et à vous rapprocher du responsable technique de la configuration que vous utilisez si vous avez des questions.

-------------------------------------------------------------------

2023/01/18 Information about Relief.nc and landiceref.nc and ce0l: correction on path

Dear all,

This is an information if you run the program ce0l (create_etat0_limit) from LMDZ, the program used to create start.nc/startphy.nc and limit.nc files for LMDZ.
 
1) Currently in the configurations v6.2_work, v6.3_work or v6.4_work, this program doesn't work. The experiences CREATE_ will fail if you uses the latest version of these configurations. 
It'll be resolved as soon as possible. Use an older version for ce0l for the moment.

2) The input files Relief.nc and landiceref.nc stored in the shared repository, where only links to newer files. These links have now been removed because it made confusion as they before 
pointed to older versions. Instead you should put in lmdz.card (if you use libIGCM) the specific file you need. For example, to use the new files (recommanded) :

        (${R_IN}/ATM/Relief_GMTED2010_15n015_00625deg_plus_GA.nc,  Relief.nc), \
        (${R_IN}/ATM/landiceref_GA_HR.new.nc,                      landiceref.nc) \


To use the original files, necessary if you need to be coherent with the coupled model:

        (${R_IN}/ATM/INPUT_CE0L/Relief_orig.nc,           Relief.nc), \
        (${R_IN}/ATM/INPUT_CE0L/landiceref_orig.nc,   landiceref.nc) \

If you use another run script, you should just remember to rename the files before starting ce0l.

2022

2022/10/24 gestion du storedir sur les machines du TGCC / management of the storedir on TGCC computers

(english below)

Bonjour,

Le TGCC a mis à jour les deux commandes ccc_quota et ccc_tree afin
qu'elles puissent aider à surveiller l'usage que l'on fait du store et à
respecter les critères définis par le TGCC.

Pour rappel le TGCC a un système de surveillance qui enclenche des
alertes lorsque sur le storedir d'un compte il y a : plus de 500
fichiers ET  { 80% des fichiers ont une taille inférieure à 500Mo OU la
taille moyenne des fichiers est inférieure à 1Go }

La commande ccc_quota vous permet de connaître, entre autre, vos
pourcentages de fichiers dont la taille est inférieure à 500Mo (très
petits fichiers), idem pour ceux dont la taille est inférieure à 1Go
(petits fichiers), et vous donne la taille moyenne de vos fichiers.

La commande ccc_tree vous indiquera répertoire par répertoire la
proportion de très petits fichiers, petits fichiers, la taille moyenne
des fichiers, mais également le nombre d'inodes qui ne sont pas des
fichiers. Tout cela correspond à une note sur 20. Si la note est
supérieure à 10/20 le répertoire apparaît en vert, sinon il apparaît en
rouge. Idéalement vous devriez avoir un maximum de répertoires verts.

Pour rappel au delà des conditions de déclenchement d'alerte, il est
fortement recommandé de ne stocker que des gros fichiers ur le storedir
et de stocker les petits fichiers sur le workdir.

Anne

--------------------------------------------------------------------------------------

Hello,

The TGCC has updated the two commands ccc_quota and ccc_tree so that
they can help to monitor the use of the store and to respect the
criteria defined by the TGCC.

As a reminder the TGCC has a monitoring system that triggers alerts when
on the storedir of an account there are: more than 500 files AND { 80%
of the files are smaller than 500MB OR the average file size is smaller
than 1GB }

The ccc_quota command allows you to know, among other things, the
percentage of files whose size is less than 500MB (very small files), as
well as those whose size is less than 1GB (small files), and gives you
the average size of your files.

The ccc_tree command will tell you directory by directory the proportion
of very small files, small files, the average size of the files, but
also the number of inodes that are not files. All this corresponds to a
score out of 20. If the score is higher than 10/20 the directory appears
in green, otherwise it appears in red. Ideally you should have a maximum
of green directories.

As a reminder, beyond the conditions for triggering an alert, it is
strongly recommended to store only large files on the storedir and to
store small files on the workdir.

Anne


2022/09/23 ORCHIDEE at Spirit and Spiritx on IPSL ESPRI MESO center

Dear all,

ORCHIDEE offline configurations using modipsl and libIGCM are now updated to compile and run at the new Spirit and Spiritx cluster at IPSL ESPRI MESO center. No specific environment needs to be installed before compiling.

Note that:
- The file system is the same between ciclad and spirit and between climserv and spritx. But you cannot run an executable compiled at ciclad/climserv on the new spirit and spritx. You need to make a new installation and new compilation.
- Following configurations have been update : ORCHIDEE trunk, ORCHIDEE_4_1, ORCHIDEE_3 and ORCHIDEE_2_2
- For these configurations, there is no need to install or load specific modules before compilation. All is included and automatically loaded during compilation and running.
- If you use another version or personal version, you need to update the file ORCHIDEE_OL/compile_orchidee_ol.sh and add ORCHIDEE_OL/ARCH/arch-ifort-MESOIPSL.env, ORCHIDEE/arch/arch-ifort_MESOIPSL.path and arch-ifort_MESOIPSL.fcm.
- Using libIGCM at Spirit(x), creating, launching job and post-processing TS and SE have been done. Pack tools and monitoring have not been set up at Spirit(x).
- Compilation with old method using Makefiles based on AA_make has not been adapted to spirit/spiritx
- Launching is done using slurm at spirit and spiritx. For you that work on Jean-Zay/IDIRS, this is the same system. Launch using sbatch Job_, use squeue to see the queue status and scancel jobid to kill a job in the queue.
- If problems, use orchidee-help@listes.ipsl.fr
- LMDZOR configurations to come hopefully soon (let me know if you're waiting for it)

Kind regards,
Josefine

2022/09/09 Arrêt du superviseur Hermes

Bonjour,

Nous vous informons que le service de supervision Hermes sera arrêté
définitivement d'ici quelques jours. Pour rappel, le superviseur Hermes
avait été mis en service en 2016 afin de pouvoir suivre les simulations
utilisant la chaîne de calcul et de post-traitement de l'IPSL sur les
différents centres de calcul. Il a été très utilisé durant la phase de
production de l'exercice CMIP6 et n'est plus fonctionnel depuis début 2021.

Le groupe PlateForme

--------------------------------------------------------------------

Hello,

We would like to inform you that the Hermes supervisor service will be
permanently stopped within a few days. As a reminder, the Hermes
supervisor had been put in service in 2016 in order to be able to follow
the simulations using the IPSL computation and post-processing chain on
the different computing centers. It was used extensively during the
production phase of the CMIP6 exercise and is no longer functional since
early 2021.

The PlateForme group


2022/01/03 configuration IPSLCM5A2.2

Bonjour,

(English below)


The IPSLCM5A2.2 configuration is now available when downloading and installing the IPSLCM model with modipsl (./model)

The changes and bug fixes are mostly directed at pelo simulations.

For people that use this configuration in Prindustrial or actual simulation the change from IPSLCM5A2.1 to IPSLCM5A2.2 should be transparent.
It is now possible to use MOSAIX interpolation weights instead of MOSAIC.
If using MOSAIX you should set cpl_old_calving = n in run.def.

For paleo simulations:
- there are now paleo experiments available for IPSLCM and LMDZOR when copying config.card to create your experiment directory.
- there are also namelist file in PARAM named filesomething_paleo that are to be used for paleo simulations (driver files have been modified and .card file may need editing)

A bug in orchidee that forced to introduce an ice point at the south pole has been corrected.

The IPSLCM5A documentation will be updated to account for these modifications

If you have any pronlem when trying to use IPSLCM5A2.2 please let me know so that I can correct what is wrong if needed.

Best !
Sébastien Nguyen (Engineer in charge of IPSLCM5A configuration within IPSL)

2021

2021/12/15 [INFORMATION] New THREDDS web-service


We deployed a new THREDDS server into production: https://thredds-su.ipsl.fr/ It replace https://vesg.ipsl.upmc.fr/ to provide access to WORK THREDDS spaces from TGCC,
IDRIS and CICLAD. This implies new URLs to access data on your THREDDS spaces. 
For the next 3 months, we won't modify the configuration of vesg.ipsl.upmc.fr to avoid any disruption of service. 

We strongly recommend that you do NOT wait to adapt your scripts and tools.


2021/07/09 Clean of old configurations in mod.def

We cleaned all old configurations in file mod.def that were not supported anymore. 
 
Removed configurations are : 
- IPSLCM6.0.0 => IPSLCM6.0.15 
- IPSLCM6.1.0 => IPSLCM6.1.9 
- IPSLCM5A and IPSLCM5B 
- LMDZOR_v6.1.2 => LMDZOR_v6.1.9 
- LMDZ5A2.1_ISO 
- LMDZORINCAREPR_v6 
- ORCHIDEE_trunk_TP


If you want to checkout a mod.def with all informations on old versions, you can use the 3 following command lines: 
svn co -r 5879 http://forge.ipsl.fr/igcmg/svn/modipsl/trunk/util/ --depth empty
cd util
svn up mod.def

or you can extract all modipsl architecture with the version of mod.def containing old revisions:
svn co -r 5879 http://forge.ipsl.fr/igcmg/svn/modipsl/trunk modipsl 

2020

2020/05/20 IPSL-CMC workflow adapted to Irene-amd/TGCC

The IPSL-CMC running environment for calculation and post-treatment has now been adapted for use at Irene-amd the new supercomputer at TGCC. Current list of maintained configurations can now be used at Jean-Zay.

The main information related to the use of Irene-amd can be found in the documentation https://forge.ipsl.fr/igcmg_doc/wiki/Doc/ComputingCenters/TGCC/IreneAmd. You have first to set up working environment as indicated in the documentation. Note that front-end machines are specific to irene-skl and irene-amd and the connexion method are different. Beware to separate model sources compiled on irene-skl and irene-amd because executables are note compatibles.

A technical control quality have been performed on the IPSLCM6.1.11-LR and IPSLCM6.2_work configurations. A scientific validation has been done on the coupled experiences piControl during 50 years and scenario with the configuration IPSLCM6.1.11. See information about these tests here: https://forge.ipsl.fr/igcmg/wiki/Calculateurs/IreneAmd#Validation.
We encourage you strongly to also validate personally the configuration you use. 

The way to use this new supercomputer is very close to Irene-skl machine concerning filesystems, use of modules, compilers as well as method to submit computing jobs. Nevertheless, we draw your attention to the architecture of AMD Rome processor (128 cores per node, 2 Gb per core) that is different from Irene-skl : computing performances are better on Irene-skl, by comparing core to core. See some performances tests here : https://forge.ipsl.fr/igcmg/wiki/Calculateurs/IreneAmd#Performances. It is possible to improve computing performances by increasing the number of cores used per process as well as dedicating nodes to perform IOs.The users guide of theses functionailties is available here : https://forge.ipsl.fr/igcmg_doc/wiki/Doc/ComputingCenters/TGCC/IreneAmd#Useofspecificoptionstoincreasecomputingperformances. Beware the use of these functionalities should increase the hour-consumption of yuo run, that's why their use must be validated by the person in charge of your configuration.

Note for CMIP6 simulations to be run on this machine : you have to modify the config.card as indicated here : https://forge.ipsl.fr/igcmg/wiki/IPSLCM6/IPSL-CM6A-LR#IPSL-CM6A-LR.

If you have any questions, don't hesitate to send them to the list "platform-users@listes.ipsl.fr" so that more people can learn from the answers. You can also answer questions from your colleges at this list. 

2020/04/28 server svn for LMDZ on TGCC

Dear all,

I'm not sure that this information was already give on the plateform_users list.
The svn LMDZ server was modify at the beginning of april. The new one doesn't work on irene, so LMDZ team done a copy of the old one, by consequence if you want to extract a configuration with LMDZ in it on IRENE or IRENE AMD you need to make some modification on file modipsl/util/mod.def before :

Line

#-S- 11 svn http://svn.lmd.jussieu.fr/LMDZ


will become

#-S- 11 svn http://oldsvn.lmd.jussieu.fr/LMDZ


Have a good day
Anne


2019

you can read archive here

2016

you can read archive here

2015

you can read archive here

Last modified 6 weeks ago Last modified on 03/13/24 16:34:17