I am in the process of running a 5-day simulation of MPAS-A on Cheyenne using a 15-3km variable mesh obtained here: MPAS-Atmosphere mesh downloads. The mesh comes with a range of mesh partition files (e.g., x5.6488066.graph.info.part.256, x5.6488066.graph.info.part.1024... etc) for running Atmosphere in parallel.
What is the most efficient method of running MPAS in parallel with high-res meshes like 15-3km? In other words 1) how many MPI tasks should be used, and 2) how should they be split among Cheyenne nodes?
See attached for a sample PBS batch script I used to submit a job. This is modified from /glade/p/mmm/wmr/mpas_tutorial/job_scripts/run_model.pbs. Here, I used 8 MPIprocs on 32 nodes for a total of 256 MPI tasks. MPAS ran error free, but it took 12 hour of walltime for 12 hours of simulation time. I would like to speed this up in a way that is not overwhelming the Cheyenne compute nodes.
Thanks!
What is the most efficient method of running MPAS in parallel with high-res meshes like 15-3km? In other words 1) how many MPI tasks should be used, and 2) how should they be split among Cheyenne nodes?
See attached for a sample PBS batch script I used to submit a job. This is modified from /glade/p/mmm/wmr/mpas_tutorial/job_scripts/run_model.pbs. Here, I used 8 MPIprocs on 32 nodes for a total of 256 MPI tasks. MPAS ran error free, but it took 12 hour of walltime for 12 hours of simulation time. I would like to speed this up in a way that is not overwhelming the Cheyenne compute nodes.
Thanks!