Quick Start (RCESM Model Workflow)¶
The following quick start guide is for versions of RCESM that have already been ported to the local target machine. RCESM is built on the CIME (Common Infrastructure for Modeling Earth) framework. Please refer to the CIME Porting Documentation if CIME has not yet been ported to the target machine.
If you are new to RCESM or CESM2, please consider reading the CIME Case Control System Part 1: Basic Usage guide first.
This is the procedure for quickly setting up and running a RCESM case.
Download the RCESM source (see Downloading RCESM).
Select a component set, and a resolution for your case. Details of available
component sets and resolutions are available from the query_config tool located
in the my_cresm_sandbox/cime/scripts
directory
cd my_cresm_sandbox/cime/scripts
./query_config --help
See the available RCESM component sets, supported model resolutions and supported machines for a list of RCESM supported component sets, grids and computational platforms.
Note
The CRESM was developed and tested around a fully-coupled Gulf of Mexico (GOM) case. We will use this case as the example throughout this quick start guide. There are options for other grids and configurations available. Please see the available RCESM component sets and supported model resolutions for more information.
Note
Variables presented as $VAR
in this guide typically refer to variables in XML files
in a CESM case. From within a case directory, you can determine the value of such a
variable with ./xmlquery VAR
. In some instances, $VAR
refers to a shell
variable or some other variable; we try to make these exceptions clear.
Create a case¶
The create_newcase command creates a case directory containing the scripts and XML
files to configure a case (see below) for the requested resolution, component set, and
machine. create_newcase has three required arguments: --case
, --compset
and
--res
(invoke create_newcase –help for help).
On machines where a project or account code is needed (including NCAR’s machines), you
must either specify the --project
argument to create_newcase or set the
$PROJECT
variable in your shell environment.
If running on a supported machine, that machine will
normally be recognized automatically and therefore it is not required
to specify the --machine
argument to create_newcase.
Invoke create_newcase as follows:
./create_newcase --case CASENAME --compset COMPSET --res GRID
where:
CASENAME
defines the name of your case (stored in the$CASE
XML variable). This is a very important piece of metadata that will be used in filenames, internal metadata and directory paths. create_newcase will create the case directory with the same name as theCASENAME
. IfCASENAME
is simply a name (not a path), the case directory is created in the directory where you executed create_newcase. IfCASENAME
is a relative or absolute path, the case directory is created there, and the name of the case will be the last component of the path. The full path to the case directory will be stored in the$CASEROOT
XML variable. RCESM case names should follow the CESM naming conventions. See CESM2 Experiment Casenames for details regarding CESM experiment case naming conventions.COMPSET
is the component set.GRID
is the model resolution.run-unsupported
is required for all RCESM compsets as these should all be considered scientifically experimental within the RCESM/CESM code base.
Here is an example on NCAR machine cheyenne with the $USER
shell environment variable
set to your cheyenne login name:
my_cresm_sandbox/cime/scripts/create_newcase --case my_case_dirs/p.r10.pbsgulf2010.tx9k_g3x.my_new_test.001 --compset PBSGULF2010 -res tx9k_g3x --run-unsupported
Setting up the case run script¶
Issuing the case.setup command creates scripts needed to run the model
along with namelist user_nl_xxx
files, where xxx denotes the set of components
for the given case configuration. Before invoking case.setup, modify
the env_mach_pes.xml
file in the case directory using the xmlchange command
as needed for the experiment.
cd to the case directory. Following the example from above:
cd my_case_dirs/p.r10.pbsgulf2010.tx9k_g3x.my_new_test.001
Modify settings in env_mach_pes.xml
(optional). (Note: To edit any of
the env xml files, use the xmlchange command.
invoke xmlchange –help for help.)
Invoke the case.setup command.
./case.setup
Build the executable using the case.build command¶
Modify build settings in env_build.xml
(optional).
Run the build script.
./case.build
Users of the NCAR cheyenne system should consider using qcmd to compile RCESM on a compute node as follows:
qcmd -- ./case.build
The RCESM executable will appear in the directory given by the XML variable $EXEROOT
,
which can be queried using:
./xmlquery EXEROOT
Run the case¶
Modify runtime settings in env_run.xml
(optional).
Run length: By default, the model is set to run for 5 days based on the $STOP_N
and
$STOP_OPTION
variables:
./xmlquery STOP_OPTION,STOP_N
These default settings can be useful in troubleshooting runtime problems before submitting for a longer time. The Gulf of Mexico example from above has mostly been tested for this 5 day period. If you wanted to run it for an entire month, you would change:
./xmlchange STOP_OPTION=nmonths,STOP_N=1
After setting the run lenght, submit the job to the batch queue using the case.submit command.
./case.submit
When the job is complete, most output will NOT be written under the case directory, but
instead under some other directories (on NCAR’s cheyenne machine, these other directories
will be in /glade/scratch/$USER
). Review the following directories and files, whose
locations can be found with xmlquery (note: xmlquery can be run with a list of
comma separated names and no spaces):
./xmlquery RUNDIR,CASE,CASEROOT
$RUNDIR
This directory is set in the
env_run.xml
file. This is the location where RCESM was run. There should be log files there for every component (i.e. of the form cpl.log.yymmdd-hhmmss). Each component writes its own log file. Also see whether any restart or history files were written. To check that a run completed successfully, check the last several lines of the cpl.log file for the string “SUCCESSFUL TERMINATION OF CPL7-cesm”.$CASEROOT
There could be standard out and/or standard error files output from the batch system.
$CASEROOT/CaseDocs
The case namelist files are copied into this directory from the
$RUNDIR
.$CASEROOT/timing
There should be two timing files there that summarize the model performance.
Restart the case¶
The RCESM supports the ability to restart from any point that restart files are written. To
set the frequency that restart files are written, first check the $REST_N
and
$REST_OPTION
variables:
./xmlquery REST_OPTION,REST_N
The default values for these variables are set to the end of the simulation (so they should
be the same as $STOP_N
and $STOP_OPTION
variables initially). To set the
frequency so that restart files are written more often, change the variables to reflect the
values you want. For example, to write daily restart files, set:
./xmlchange REST_OPTION=ndays,REST_N=1
Once your simulation is finished and you want to restart the run, you will need to change the
$CONTINUE_RUN
xml variable to true, and then submit the run again.
./xmlchange CONTINUE_RUN=TRUE
./case.submit
This change on its own will restart the model from the point of the last written restart. This
is because the model uses the rpointer.*
files in the case run directory to determine
the restart date and time. These files are automatically set to point to the most recently written
set of restart files. If you were interested in restarting from an earlier write, you would need to
manually edit the filenames in each of the rpointer.*
files in the case run directory.