TOMAS setup guide

From Geos-chem
Jump to navigation Jump to search

This page describes how to acquire the latest source code, data, and libraries required to build and run GEOS-Chem with TOMAS aerosol microphysics on the ace-net glooscap cluster.

Overview

The latest public release of GEOS-Chem with TOMAS does not include many of the recent developments in aerosol science. It also cannot take advantage of parallel computing technologies. However, the 'bleeding edge' code has many recent developments in GEOS-Chem/TOMAS that are not included in the public release, including parallel computing.

Getting Set Up

Compiler

GEOS-Chem works best (only) with the Intel Ifort Fortran compiler - v11.1 There is an instance of the compiler installed on glooscap, which you can load by doing

module load intel/11.1.073

Alternatively, I have installed ifort version 11.1.080. This also gives you access to the iidb debugger. To use this version, add the following to your .bashrc

export LD_LIBRARY_PATH="/home/sfarina/geos-chem-libraries-intel11/lib"
export PATH="/home/sfarina/geos-chem-libraries-intel11/Compiler/11.1/080/bin/intel64:/home/sfarina/opt/bin:$PATH"
export LD_LIBRARY_PATH="/usr/local/gnu/lib64:/usr/local/gnu/lib:/home/sfarina/geos-chem-libraries-intel11/lib:/home/sfarina/geos-chem-libraries-intel11/Compiler/11.1/080/lib/intel64/:/home/sfarina/geos-chem-libraries-intel11/Compiler/11.1/080/idb/lib/intel64"
export INTEL_LICENSE_FILE="/home/sfarina/geos-chem-libraries-intel11/software/intel/Compiler/11.1/080/Licenses"
source /home/sfarina/geos-chem-libraries-intel11/Compiler/11.1/080/bin/ifortvars.sh intel64
export FC="ifort"

Code

The latest stable version of TOMAS will be included with the next public release. Currently, the latest code can be obtained using git

git clone https://bitbucket.org/gcst/geos-chem

Libraries

geos-chem-libraries-intel11 is a bundle of software required to build and run the latest version of GEOS-Chem. Included in this package:

  • NetCDF - Network Common Data Format libraries - required to read and write certain datasets
  • HDF5 - Hierarchical Data Format - required to read and write certain datasets
  • other dependencies - required for netcdf and hdf5

Please follow the directions for Installing libraries for GEOS-Chem wiki before proceeding. You will need to install the netCDF-4.2 libraries.

Environment

After installing the libraries, your .bashrc should include a similar section to the following

ROOT_LIBRARY_DIR="/home/sfarina/geos-chem-libraries-intel11"
GC_BIN=$ROOT_LIBRARY_DIR/bin
GC_INCLUDE=$ROOT_LIBRARY_DIR/include
GC_LIB=$ROOT_LIBRARY_DIR/lib
export GC_BIN
export GC_INCLUDE
export GC_LIB
export LD_LIBRARY_PATH=$GC_LIB:$LD_LIBRARY_PATH

Once the compiler and libraries are installed in ~/geos-chem-libraries-intel11

source ~/.bashrc
ifort --version

If ifort returns

ifort (IFORT) 11.1 20101201

you should be all set to start compiling

Data

To set up the necessary data (meteorology, emissions, land use, etc.) for GEOS-Chem, simply

cd ~
ln -s /home/sfarina/data .

This will allow you to link to my data directory, which is mostly a collection of links to the data at /home/rmartin/group/ctm/ with some changes and additions due to recent GC development and TOMAS specifics. DO NOT copy this directory, as it is many many many gigabytes, and is probably beyond your disk quota on glooscap.

Restart Files

There are restart files for TOMAS at 4x5 resolution at

/net/samqfs/pierce/sfarina/standard_run_directories/restart.TOMASXX

Where XX is the number of bins. These restart files use an "empty" restart file for 2005/06/01 and spin-up times can be calculated accordingly. I will be adding to this directory in the coming week or two. Restart files for 2x2.5 are located at

/net/samqfs/pierce/sfarina/standard_run_directories/2x2.5/restart.ires.TOMAS15

So far, I have only used TOMAS15 at this model resolution.

The North American nested grid is under active development for TOMAS.

Building GEOS-Chem/TOMAS

Bash versions

Geos Chem now requires bash > v3.2 in order to compile properly. There are two ways to ensure you are using the correct version:

  • install your own instance of bash, and point to it in Makefile_header.mk
  • on glooscap, type 'module load bash' before compiling

Compile Flags

Choice of GEOS-Chem model resolution is now done using compile time flags. Full instructions are available here.

Example: To build TOMAS for simulations on a global 4x5 degree grid, using geos5 meteorology, I invoke make as follows for each version:

make GRID=4x5 MET=geos5 tomas12
make GRID=4x5 MET=geos5 tomas15
make GRID=4x5 MET=geos5 tomas
make GRID=4x5 MET=geos5 tomas40

Note: "make tomas" is shorthand for "make TOMAS=yes all"

Note: "make tomas15" os shorthand for "make TOMAS=yes TOMAS15=yes all"

--Salvatore Farina 16:13, 3 March 2014 (EST)

Make

qrsh allows you to use multicore interactive shells to do heavy processing. I invoke a 16 core shell to build geoschem. put this in your .bashrc:

alias pshell16="qrsh -V -cwd -l h_rt=08:00:00 -l h_vmem=2.0G -l h_stack=12.5G -N IA_16 -pe openmp 16 bash"
alias pshell8="qrsh -V -cwd -l h_rt=08:00:00 -l h_vmem=2.0G -l h_stack=12.5G -N IA_8 -pe openmp 8 bash"

Then you can do, for example"

cd YOUR_CODE_DIR/GeosCore
pshell16
make -j16 GRID=4x5 MET=geos5 tomas40

This will build GEOS-Chem with 40 bin TOMAS using 16 processors at a time. As an added bonus, this will not choke up the rest of the users on the system.

The available target names are:

tomas                 <--TOMAS 30
tomas12
tomas15
tomas40

Alternatively, you can use the following to define a tomas version when compiling:

make TOMAS=yes all
make TOMAS40=yes all
etc.

Important!

When changing tomas versions, always always always do

make realclean

Running GEOS-Chem with TOMAS

Run Directories

There are run directories for each of the tomas versions at:

/net/samqfs/pierce/sfarina/standard_run_directories/

Copy the tarballs (named 40.tgz, 30.tgz, etc.) to a standard location. You can then do

tar zxvf YOUR_STANDARD_LOCATION/40.tgz

to extract the appropriate run directory to your current working directory. The folder will be named run.TOMASXX, where XX is 12, 15,30, or 40 depending on the version you would like to run.

Once you have the appropriate version of geostomas compiled and your run directory extracted, copy the executable to your run directory.

input.geos

The input.geos file is where most of the runtime options for geoschem are configured. There are currently no TOMAS specific entries in the input.geos file, save for diagnostic output quantities. Please see the Users' Guide for more information.

Submitting Jobs to the Parallel Queue

In each folder is a file called parallel.sh. Below is a description of some of the parameters:

#!/bin/bash
# $ -S /bin/bash
./etc/profile
#$ -o job_output
#$ -l h_rt=100:00:00                            #wall clock time requested from grid engine. Lower request times will have higher priority in the queue
#$ -l h_vmem=2.0G                               #vmem requested from grid engine. 2.0 is sufficient for all versions at 4x5 and TOMAS15 at 2x2.5 on 16 cores
#$ -l h_stack=12.5G                             #stack memory requested from grid engine
#$ -N RUN_NAM                                   #a name for your run
#$ -pe openmp 16                                #number of cores you are requesting from grid engine
#$ -cwd                                         #inherit properties from your current shell
export OMP_NUM_THREADS=16                       #number of openMP threads
export KMP_STACKSIZE=500000000                  #stacksize memory limit for each thread

ulimit -t unlimited              # cputime
ulimit -f unlimited              # filesize
ulimit -c unlimited              # coredumpsize
ulimit -m unlimited              # memoryuse
ulimit -l unlimited              # memorylocked

cd YOUR_RUN DIRECTORY
./geostomas > log

You'll need to edit it slightly (run name and working directory), then run:

qsub parallel.sh

You can check on the status in the queue with

qstat

You can watch the logfile output of your simulation with

tail -f log

With some minimal editing, you can find some summary information from your runs using the script here

/net/samqfs/pierce/sfarina/testruns/informed/hourstat.sh

A Note about Speed

Choosing the appropriate version of tomas for your needs includes consideration of time and resources. Using 16 processors on glooscap at 4x5 resolution, the model time : real time ratio is roughly as follows:

version   |  speedup
40 bin    -     46
30 bin    -     60
15 bin    -    125
12 bin    -    131

Developing

Writing for GEOS_Chem is pretty straightforward. Please try to follow the style guide as much as possible. Most of TOMAS is contained within tomas_mod.F90, and you should be able to find what you need with a little work and a few invocations of grep. If you can't find what you need, ask.

Version Control

Git! You should definitely use git to track your changes. To use git on glooscap:

module load git

Branching and Commits

Once you have your source code directory, make a separate branch for yourself before making any changes. This will simplify trading and tracking updates/advances/bugfixes.

git checkout -b MY_NEW_BRANCH
vi fictional_example_mod.F90
git status
git add fictional_example_mod.F90
git commit

Patching

If I make some new changes to my branch of code, you will need to do a patch and merge. My current branch in git is called tomasmerge. If I provide you with update.patch, this should do the trick:

git checkout tomasmerge
git apply update.patch
git checkout MY_BRANCH
git merge tomasmerge

Reference

There are many useful resources for git on the web. Here are some I found useful:

Debugging

There are two major ways of debugging: inserting massive amounts of print statements, or using a debugger. Both are useful.

ifort comes with a debugger similar to gdb: iidb.

geos-chem-libraries-intel11/Compiler/11.1/080/bin/intel64/iidb

In order to use it, you must compile geostomas as follows

make realclean
make DEBUG=yes tomas

Apart from the debugger and normal print statements, TOMAS has a very useful builtin called DEBUGPRINT, that prints the values of the TOMAS size bins in a big table.

Post Processing

Now that you've successfully run the model, there are a few more hurdles to inspect your data.

Installing IDL

Copy the IDL / gamap scripts from my home directory.

cp -r ~sfarina/IDL ~

Edit the following as needed, and add it to your .bashrc

IDL_STARTUP="/home/sfarina/IDL/idl_startup/idl_startup.pro"
IDL_DIR="/usr/local/itt/idl/idl80/"
IDL_PATH="$IDL_DIR:/home/sfarina/IDL"
module load idl/8.0

Processing

GEOS-Chem currently outputs all data in the form of a binary punch file (.bpch). These files must be handled using IDL. The process is outlined below:

Copy

Copy the relevant files to your postprocessing directory for a given run

ctm.bpch
diaginfo.dat
tracerinfo.dat
proc_one.pro
averageCNCCN_XX.py             <-- XX is TOMAS version
plotCNCCN.py

Split

Use the script Bpch_Sep_Sal interactively from within the IDL environment to ctm.bpch into separate months For example, to extract august, 2005 from ctm.bpch

idl
> Bpch_Sep_Sal,'ctm.bpch','ctm.08.bpch',Tau0=nymd2tau(20050801) 
> exit

Create netcdf output

Using the IDL script proc_one.pro, we extract information from the monthly .bpch files and save it to the standard netCDF Edit proc_one.pro to use the correct infile/outfiles Execute proc_one from your shell:

idl proc_one.pro

Counting CN and CCN

Run averageCNCCN_XX.py, where XX is the model version For example, to bin and average the August results from TOMAS15:

./averageCNCCN_15.py 08

Plotting the Results

Edit your directory name to be of the format YYY_run.TOMASXX, where YYY is a run number, and XX is the TOMAS version. plotCNCCN.py will automatically detect the model version and customize map names. To plot the surface and zonal average concentrations of CN3, CN10, CN40, and CN80 for august:

./plotCNCCN.py 08

Once you have completed this process, you will have a zonal and surface level map of CN3, CN10, CN40 and CN80 predicted by the model.

NCview

You can also use ncview on the file ctm.nc to view individual species concentrations or nucleation rates.

ncview ctm.nc
ncview ctm_nuc.nc

Other Advice / Issues

  • If you have followed these instructions and geoschem crashes without any output, try (un)commenting the "welcome to geoschem" and the following call flush lines from main.F. This is a hard to track bug related to ongoing development of grid independent geoschem.
  • I use the GNU Bourne Again SHell (bash). I suggest you do the same. The csh is fine, but I have written all of my scripts using bash. Your life will probably be easier if you use bash.
  • If you are trying to run geoschem outside of a qrsh (grid engine) environment (i.e. on the head node), you will need to add ulimit -S -s unlimited to your .bashrc
  • It is a good idea to TAKE NOTES on the details of your simulations.
  • Making a backup of your code and any important files is a good idea. Making two backups in different physical locations is a better idea.
  • if you have any questions or you are running into trouble, please ask either myself, Sajeev, or Jeff for help. I am usually able to respond to emails within a day, and am willing to use gchat or skype if need be.

--Salvatore Farina 17:28, 25 July 2013 (EDT)