TOMAS setup guide: Difference between revisions
(→Code) |
|||
Line 226: | Line 226: | ||
== Post Processing == | == Post Processing == | ||
NB: The following section is out of date. Please contact Jeff Pierce or Peter Adams for TOMAS resources. | |||
=== Installing IDL === | === Installing IDL === |
Revision as of 23:38, 13 August 2018
This page describes how to acquire the latest source code, data, and libraries required to build and run GEOS-Chem with TOMAS aerosol microphysics on the ace-net glooscap cluster.
Overview
The latest public release of GEOS-Chem with TOMAS does not include many of the recent developments in aerosol science. It also cannot take advantage of parallel computing technologies. However, the 'bleeding edge' code has many recent developments in GEOS-Chem/TOMAS that are not included in the public release, including parallel computing.
Getting Set Up
Compiler
GEOS-Chem works best (only) with the Intel Ifort Fortran compiler - v11.1 There is an instance of the compiler installed on glooscap, which you can load by doing
module load intel/11.1.073
Alternatively, I have installed ifort version 11.1.080. This also gives you access to the iidb debugger. To use this version, add the following to your .bashrc
export LD_LIBRARY_PATH="/home/sfarina/geos-chem-libraries-intel11/lib" export PATH="/home/sfarina/geos-chem-libraries-intel11/Compiler/11.1/080/bin/intel64:/home/sfarina/opt/bin:$PATH" export LD_LIBRARY_PATH="/usr/local/gnu/lib64:/usr/local/gnu/lib:/home/sfarina/geos-chem-libraries-intel11/lib:/home/sfarina/geos-chem-libraries-intel11/Compiler/11.1/080/lib/intel64/:/home/sfarina/geos-chem-libraries-intel11/Compiler/11.1/080/idb/lib/intel64" export INTEL_LICENSE_FILE="/home/sfarina/geos-chem-libraries-intel11/software/intel/Compiler/11.1/080/Licenses" source /home/sfarina/geos-chem-libraries-intel11/Compiler/11.1/080/bin/ifortvars.sh intel64 export FC="ifort"
Code
The latest stable version of TOMAS will be included with the next public release. Currently, the latest code can be obtained using git
git clone https://bitbucket.org/gcst/geos-chem
Libraries
geos-chem-libraries-intel11 is a bundle of software required to build and run the latest version of GEOS-Chem. Included in this package:
- NetCDF - Network Common Data Format libraries - required to read and write certain datasets
- HDF5 - Hierarchical Data Format - required to read and write certain datasets
- other dependencies - required for netcdf and hdf5
Please follow the directions for Installing libraries for GEOS-Chem wiki before proceeding. You will need to install the netCDF-4.2 libraries.
Environment
After installing the libraries, your .bashrc should include a similar section to the following
ROOT_LIBRARY_DIR="/home/sfarina/geos-chem-libraries-intel11" GC_BIN=$ROOT_LIBRARY_DIR/bin GC_INCLUDE=$ROOT_LIBRARY_DIR/include GC_LIB=$ROOT_LIBRARY_DIR/lib export GC_BIN export GC_INCLUDE export GC_LIB export LD_LIBRARY_PATH=$GC_LIB:$LD_LIBRARY_PATH
Once the compiler and libraries are installed in ~/geos-chem-libraries-intel11
source ~/.bashrc ifort --version
If ifort returns
ifort (IFORT) 11.1 20101201
you should be all set to start compiling
Data
To set up the necessary data (meteorology, emissions, land use, etc.) for GEOS-Chem, simply
cd ~ ln -s /home/sfarina/data .
This will allow you to link to my data directory, which is mostly a collection of links to the data at /home/rmartin/group/ctm/ with some changes and additions due to recent GC development and TOMAS specifics. DO NOT copy this directory, as it is many many many gigabytes, and is probably beyond your disk quota on glooscap.
Restart Files
There are restart files for TOMAS at 4x5 resolution at
/net/samqfs/pierce/sfarina/standard_run_directories/restart.TOMASXX
Where XX is the number of bins. These restart files use an "empty" restart file for 2005/06/01 and spin-up times can be calculated accordingly. I will be adding to this directory in the coming week or two. Restart files for 2x2.5 are located at
/net/samqfs/pierce/sfarina/standard_run_directories/2x2.5/restart.ires.TOMAS15
So far, I have only used TOMAS15 at this model resolution.
The North American nested grid is under active development for TOMAS.
Building GEOS-Chem/TOMAS
Bash versions
Geos Chem now requires bash > v3.2 in order to compile properly. There are two ways to ensure you are using the correct version:
- install your own instance of bash, and point to it in Makefile_header.mk
- on glooscap, type 'module load bash' before compiling
Compile Flags
Choice of GEOS-Chem model resolution is now done using compile time flags. Full instructions are available here.
Example: To build TOMAS for simulations on a global 4x5 degree grid, using geos5 meteorology, I invoke make as follows for each version:
make GRID=4x5 MET=geos5 tomas12 make GRID=4x5 MET=geos5 tomas15 make GRID=4x5 MET=geos5 tomas make GRID=4x5 MET=geos5 tomas40
Note: "make tomas" is shorthand for "make TOMAS=yes all"
Note: "make tomas15" os shorthand for "make TOMAS=yes TOMAS15=yes all"
--Salvatore Farina 16:13, 3 March 2014 (EST)
Make
qrsh allows you to use multicore interactive shells to do heavy processing. I invoke a 16 core shell to build geoschem. put this in your .bashrc:
alias pshell16="qrsh -V -cwd -l h_rt=08:00:00 -l h_vmem=2.0G -l h_stack=12.5G -N IA_16 -pe openmp 16 bash" alias pshell8="qrsh -V -cwd -l h_rt=08:00:00 -l h_vmem=2.0G -l h_stack=12.5G -N IA_8 -pe openmp 8 bash"
Then you can do, for example"
cd YOUR_CODE_DIR/GeosCore pshell16 make -j16 GRID=4x5 MET=geos5 tomas40
This will build GEOS-Chem with 40 bin TOMAS using 16 processors at a time. As an added bonus, this will not choke up the rest of the users on the system.
The available target names are:
tomas <--TOMAS 30 tomas12 tomas15 tomas40
Alternatively, you can use the following to define a tomas version when compiling:
make TOMAS=yes all make TOMAS40=yes all etc.
Important!
When changing tomas versions, always always always do
make realclean
Running GEOS-Chem with TOMAS
Run Directories
There are run directories for each of the tomas versions at:
/net/samqfs/pierce/sfarina/standard_run_directories/
Copy the tarballs (named 40.tgz, 30.tgz, etc.) to a standard location. You can then do
tar zxvf YOUR_STANDARD_LOCATION/40.tgz
to extract the appropriate run directory to your current working directory. The folder will be named run.TOMASXX, where XX is 12, 15,30, or 40 depending on the version you would like to run.
Once you have the appropriate version of geostomas compiled and your run directory extracted, copy the executable to your run directory.
input.geos
The input.geos file is where most of the runtime options for geoschem are configured. There are currently no TOMAS specific entries in the input.geos file, save for diagnostic output quantities. Please see the Users' Guide for more information.
Submitting Jobs to the Parallel Queue
In each folder is a file called parallel.sh. Below is a description of some of the parameters:
#!/bin/bash # $ -S /bin/bash ./etc/profile #$ -o job_output #$ -l h_rt=100:00:00 #wall clock time requested from grid engine. Lower request times will have higher priority in the queue #$ -l h_vmem=2.0G #vmem requested from grid engine. 2.0 is sufficient for all versions at 4x5 and TOMAS15 at 2x2.5 on 16 cores #$ -l h_stack=12.5G #stack memory requested from grid engine #$ -N RUN_NAM #a name for your run #$ -pe openmp 16 #number of cores you are requesting from grid engine #$ -cwd #inherit properties from your current shell export OMP_NUM_THREADS=16 #number of openMP threads export KMP_STACKSIZE=500000000 #stacksize memory limit for each thread ulimit -t unlimited # cputime ulimit -f unlimited # filesize ulimit -c unlimited # coredumpsize ulimit -m unlimited # memoryuse ulimit -l unlimited # memorylocked cd YOUR_RUN DIRECTORY ./geostomas > log
You'll need to edit it slightly (run name and working directory), then run:
qsub parallel.sh
You can check on the status in the queue with
qstat
You can watch the logfile output of your simulation with
tail -f log
With some minimal editing, you can find some summary information from your runs using the script here
/net/samqfs/pierce/sfarina/testruns/informed/hourstat.sh
A Note about Speed
Choosing the appropriate version of tomas for your needs includes consideration of time and resources. Using 16 processors on glooscap at 4x5 resolution, the model time : real time ratio is roughly as follows:
version | speedup 40 bin - 46 30 bin - 60 15 bin - 125 12 bin - 131
Developing
Writing for GEOS_Chem is pretty straightforward. Please try to follow the style guide as much as possible. Most of TOMAS is contained within tomas_mod.F90, and you should be able to find what you need with a little work and a few invocations of grep. If you can't find what you need, ask.
Version Control
Git! You should definitely use git to track your changes. To use git on glooscap:
module load git
Branching and Commits
Once you have your source code directory, make a separate branch for yourself before making any changes. This will simplify trading and tracking updates/advances/bugfixes.
git checkout -b MY_NEW_BRANCH vi fictional_example_mod.F90 git status git add fictional_example_mod.F90 git commit
Patching
If I make some new changes to my branch of code, you will need to do a patch and merge. My current branch in git is called tomasmerge. If I provide you with update.patch, this should do the trick:
git checkout tomasmerge git apply update.patch git checkout MY_BRANCH git merge tomasmerge
Reference
There are many useful resources for git on the web. Here are some I found useful:
Debugging
There are two major ways of debugging: inserting massive amounts of print statements, or using a debugger. Both are useful.
ifort comes with a debugger similar to gdb: iidb.
geos-chem-libraries-intel11/Compiler/11.1/080/bin/intel64/iidb
In order to use it, you must compile geostomas as follows
make realclean make DEBUG=yes tomas
Apart from the debugger and normal print statements, TOMAS has a very useful builtin called DEBUGPRINT, that prints the values of the TOMAS size bins in a big table.
Post Processing
NB: The following section is out of date. Please contact Jeff Pierce or Peter Adams for TOMAS resources.
Installing IDL
Copy the IDL / gamap scripts from my home directory.
cp -r ~sfarina/IDL ~
Edit the following as needed, and add it to your .bashrc
IDL_STARTUP="/home/sfarina/IDL/idl_startup/idl_startup.pro" IDL_DIR="/usr/local/itt/idl/idl80/" IDL_PATH="$IDL_DIR:/home/sfarina/IDL" module load idl/8.0
Processing
GEOS-Chem currently outputs all data in the form of a binary punch file (.bpch). These files must be handled using IDL. The process is outlined below:
Copy
Copy the relevant files to your postprocessing directory for a given run
ctm.bpch diaginfo.dat tracerinfo.dat proc_one.pro averageCNCCN_XX.py <-- XX is TOMAS version plotCNCCN.py
Split
Use the script Bpch_Sep_Sal interactively from within the IDL environment to ctm.bpch into separate months For example, to extract august, 2005 from ctm.bpch
idl > Bpch_Sep_Sal,'ctm.bpch','ctm.08.bpch',Tau0=nymd2tau(20050801) > exit
Create netcdf output
Using the IDL script proc_one.pro, we extract information from the monthly .bpch files and save it to the standard netCDF Edit proc_one.pro to use the correct infile/outfiles Execute proc_one from your shell:
idl proc_one.pro
Counting CN and CCN
Run averageCNCCN_XX.py, where XX is the model version For example, to bin and average the August results from TOMAS15:
./averageCNCCN_15.py 08
Plotting the Results
Edit your directory name to be of the format YYY_run.TOMASXX, where YYY is a run number, and XX is the TOMAS version. plotCNCCN.py will automatically detect the model version and customize map names. To plot the surface and zonal average concentrations of CN3, CN10, CN40, and CN80 for august:
./plotCNCCN.py 08
Once you have completed this process, you will have a zonal and surface level map of CN3, CN10, CN40 and CN80 predicted by the model.
NCview
You can also use ncview on the file ctm.nc to view individual species concentrations or nucleation rates.
ncview ctm.nc ncview ctm_nuc.nc
Other Advice / Issues
- If you have followed these instructions and geoschem crashes without any output, try (un)commenting the "welcome to geoschem" and the following call flush lines from main.F. This is a hard to track bug related to ongoing development of grid independent geoschem.
- I use the GNU Bourne Again SHell (bash). I suggest you do the same. The csh is fine, but I have written all of my scripts using bash. Your life will probably be easier if you use bash.
- If you are trying to run geoschem outside of a qrsh (grid engine) environment (i.e. on the head node), you will need to add ulimit -S -s unlimited to your .bashrc
- It is a good idea to TAKE NOTES on the details of your simulations.
- Making a backup of your code and any important files is a good idea. Making two backups in different physical locations is a better idea.
- if you have any questions or you are running into trouble, please ask either myself, Sajeev, or Jeff for help. I am usually able to respond to emails within a day, and am willing to use gchat or skype if need be.
--Salvatore Farina 17:28, 25 July 2013 (EDT)