Retrieve larger data sets from MARS on ECgate or HPCF

From mn/geo/geoit
Revision as of 17:35, 12 March 2013 by Nik (talk | contribs) (RETRIEVE, LIST, READ, COMPUTE and WRITE CALLS)

Jump to: navigation, search

Back to ECMWF overview[1]

Updated by nik: 12 March 2013

To retrieve large numbers of files from MARS you need to log in to the ECgate server and issue jobs in form of scripts.


The retrieval of data from MARS is done through a submission of a shell-script.

When running a program on the UNIX system use the batch system (not interactive mode). That means you submit the job with explicit commands so that the job is run unattended under Unix.

nohup is a way to submit a job to run unattended on a Unix system, but there exists more sophisticated batch systems for handling jobs.

The batch system currently available on ECgate and HPCF is called LoadLeveler and jobs are submitted with the command llsubmit.

OBS OBS! From June 2013 the batch system on ECgate will change from Loadleveler to SLURM and jobs are submitted with the command sbatch.

This page should be updated to contain the commands for the new batch system when it is up running.


Log in to ECgate.

Create a ksh script on your home directory on ecgate.

(The default shell is either Korn (ksh) or C-shell (csh)).

In the beginning of the script, set the Batch System keywords:

#@ shell        = /usr/bin/ksh                   (Specify the shell)
#@ job_type     = serial                         (Indicates that this is a serial job)
#@ job_name     = request_to_MARS                (Name of job)
#@ initialdir   = /scratch/ms/no/sb9/test/       (Pathname to the initial working directory. OBS: Do not user environmental variables like $USER or $SCRATCH)
#@ output       = $(job_name).$(jobid).out       (*.out file)
#@ error        = $(job_name).$(jobid).err       (Error file)
#@ class        = normal                         (indicates the priority of the job, usually this is "normal")
#@ notification = complete                       (Sends notification on completion)
#@ notify_user = <userId>              (change to your userID, by default your userID )
#@ account_no   = spnoflex                       (FLEXPART account)
#@ queue                                         (indicates the end of the keywords, mandatory)

OBS: Do not user environmental variables like $USER or $SCRATCH in the LL keywords!

Then add one or more requests which might look like this:

 class    = od,                                             ("Operational archive")
 stream   = oper,                                           ("Which model forecast" - here:operational Atmospheric model)
 expver   = 1,                                              ("Experiment version", always use 1 which is the verified version)
 date     = 20130312,                                       ("Specifies the Analysis/Forecast base date")
 time     = 0,                                              ("Specifies the Analysis/Forecast base time"
 step     = 0/to/72/by/6,                                   ("Specifies the Analysis/Forecast time steps to retrieve data for - from the base time")
 type     = an,                                             ("Type of field", Forecast=fc, Analysis=an)
 levtype  = pl,                                             ("Type of level", pressure level=pl, model level=ml, surface=sfc...)
 levelist = 100/150/200/250/300/400/500/700/850/925/1000,   ("Levels for the specified levtype", this can also be specified as "all")
 param    = 129.128/130.128/131.128/132.128/133.128,        ("Meteorological parameters to retrieve", can be specified in various ways, e.g. t, temperature, 130,30.128)
 grid     = 0.5/0.5,                                        ("Post-processing". The data are interpolated to a grid of lat/long increments)
 target   = "ecmwf_data.grib"                               ("Storage", pathname where retrieved data is stored)

You can repeat this retrieve section again and the values are inherited to the next retrieval session so that you only need to specify the ones you want to change!

A summary of MARS keywords can be found here:

The GRIB data in MARS are in Spherical Harmonics for Upper-Air fields, but by specifying grid one can interpolate this data to a lon/lat grid, or by specifying gaussian one can interpolate the data to a regular or reduced Gaussian grid.

To transfer files to zardoz add the following to your ksh script:

ectrans -remote my_ms_association@genericFtp \
-source ecmwf_data.grib \
-mailto \
-onfailure \
-remove \


Instead of the retrieve keyword in the above script you can use other keywords to list, read, compute or write data.

list     Lists the data in the archive (metadata). See amount of data, number of fields, number of tapes. Alternative to the archive catalogue on web-MARS.
read     Reads fields from a local disk/file (already retrieved data). Used to filter/manipulate already retrieved data. Need to specify source.
compute  Performs computation on GRIB fields. E.g. u

As a rule - you should have as few retrieve calls as possible, but rather have each retrieval call collect as much data as possible.

This is to limit the number of calls to the archive considering the architectural structure of the archive. Some data are stored on tape, and this is collected "manually".

Accessing the same tape multiple times in form of many retrieval commands is a large effort.

You can collect a lot of data in one single retrieve call and subsequently change the output file and create multiple files of the data with the read/compute/write commands.

However, to retrieve e.g. ERA-INTERIM data for a full year, it would be most appropriate to split the retrieval calls up in 12 (one for each month) considering

- the architecture of MARS
- volume of data extracted
- restrictions on disk space
- restart availability
- queuing (the larger the request the slower)

By optimizing your request you can jump forward in the queue!


Submit your job as a batch job to LoadLeveler.

To submit your script:

llsubmit myscript


llq -u <UserId> To view where your job is in the queue
llq -l <jobId>  To get a detailed description for a job
llq -s <jobId>  To determine why the job is not running
llcancel <jobId>To cancel your script

See man llq for more options