Retrieve larger data sets from MARS on ECgate or HPCF
Back to ECMWF overview
Updated by nik: 12 March 2013
To retrieve large numbers of files from MARS you need to log in to the ECgate server and issue jobs in form of scripts.
The retrieval of data from MARS is done through a submission of a shell-script.
When running a program on the UNIX system use the batch system (not interactive mode). That means you submit the job with explicit commands so that the job is run unattended under Unix.
nohup is a way to submit a job to run unattended on a Unix system, but there exists more sophisticated batch systems for handling jobs.
The batch system currently available on ECgate and HPCF is called LoadLeveler and jobs are submitted with the command llsubmit.
OBS OBS! From June 2013 the batch system on ECgate will change from Loadleveler to SLURM and jobs are submitted with the command sbatch.
This page should be updated to contain the commands for the new batch system when it is up running.
CREATING A SCRIPT
Log in to ECgate.
Create a ksh script on your home directory on ecgate.
(The default shell is either Korn (ksh) or C-shell (csh)).
In the beginning of the script, set the Batch System keywords:
#@ shell = /usr/bin/ksh (Specify the shell) #@ job_type = serial (Indicates that this is a serial job) #@ job_name = request_to_MARS (Name of job) #@ initialdir = /scratch/ms/no/sb9/test/ (Pathname to the initial working directory. OBS: Do not user environmental variables like $USER or $SCRATCH) #@ output = $(job_name).$(jobid).out (*.out file) #@ error = $(job_name).$(jobid).err (Error file) #@ class = normal (indicates the priority of the job, usually this is "normal") #@ notification = complete (Sends notification on completion) #@ notify_user = <userId>@ecmwf.int (change to your userID, by default your userID ) #@ account_no = spnoflex (FLEXPART account) #@ queue (indicates the end of the keywords, mandatory)
OBS: Do not user environmental variables like $USER or $SCRATCH in the LL keywords!
Then add one or more requests which might look like this:
retrieve, class = od, ("Operational archive") stream = oper, ("Which model forecast" - here:operational Atmospheric model) expver = 1, ("Experiment version", always use 1 which is the verified version) date = 20130312, ("Specifies the Analysis/Forecast base date") time = 0, ("Specifies the Analysis/Forecast base time" step = 0/to/72/by/6, ("Specifies the Analysis/Forecast time steps to retrieve data for - from the base time") type = an, ("Type of field", Forecast=fc, Analysis=an) levtype = pl, ("Type of level", pressure level=pl, model level=ml, surface=sfc...) levelist = 100/150/200/250/300/400/500/700/850/925/1000, ("Levels for the specified levtype", this can also be specified as "all") param = 129.128/130.128/131.128/132.128/133.128, ("Meteorological parameters to retrieve", can be specified in various ways, e.g. t, temperature, 130,30.128) grid = 0.5/0.5, ("Post-processing". The data are interpolated to a grid of lat/long increments) target = "ecmwf_data.grib" ("Storage", pathname where retrieved data is stored)
You can repeat this retrieve section again and the values are inherited to the next retrieval section so that for the second section you only need to specify the ones you want to change from the first section.
A summary of MARS keywords can be found here:
The GRIB data in MARS are in Spherical Harmonics for Upper-Air fields, but by specifying grid one can interpolate this data to a lon/lat grid, or by specifying gaussian one can interpolate the data to a regular or reduced Gaussian grid.
To transfer files to zardoz add the following to your ksh script:
ectrans -remote my_ms_association@genericFtp \ -source ecmwf_data.grib \ -mailto firstname.lastname@example.org \ -onfailure \ -remove \ -verbose
RETRIEVE, LIST, READ, COMPUTE and WRITE CALLS
Instead of the retrieve keyword in the above script you can use other keywords to list, read, compute or write data.
list Lists the data in the archive (metadata). See amount of data, number of fields, number of tapes. Alternative to the archive catalogue on web-MARS. read Reads fields from a local disk/file (already retrieved data). Used to filter/manipulate already retrieved data. Need to specify source. compute Performs computation on GRIB fields. Need to specify formula, e.g. formula="sqrt(u*u + v*v)" to calculate windspeed from the u and v velocities. write Writes the computed data to a target GRIB file.
With these keywords it is useful to use fieldset for temporary storage for further processing. A variable is then set which can be referenced in a further request.
At the end of the call to MARS, all fieldsets are released.
As a rule - you should have as few retrieve calls as possible, but rather have each retrieval call collect as much data as possible.
This is to limit the number of calls to the archive considering the architectural structure of the archive. Some data are stored on tape, and this is collected "manually".
Accessing the same tape multiple times in form of many retrieval commands is a large effort.
You can collect a lot of data in one single retrieve call and subsequently change the output file and create multiple files of the data with the read/compute/write commands.
However, to retrieve e.g. ERA-INTERIM data for a full year, it would be most appropriate to split the retrieval calls up in 12 (one for each month) considering
- the architecture of MARS - volume of data extracted - restrictions on disk space - restart availability - queuing (the larger the request the slower)
By optimizing your request you can jump forward in the queue!
SUBMIT YOUR JOB
Submit your job as a batch job to LoadLeveler.
To submit your script:
MONITOR YOUR JOB
llq -u <UserId> To view where your job is in the queue
llq -l <jobId> To get a detailed description for a job
llq -s <jobId> To determine why the job is not running
llcancel <jobId>To cancel your script
See man llq for more options