Home About Software Documentation Support Outreach Ecosystem Dev Awards Team & Sponsors

User's Guide 5.6

Introduction

How GridWay operates

GridWay enables you to treat your jobs as if they were Unix processes. Each job is given a numerical identifier, analogous to the PID of a process. This value is called the Job identifier, JID for short. If the job belongs to an array job, it will also have an array identifier, AID for short. A job index within an array is called the task identifier, TID for short.

Jobs are submitted using the gwsubmit command. A job is described by its template file. Here you can specify the job's executable file, its command line arguments, input/output files, standard stream redirection as well as other aspects.

Jobs can be monitored using the gwps command. You can control your jobs at runtime using the gwkill command. You can synchronize your jobs using the gwwait command. You can find out what resources your job has used with the gwhistory command.

System monitoring commands allow you to gather information of the GridWay system and the Grids you are using. These commands are: gwuser to show information about the users using GridWay; gwhost to monitor the available hosts in the testbed; and gwacct to print usage (accounting) information per user or host.

Note
Every command has a -h option which shows its usage and available options.

Job life-cycle in GridWay

A job can be in one of the following dispatch states (DM state):

  • Pending (pend): The job is waiting for a resource to run on. The job reaches this state when it is initially submitted by the user or when it is restarted after a failure, stop or self-migration.
  • Hold (hold): The owner (or GridWay administrator) has held the job. It will not be scheduled until it receives a release signal.
  • Prolog (prol): The job is preparing the remote system, by creating the execution directory in the remote host and transferring the input and restart (in case of migration) files to it.
  • Pre-wrapper (prew): The job is making some advanced preparation tasks in the remote resource, like getting some data from a service, obtaining software licenses, etc.
  • Wrapper (wrap): The job is executing the Wrapper, which in turns executes the actual application. It also starts a self-monitoring program if specified. This monitor, watches the raw performance (CPU usage) obtained by the application.
  • Epilog (epil): The job is finalizing. In this phase it transfers the output and restart (in case of failure, stop or self-migration) files and cleaning up the remote system directory.
  • Migrate (migr): The job is migrating from one resource to another, by canceling the execution of Wrapper and performing finalization tasks in the old resource (like in Epilog state) and preparation tasks in the new resource (like in Prolog state).
  • Stopped (stop): The job is stopped. If restart files have been defined in the Job Template, they are transferred back to the client, and will be used when the job is resumed.
  • Failed (fail): The job failed.
  • Done (done): The job is done and the user can check the exit status.

Figure 1. Simplified state machine of the GridWay Metascheduler.

Simplified state machine of the GridWay Metascheduler.

When a job is in Wrapper dispatch state, it can be in one of the following execution states (EM state), which are a subset of the available Globus GRAM states:

  • Pending (pend): The job has been successfully submitted to the local DRM system and it is waiting for the local DRM system to execute it.
  • Suspended (susp): The job has been suspended by the local DRM system.
  • Active (actv): The job is being executed by the local DRM system
  • Failed (fail): The job failed.
  • Done (done): The job is done.

Finally, The following flags are associated with a job (RWS flags):

  • Restarted (R): Number of times the job was restarted or migrated.
  • Waiting (W): Number of clients waiting for this job to end.
  • Rescheduled (S): 1 if this job is waiting to be rescheduled, 0 otherwise.

A grid-aware application model

In order to obtain a reasonable degree of both application performance and fault tolerance, a job must be able to adapt itself according to the availability of the resources and the current performance provided by them. Therefore, the classical application model must be extended to achieve such functionality.

The GridWay system assumes the following application model:

  • Executable: The executable must be compiled for the remote host architecture. GridWay provides a straightforward method to select the appropriate executable for each host. The variable GW_ARCH, as provided by the Information MAD, can be used to define the executable in the Job Template (for example, EXECUTABLE=sim_code.${GW_ARCH})
  • Input files: These files are staged to the remote host. GridWay provides a flexible way to specify input files and supports Parameter Sweep like definitions. Please note that these files may be also architecture dependent.
  • Output files: These files are generated on the remote host and transferred back to the client once the job has finished.
  • Standard streams: The standard input (STDIN) file is transferred to the remote system previous to job execution. Standard output (STDOUT) and standard error (STDERR) streams are also available at the client once the job has finished. These files could be extremely useful for debugging.
  • Restart files: Restart files are highly advisable if dynamic scheduling is performed. User-level checkpointing managed by the programmer must be implemented because system-level checkpointing is not possible among heterogeneous resources.

Migration is commonly implemented by restarting the job on the new candidate host. Therefore, the job should generate restart files at regular intervals in order to restart execution from a given point. However, for some application domains the cost of generating and transferring restart files could be greater than the saving in compute time due to checkpointing. Hence, if the checkpointing files are not provided the job is restarted from the beginning. In order not to reduce the number of candidate hosts where a job can migrate, the restart files should be architecture independent.

User environment configuration

Important
You should include the following environment variables in your shell configuration file. (example $HOME/.bashrc)

In order to set the user environment, follow these steps:

  • 1. Set up Globus user environment:
    $ source $GLOBUS_LOCATION/etc/globus-user-env.sh
    or
    $ . $GLOBUS_LOCATION/etc/globus-user-env.csh
    depending on the shell you are using.
  • 2. Set up the GridWay user environment:
    $ export GW_LOCATION=<path_to_GridWay_installation>
    $ export PATH=$PATH:$GW_LOCATION/bin 
    or
    $ setenv GW_LOCATION <path_to_GW_location>
    $ setenv PATH $PATH:$GW_LOCATION/bin
    depending on the shell you are using
  • 3. Optionally, you can set up your environment to use the GridWay DRMAA library:
    $ export LD_LIBRARY_PATH=$LD_LIBRARY_PATH:$GW_LOCATION/lib
    or:
    $ setenv LD_LIBRARY_PATH $LD_LIBRARY_PATH:$GW_LOCATION/lib
  • 4. If GridWay has been compiled with accounting support, you may need to set up the DB library. For example, if DB library has been installed in /usr/local/BerkeleyDB.4.4:
    $ export LD_LIBRARY_PATH=$LD_LIBRARY_PATH:/usr/local/BerkeleyDB.4.4/lib
Note
This step is only needed if your environment has not been configured, ask your administrator.
  • 5. DRMAA extensions for all the languages use the dynamic drmaa libraries provided by GridWay. To use this libraries it is needed to tell the operating system where to look for them. Here are described the steps needed to do this in Linux and MacOS X.
    • 1. In linux we have two ways to do this, one is using environment variables and the other one is modifying systemwide library path configuration. You only need to use one of this methods. If you do not have root access to the machine you are using or you do not want to setup it for every user in your system you have to use the environment variable method.
    • 1.1 The environment variable you have to set so the extensions find the required DRMAA library is LD_LIBRARY_PATH with a line similar to:
      $ export LD_LIBRARY_PATH=$GW_LOCATION/lib
      If you want to setup this systemwide you can put this line alongside GW_LOCATION setup into /etc/profile. If you do not have root access or you want to do it per user the best place to do it is in the user's .bashrc.You can also do this steps in the console before launching your scripts as it will have the same effect.
    • Systems that use GNU/libc (GNU/Linux is one of them) do have a systemwide configuration file with the paths where to look for dynamic libraries. You have to add this line to /etc/ld.so.conf:
      <path_to_gridway_installation>/lib

      After doing this you have to rebuild the library cache issuing this command:

      # ldconfig

  • In MacOS X you have to use the environment variable method described for Linux but this time the name of the variable is DYLD_LIBRARY_PATH.

Job description overview

Job Templates allow you to configure your job requirements, in terms of needed files, generated files, requirements and ranks of execution hosts, as well as other options.

Syntax:

  <VARIABLE> = ["]<VALUE>["]
  # <Comments>
Important
Default values for EVERY Job Template are read from $GW_LOCATION/etc/job_template.default.

Job Template options

Table 1. Job Template options.

General
NAME Name of the job (filename of the Job Template by default).
Execution
EXECUTABLE The executable file. Example: EXECUTABLE = bin.${ARCH}
ARGUMENTS Arguments to the above executable. Example: ARGUMENTS = “${TASK_ID}”
ENVIRONMENT User defined, comma-separated, environment variables. Example: ENVIRONMENT = SCRATCH_DIR /tmp, LD_LIBRARY_PATH=/usr/local/lib
TYPE Type of job. Possible values are single (default), multiple and mpi, with similar behaviour to that of GRAM jobs.
NP Number of processors in MPI jobs. For multiple and single jobs it defines the count parameter in the RSL.
I/O files
INPUT_FILES A comma-separated pair of local remote filenames. If the remote filename is missing, the local filename will be preserved in the execution host. Example: INPUT_FILES = param.${TASK_ID} param, inputfile
OUTPUT_FILES A comma-separated pair of remote filename local filename. If the local filename is missing, the remote filename will be preserved in the client host. Example: OUTPUT_FILES = outputfile, binary binary.${ARCH}.${TASK_ID}
Standard streams
STDIN_FILE Standard input file. Example: STDIN_FILE = /dev/null
STDOUT_FILE Standard output file. Example: STDOUT_FILE = stdout_file.${JOB_ID}
STDERR_FILE Standard error file. Example: STDERR_FILE = stderr_file.${JOB_ID}
Checkpointing
RESTART_FILES Checkpoint Files. These files are managed by the programmer and should be architecture independent (NO URLS HERE, you can use a checkpoint server using CHECKPOINT_URL). Example: RESTART_FILES = checkpoint
CHECKPOINT_INTERVAL How often (seconds) restart files are transferred from the execution host to the checkpointing server
CHECKPOINT_URL GridFTP URL to store/access checkpoint files (Default is log directory in localhost). Example: CHECKPOINT_URL = gsiftp://hydrus.ucm.es/var/checkpoints/
Resource selection (See Section ''Resource selection expressions'' for more information)
REQUIREMENTS A Boolean expression evaluated for each available host, if the evaluation returns true the host will be considered to submit the job. Example: REQUIREMENTS = ARCH = “i686” & CPU_MHZ > 1000
RANK A numerical expression evaluated for each candidate host (those for which the requirement expression is true). Those candidates with higher ranks are used first to execute your jobs. Example: RANK = (CPU_MHZ * 2) + FREE_MEM_MB
Scheduling
RESCHEDULING_INTERVAL How often GridWay searches the Grid for better resources to run this job. (0 = never)
RESCHEDULING_THRESHOLD If a better resource is found and the job has been running less than this threshold (in (seconds), it will migrate to the new host
DEADLINE Deadline (format [[D:]H:]M) to start the job (0 = none)
Performance
SUSPENSION_TIMEOUT Maximum suspension time (seconds) in the local job management system. If exceeded the job is migrated to another host. (0 = never)
CPULOAD_THRESHOLD If the CPU assigned to your job is less than this given percentage, the job will be migrated
MONITOR Optional program to monitor job performance
Fault tolerance
RESCHEDULE_ON_FAILURE Behavior in case of failure. Possible values are yes or no
NUMBER_OF_RETRIES Number of retries in case of failure. GridWay follows a linear backoff strategy for re-trying file transfers and job submissions.
Advanced job execution
WRAPPER Script for wrapper. stdout and stderr streams of this program can be found in directory $GW_LOCATION/var/$GW_JOB_ID as files stdout.wrapper.$GW_RESTARTED and stderr.wrapper.$GW_RESTARTED
PRE_WRAPPER Optional program that is executed before the execution of the job, to perform an additional remote setup (e.g. access a web service). This job is ALWAYS submitted to the FORK job-manager. stdout and stderr streams of this program can be found in directory $GW_LOCATION/var/$GW_JOB_ID as files stdout.pre_wrapper.$GW_RESTARTED and stderr.pre_wrapper.$GW_RESTARTED
PRE_WRAPPER_ARGUMENTS Arguments to the pre-wrapper program

File definition in Job Templates

Input and output files are in general specified in a comma-separated, source/destination pair.

  SRC1 DST1, SRC2 DST2,...

We next describe the available alternatives and protocols that you can use to choose the best staging strategy for your applications.

Defining input files

Input files (if staged to the remote host) are always placed in the remote experiment directory. However you can specify the name used for the file in the remote directory with the destination name (DST above). This feature is very useful when your executable always expect a fixed input filename, and you want to process different input files (as is common in parametric computations).

Important
The destination names for input files MUST be a single name, do not use absolute paths or URLs.

You can specify the input files using:

  • Absolute path: In this case no staging is performed. File is assumed to be in that location in the remote host.
    Example:
    EXECUTABLE  = /bin/ls      #Will use remote ls!
  • GridFTP URL: The file will be downloaded from the given GridFTP url. If no destination is given the filename in the URL will be used in the remote host.Example, will copy file input_exp1 from /tmp in machine to the remote host with name input.
    INPUT_FILES  = gsiftp://machine/tmp/input_exp1 input
  • File URL: The file will copied from an absolute path in the local host. If no destination is given the filename in the URL will be used in the remote host.Example:
    INPUT_FILES  = file:///etc/passwd #Will copy local /etc/passwd

    file to remote dir, with name passwd

  • Name: Use simple names to stage files in your local experiment directory (directory where the Job Template file is placed). If no destination is given the filename will be preserved.Example:
    INPUT_FILES  = test_case.bin
Note
The executable is treated as an input file, so the same remarks are applicable for the “EXECUTABLE” Job Template parameter.

Defining output files

Output files are always copied FROM the remote experiment directory. However you can specify the destination of the output files of your applications.

Important
The source names for output files MUST be a single name, do not use absolute paths or URLs.

You can specify the destination of output files using:

  • Absolute path: The remote file name will be copied to the absolute path in the local host. Note that you can also use ''file:///'' protocol.
    Example, to copy the output file output.bin in the /tmp directory of the local host with name outfile:
    OUTPUT_FILES  = output.bin /tmp/outfile
  • GridFTP URL: The file will be copied to the given GridFTP url.Example, you can also use variable substitution in URLs, see next section.
    OUTPUT_FILES  = out gsiftp://storage_servere/~/output.${TASK_ID}
  • Name: Use simple names to stage files to your local experiment directory (directory where the Job Template file is placed). If no destination is given the filename will be preserved.Example:
    OUTPUT_FILES  = test_case.bin

    Tip: You can organize your output files in directories in the experiment directory (They MUST exist!), using a relative path

    OUTPUT_FILES = outfile Out/file.${JOB_ID} #Directory Out must exist in the experiment directory

Defining standard streams

Standard streams includes the standard input for your executable (“STDIN_FILE”) and its standard output and error (“STDOUT_FILE” and “STDERR_FILE”).

The STDIN_FILE can be defined using any of the methods described above for the input files. However you can not specified a destination name for the standard input stream, as is internally handled by the system. Note also that only one standard input file can be specified.

Example:

  STDIN_FILE = In/input.${JOB_ID} #Will use input from directory In

The STDOUT_FILE and STDERR_FILE parameters can be defined using any of the methods described above for the output files. However you can not specified a source name (only destination). Note also that only one standard output and error files can be specified.

Example:

  STDOUT_FILE = Out/ofile #Will place stdout in Out directory with name ofile

Defining restart files

Restart files are periodically copied to the job directory ($GW_LOCATION/var/$JOB_ID/). Restart files are only specified with its name. Note also that you can defined a checkpointing server with the “CHECKPOINT_URL Job Template parameter.

Example:

  RESTART_FILES = tmp_file

Variable substitution

You can use variables in the value string of each option, with the format:

  ${GW_VARIABLE}

These variables are substituted at run time with its corresponding value. For example:

  STDOUT_FILE = stdout.${JOB_ID}

will store the standard output of job 23 in the file stdout.23

The following table lists the variables available to define job options, along with their description.

Table 2. Substitution variables.

Variable Description
${JOB_ID} The job identifier
${ARRAY_ID} The job array identifier (-1 if the job does not belong to any)
${TASK_ID} The task identifier within the job array (-1 if the job does not belong to any)
${TOTAL_TASKS} The total number of tasks in the job array (-1 if the job does not belong to any)
${ARCH} The architecture of the selected execution host
${PARAM} Allows the assignment of arbitrary start and increment values for array jobs (start + increment*GW_TASK_ID). Useful to generate file naming patterns or task processing. The values for start and increment will be specified with options -s (for start, with 0 by default) and -i (for increment, with 1 by default) of the gwsubmit command.
${MAX_PARAM} Upper bound of the ${PARAM} variable.


Important
The above variables can be used in any job option.

Resource selection expressions

Requirement expression syntax

The syntax of the requirement expressions is defined as:

  stmt::= expr';'
  expr::= VARIABLE '=' INTEGER
          | VARIABLE '>' INTEGER
          | VARIABLE '<' INTEGER
          | VARIABLE '=' STRING
          | expr '&' expr
          | expr '|' expr
          | '!' expr
          | '(' expr ')'

Each expression is evaluated to 1 (TRUE) or 0 (FALSE). Only those hosts for which the requirement expression is evaluated to TRUE will be considered to execute the job.

Logical operators are as expected ( less '<', greater '>', '&' AND, '|' OR, '!' NOT), '=' means equals with integers. When you use '=' operator with strings, it performs a shell wildcard pattern matching.

Examples:

  REQUIREMENTS = LRMS_NAME = "*pbs*"; # Only use pbs like jobmanagers
  REQUIREMENTS = HOSTNAME = "*.es"; #Only hosts in Spain
  REQUIREMENTS = HOSTNAME = "hydrus.dacya.ucm.es"; #Only use hydrus.dacya.ucm.es

Rank expression syntax

The syntax of the rank expressions is defined as:

  stmt::= expr';'
  expr::= VARIABLE
          | INTEGER
          | expr '+' expr
          | expr '-' expr
          | expr '*' expr
          | expr '/' expr
          | '-' expr
          | '(' expr ')'

Rank expressions are evaluated using each host information. '+', '-', '*', '/' and '-' are arithmetic operators, so only integer values should be used in rank expressions.

Requirement and rank variables

To set the REQUIREMENTS and RANK parameter values the following extended set of variables, provided by the Information Manager, can be used:

Table 3. Variables that can be used to define the job REQUIREMENTS and RANK.

Variable Description
HOSTNAME FQDN (Fully Qualified Domain Name) of the execution host (e.g. hydrus.dacya.ucm.es)
ARCH Architecture of the execution host (e.g. i686, alpha)
OS_NAME Operating System name of the execution host (e.g. Linux, SL)
OS_VERSION Operating System version of the execution host (e.g. 2.6.9-1.66, 3)
CPU_MODEL CPU model of the execution host (e.g. Intel(R) Pentium(R) 4 CPU 2, PIV)
CPU_MHZ CPU speed in MHz of the execution host
CPU_FREE Percentage of free CPU of the execution host
CPU_SMP CPU SMP size of the execution host
NODECOUNT Total number of nodes of the execution host
SIZE_MEM_MB Total memory size in MB of the execution host
FREE_MEM_MB Free memory in MB of the execution hosts
SIZE_DISK_MB Total disk space in MB of the execution hosts
FREE_DISK_MB Free disk space in MB of the execution hosts
LRMS_NAME Name of local DRM system (job manager) for execution, usually not fork (e.g. jobmanager-pbs, PBS, jobmanager-sge, SGE)
LRMS_TYPE Type of local DRM system for execution (e.g. pbs, sge)
QUEUE_NAME Name of the queue (e.g. default, short, dteam)
QUEUE_NODECOUNT Total node count of the queue
QUEUE_FREENODECOUNT Free node count of the queue
QUEUE_MAXTIME Maximum wall time of jobs in the queue
QUEUE_MAXCPUTIME Maximum CPU time of jobs in the queue
QUEUE_MAXCOUNT Maximum count of jobs that can be submitted in one request to the queue
QUEUE_MAXRUNNINGJOBS Maximum number of running jobs in the queue
QUEUE_MAXJOBSINQUEUE Maximum number of queued jobs in the queue
QUEUE_DISPATCHTYPE Dispatch type of the queue (e.g. batch, inmediate)
QUEUE_PRIORITY Priority of the queue
QUEUE_STATUS Status of the queue (e.g. active, production)


Job environment

Job environment variables can be easily set with the “ENVIRONMENT ” parameter of the Job Template. These environment variables are parsed, so you can use the GridWay variables defined in Section “Variable substitution”, to set the job environment.

Note
The variables defined in the ENVIRONMENT are sourced in a bash shell. In this way you can take advantage of the bash substitution capabilities and built-in functions. For example:
 ENVIRONMENT = VAR = "`expr ${JOB_ID} + 3`" # will set VAR to JOB_ID + 3 

In addition to those variables set in the “ENVIRONMENT ” parameter, GridWay set the following variables, that can be used by your applications:

  • GW_RESTARTED
  • GW_EXECUTABLE
  • GW_HOSTNAME
  • GW_ARCH
  • GW_CPU_MHZ
  • GW_MEM_MB
  • GW_RESTART_FILES
  • GW_CPULOAD_THRESHOLD
  • GW_ARGUMENTS
  • GW_TASK_ID
  • GW_CPU_MODEL
  • GW_ARRAY_ID
  • GW_TOTAL_TASKS
  • GW_JOB_ID
  • GW_OUTPUT_FILES
  • GW_INPUT_FILES
  • GW_OS_NAME
  • GW_USER
  • GW_DISK_MB
  • GW_OS_VERSION

Monitor and Wrapper Scripts

If you want to specify your own monitor or wrapper script you can do it using WRAPPER or MONITOR variables in your job template (or modifying the default one located in the etc directory of your GW installation). The file you specify must be a full path name of the script or a relative path if it is located inside $GW_LOCATION. Here you can not use gsiftp or file url protocol prefixes.

Job Submission Description Language (JSDL)

JSDL overview

GridWay supports Job Submission Description Language (JSDL). This specification is a language for describing the job requirements for submission to resources. The JSDL language specification is based on XML Schema that facilitate the expression of those requirements as a set of XML elements. More info at https://forge.gridforum.org/sf/projects/jsdl-wg

JSDL document structure

The JSDL document structure is as follows:

  <JobDefinition>
  |-------<JobDescription>
  |---------------<JobIdentification>
  |-----------------------<JobName>?
  |-----------------------<Description>?
  |-----------------------<JobAnnotation>*
  |-----------------------<JobProject>*
  |-----------------------<xsd:any##other>*
  |---------------</JobIdentification>?
  |---------------<Application>
  |-----------------------<ApplicationName>?
  |-----------------------<ApplicationVersion>?
  |-----------------------<Description>?
  |-----------------------<xsd:any##other>*
  |---------------</Application>?
  |---------------<Resources>?
  |-----------------------<CandidateHosts>
  |-------------------------------<HostName>+
  |-----------------------</CandidateHosts>?
  |-----------------------<FileSystem>
  |-------------------------------<Description>?
  |-------------------------------<MountPoint>?
  |-------------------------------<MountSource>?
  |-------------------------------<DiskSpace>?
  |-------------------------------<FileSystemType>?
  |-------------------------------<xsd:any##other>*
  |-----------------------</FileSystem>*
  |-----------------------<ExclusiveExecution>?
  |-----------------------<OperatingSystem>?
  |-------------------------------<OperatingSystemType>
  |---------------------------------------<OperatingSystemName>
  |---------------------------------------<xsd:any##other>*
  |-------------------------------</OperatingSystemType>?
  |-------------------------------<OperatingSystemVersion>?
  |-------------------------------<Description>?
  |-------------------------------<xsd:any##other>*
  |-----------------------</OperatingSystem>?
  |-----------------------<CPUArchitecture>
  |-------------------------------<CPUArchitectureName>
  |-------------------------------<xsd:any##other>*
  |-----------------------</CPUArchitecture>?
  |-----------------------<IndividualCPUSpeed>?
  |-----------------------<IndividualCPUTime>?
  |-----------------------<IndividualCPUCount>?
  |-----------------------<IndividualNetworkBandwidth>?
  |-----------------------<IndividualPhysicalMemory>?
  |-----------------------<IndividualVirtualMemory>?
  |-----------------------<IndividualDiskSpace>?
  |-----------------------<TOTALCPUTime>?
  |-----------------------<TOTALCPUCount>?
  |-----------------------<TOTALPhysicalMemory>?
  |-----------------------<TOTALVirtualMemory>?
  |-----------------------<TOTALDiskSpace>?
  |-----------------------<TOTALResourceCount>?
  |-----------------------<xsd:any##other>*
  |---------------</Resources>?
  |---------------<DataStaging>
  |-----------------------<FileName>
  |-----------------------<FileSystemName>?
  |-----------------------<CreationFlag>
  |-----------------------<DeleteOnTermination>?
  |-----------------------<Source>
  |-------------------------------<URI>?
  |-------------------------------<xsd:any##other>*
  |-----------------------</Source>?
  |-----------------------<Target>
  |-------------------------------<URI>?
  |-------------------------------<xsd:any##other>*
  |-----------------------</Target>?
  |-----------------------<xsd:any##other>*
  |---------------</DataStaging>*
  |---------------<xsd:any##other>*
  </JobDefinition>
Note
The symbol “?” denotes zero or one occurrences, “*” denotes zero or more occurrences and “+“ denotes one or more occurrences.

JSDL POSIX application

This schema defines the JSDL specification for describing an application executed on a POSIX compliance system. Due to GridWay Job Template specification, this schema MUST be included in the JSDL file. The JSDL POSIX Application Schema is as follow:

  <POSIXApplication name="xsd:NCName"?>
  |-------<Executable>?
  |-------<Argument>*
  |-------<Input>?
  |-------<Output>?
  |-------<Error>?
  |-------<WorkingDirectory>?
  |-------<Environment>*
  |-------<WallTimeLimit>?
  |-------<FileSizeLimit>?
  |-------<CoreDumpLimit>?
  |-------<DataSegmentLimit>?
  |-------<LockedMemoryLimit>?
  |-------<MemoryLimit>?
  |-------<OpenDescriptorsLimit>?
  |-------<PipeSizeLimit>?
  |-------<StackSizeLimit>?
  |-------<CPUTimeLimit>?
  |-------<ProcessCountLimit>?
  |-------<VirtualMemoryLimit>?
  |-------<ThreadCountLimit>?
  |-------<UserName>?
  |-------<GroupName>?
  </POSIXApplication name="xsd:NCName"?>

More details at https://forge.gridforum.org/sf/projects/jsdl-wg

JSDL HPC Profile Application Extension

This schema defines the JSDL specification for describing a simple HPC application that is made up of an executable file running within an operating system process. It shares much in common with the JSDL POSIXApplication. The JSDL HPC Application Schema is as follow:

  <HPCProfileApplication name="xsd:NCName"?>
  |-------<Executable>?
  |-------<Argument>*
  |-------<Input>?
  |-------<Output>?
  |-------<Error>?
  |-------<WorkingDirectory>?
  |-------<Environment>*
  |-------<UserNamet>?
  </HPCProfileApplication name="xsd:NCName"?>

More details at https://forge.gridforum.org/sf/projects/jsdl-wg

Job Submission Description Language versus GridWay Job Template

Next table compares JSDL and GridWay Job Template schema, and the JSDL elements support by the current GridWay version.

Table 4. JSDL vs GWJT

JSDL Element GWJT Attribute Adoption
JobDefinition - Supported
JobDescription - Supported
JobIdentification - Supported
JobName NAME Supported
JobAnnotation - Not supported
JobProject - Not supported
Application - Supported
ApplicationName - Supported
ApplicationVersion - Supported
Description - Supported
Resources - Supported
CandidateHosts - Supported
HostName HOSTNAME Supported
FileSystem - Not supported
MountPoint - Not supported
MountSource - Not supported
DiskSpace - Not supported
FileSystemType - Not supported
ExclusiveExecution - Not supported
OperatingSystem - Supported
OperatingSystemType - Supported
OperatingSystemName OS_NAME Supported
OperatingSystemVersion OS_VERSION Supported
CPUArchitecture - Supported
CPUArchitectureName ARCH Supported
IndividualCPUSpeed CPU_MHZ Supported
IndividualCPUTime - Not supported
IndividualCPUCount NODECOUNT Supported
IndividualNetworkBandwidth - Not supported
IndividualPhysicalMemory MEM_MB Supported
IndividualVirtualMemory - Not supported
IndividualDiskSpace SIZE_DISK_MB Supported
TOTALCPUTime - Not supported
TOTALCPUCount - Not supported
TOTALPhysicalMemory - Not supported
TOTALVirtualMemory - Not supported
TOTALDiskSpace - Not supported
TOTALResourceCount - Not supported
DataStaging - Supported
FileName - Supported
FileSystemName - Not supported
CreationFlag - Supported
DeleteOnTermination - Supported
Source INPUT_FILES Supported
Target OUTPUT_FILES Supported
URI - Supported
POSIXApplication - Supported
Executable EXECUTABLE Supported
Argument ARGUMENTS Supported
Input STDIN_FILE Supported
Output STDOUT_FILE Supported
Error STDERR_FILE Supported
WorkingDirectory - Not supported
Environment ENVIRONMENT Supported
WallTimeLimit - Not supported
FileSizeLimit - Not supported
CoreDumpLimit - Not supported
DataSegmentLimit - Not supported
LockedMemoryLimit - Not supported
MemoryLimit - Not supported
OpenDescriptorsLimit - Not supported
PipeSizeLimit - Not supported
StackSizeLimit - Not supported
CPUTimeLimit - Not supported
ProcessCountLimit - Not supported
VirtualMemoryLimit - Not supported
ThreadCountLimit - Not supported
UserName - Not supported
GroupName - Not supported


Examples

A simple example

This example shows the representation of a simple job in JSDL format and the translator of this example in Gridway Job Template format.

JSDL file

  <?xml version="1.0" encoding="UTF-8"?>
 
  <jsdl:JobDefinition xmlns="http://www.example.org/"
      xmlns:jsdl="http://schemas.ggf.org/jsdl/2005/11/jsdl"
      xmlns:jsdl-posix="http://schemas.ggf.org/jsdl/2005/11/jsdl-posix"
      xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance">
    <jsdl:JobDescription>
      <jsdl:JobIdentification>
        <jsdl:JobName>Simple Application GW Template vs JSDL</jsdl:JobName>
        <jsdl:Description> This is a simple example to describe the main
               differences between GW Template and the JSDL schema.
        </jsdl:Description>
      </jsdl:JobIdentification>
      <jsdl:Application>
        <jsdl:ApplicationName>ls</jsdl:ApplicationName>
        <jsdl-posix:POSIXApplication>
          <jsdl-posix:Executable>/bin/ls</jsdl-posix:Executable>
          <jsdl-posix:Argument>-la file.txt</jsdl-posix:Argument>
          <jsdl-posix:Environment name="LD_LIBRARY_PATH">/usr/local/lib</jsdl-posix:Environment>
          <jsdl-posix:Input>/dev/null</jsdl-posix:Input>
          <jsdl-posix:Output>stdout.${JOB_ID}</jsdl-posix:Output>
          <jsdl-posix:Error>stderr.${JOB_ID}</jsdl-posix:Error>
        </jsdl-posix:POSIXApplication>
      </jsdl:Application>
      <jsdl:Resources>
        <jsdl:CandidateHost>
          <jsdl:HostName>*.dacya.ucm.es</jsdl:HostName>
        </jsdl:CandidateHost>
        <jsdl:CPUArchitecture>
          <jsdl:CPUArchitectureName>x86_32</jsdl:CPUArchitectureName>
        </jsdl:CPUArchitecture>
      </jsdl:Resources>
      <jsdl:DataStaging>
        <jsdl:FileName>file.txt</jsdl:FileName>
        <jsdl:CreationFlag>overwrite</jsdl:CreationFlag>
        <jsdl:DeleteOnTermination>true</jsdl:DeleteOnTermination>
        <jsdl:Source>
          <jsdl:URI>gsiftp://hydrus.dacya.ucm.es/home/jose/file1.txt</jsdl:URI>
        </jsdl:Source>
      </jsdl:DataStaging>
      <jsdl:DataStaging>
        <jsdl:FileName>stdout.${JOB_ID}</jsdl:FileName>
        <jsdl:CreationFlag>overwrite</jsdl:CreationFlag>
        <jsdl:DeleteOnTermination>true</jsdl:DeleteOnTermination>
        <jsdl:Target>
          <jsdl:URI>gsiftp://hydrus.dacya.ucm.es/home/jose/stdout.${JOB_ID}</jsdl:URI>
        </jsdl:Target>
      </jsdl:DataStaging>
      <jsdl:DataStaging>
        <jsdl:FileName>stderr.${JOB_ID}</jsdl:FileName>
        <jsdl:CreationFlag>overwrite</jsdl:CreationFlag>
        <jsdl:DeleteOnTermination>true</jsdl:DeleteOnTermination>
        <jsdl:Target>
          <jsdl:URI>gsiftp://hydrus.dacya.ucm.es/home/jose/stderr.${JOB_ID}</jsdl:URI>
        </jsdl:Target>
      </jsdl:DataStaging>
    </jsdl:JobDescription>
  </jsdl:JobDefinition>

GridWay Job Template file

  #This file was automatically generated by the JSDL2GWJT parser
  EXECUTABLE=/bin/ls
  ARGUMENTS=-la file.txt
  STDIN_FILE=/dev/null
  STDOUT_FILE=stdout.${JOB_ID}
  STDERR_FILE=stderr.${JOB_ID}
  ENVIRONMENT=LD_LIBRARY_PATH=/usr/local/lib
  REQUIREMENTS=HOSTNAME="*.dacya.ucm.es" & ARCH="x86_32"
  INPUT_FILES=file.txt

A HPC profile example

This example shows the representation of a HPC profile job in JSDL format and the translator of this example in Gridway Job Template format.

JSDL file

  <?xml version="1.0" encoding="UTF-8"?>
 
  <jsdl:JobDefinition xmlns="http://www.example.org/"
      xmlns:jsdl="http://schemas.ggf.org/jsdl/2005/11/jsdl"
      xmlns:jsdl-posix="http://schemas.ggf.org/jsdl/2005/11/jsdl-posix"
      xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance">
    <jsdl:JobDescription>
      <jsdl:JobIdentification>
        <jsdl:JobName>Simple Application GW Template vs JSDL</jsdl:JobName>
        <jsdl:Description> This is a simple example to describe the main
               differences between GW Template and the JSDL schema.
        </jsdl:Description>
      </jsdl:JobIdentification>
      <jsdl:Application>
        <jsdl:ApplicationName>ls</jsdl:ApplicationName>
        <jsdl-hpcpa:HPCProfileApplication>
          <jsdl-hpcpa:Executable>/bin/ls</jsdl-hpcpa:Executable>
          <jsdl-hpcpa:Argument>-la file.txt</jsdl-hpcpa:Argument>
          <jsdl-hpcpa:Environment name="LD_LIBRARY_PATH">/usr/local/lib</jsdl-hpcpa:Environment>
          <jsdl-hpcpa:Input>/dev/null</jsdl-hpcpa:Input>
          <jsdl-hpcpa:Output>stdout.${JOB_ID}</jsdl-hpcpa:Output>
          <jsdl-hpcpa:Error>stderr.${JOB_ID}</jsdl-hpcpa:Error>
        </jsdl-hpcpa:HPCProfileApplication>
      </jsdl:Application>
      <jsdl:Resources>
        <jsdl:CandidateHost>
          <jsdl:HostName>*.dacya.ucm.es</jsdl:HostName>
        </jsdl:CandidateHost>
        <jsdl:CPUArchitecture>
          <jsdl:CPUArchitectureName>x86_32</jsdl:CPUArchitectureName>
        </jsdl:CPUArchitecture>
      </jsdl:Resources>
      <jsdl:DataStaging>
        <jsdl:FileName>file.txt</jsdl:FileName>
        <jsdl:CreationFlag>overwrite</jsdl:CreationFlag>
        <jsdl:DeleteOnTermination>true</jsdl:DeleteOnTermination>
        <jsdl:Source>
          <jsdl:URI>gsiftp://hydrus.dacya.ucm.es/home/jose/file1.txt</jsdl:URI>
        </jsdl:Source>
      </jsdl:DataStaging>
      <jsdl:DataStaging>
        <jsdl:FileName>stdout.${JOB_ID}</jsdl:FileName>
        <jsdl:CreationFlag>overwrite</jsdl:CreationFlag>
        <jsdl:DeleteOnTermination>true</jsdl:DeleteOnTermination>
        <jsdl:Target>
          <jsdl:URI>gsiftp://hydrus.dacya.ucm.es/home/jose/stdout.${JOB_ID}</jsdl:URI>
        </jsdl:Target>
      </jsdl:DataStaging>
      <jsdl:DataStaging>
        <jsdl:FileName>stderr.${JOB_ID}</jsdl:FileName>
        <jsdl:CreationFlag>overwrite</jsdl:CreationFlag>
        <jsdl:DeleteOnTermination>true</jsdl:DeleteOnTermination>
        <jsdl:Target>
          <jsdl:URI>gsiftp://hydrus.dacya.ucm.es/home/jose/stderr.${JOB_ID}</jsdl:URI>
        </jsdl:Target>
      </jsdl:DataStaging>
    </jsdl:JobDescription>
  </jsdl:JobDefinition>

GridWay Job Template file

  #This file was automatically generated by the JSDL2GWJT parser
  EXECUTABLE=/bin/ls
  ARGUMENTS=-la file.txt
  STDIN_FILE=/dev/null
  STDOUT_FILE=stdout.${JOB_ID}
  STDERR_FILE=stderr.${JOB_ID}
  ENVIRONMENT=LD_LIBRARY_PATH=/usr/local/lib
  REQUIREMENTS=HOSTNAME="*.dacya.ucm.es" & ARCH="x86_32"
  INPUT_FILES=file.txt

Usage Scenarios

Single jobs: Submitting and monitoring the simplest job

GWD should be configured and running. Check the Installation Guide and Configuration Guide to do that. Do not forget to create a proxy with grid-proxy-init

To submit a job, you will need a Job Template. The most simple Job Template in GridWay could be:

EXECUTABLE=/bin/ls

Save it as file jt in directory example.

Use the gwsubmit command to submit the job:

$ gwsubmit -t example/jt

Let see how many resources are available in our Grid, with gwhost:

  $ gwhost
  HID PRIO OS              ARCH   MHZ %CPU MEM(F/T)     DISK(F/T)  N(U/F/T) LRMS  HOSTNAME
  0   1    Linux2.6.17-2-6 x86   3215  100 923/2027 105003/118812     0/1/2 Fork  cygnus.dacya.ucm.es
  1   1    Linux2.6.17-2-6 x86   3216  189 384/2027 105129/118812     0/2/2 Fork  draco.dacya.ucm.es
  2   1    Linux2.6.18-3-a x86_6 2211  100 749/1003   76616/77844     0/2/2 SGE   aquila.dacya.ucm.es
  3   1                             0    0      0/0           0/0     0/0/0       hydrus.dacya.ucm.es
  4   1    Linux2.6.18-3-a x86_6 2009   74  319/878 120173/160796     0/1/1 Fork  orion.dacya.ucm.es
  5   1    Linux2.6.16.13- x86   3200  100  224/256       114/312     0/6/6 SGE   ursa.dacya.ucm.es
Note that hydrus is down in this example, so no information is received at all and of course, this host won't be considered in future scheduling decisions. If you want to retrieve more information about a single resource, issue the gwhost command followed by the host identification (HID):
  $ gwhost 0
  HID PRIO OS              ARCH   MHZ %CPU MEM(F/T)     DISK(F/T)  N(U/F/T) LRMS  HOSTNAME
  5   1    Linux2.6.16.13- x86   3200  100  224/256       114/312     0/6/6 SGE   ursa.dacya.ucm.es

  QUEUENAME            SL(F/T) WALLT CPUT  COUNT MAXR  MAXQ  STATUS   DISPATCH   PRIORITY
  all.q                6/6     0     0     0     6     0     enabled  NULL       0
You can also check the resources that match your requirements with gwhost -m 0.
  $ gwhost -m 0
  HID QNAME      RANK  PRIO  SLOTS HOSTNAME
  0   default    0     1     2     cygnus.dacya.ucm.es
  1   default    0     1     0     draco.dacya.ucm.es
  2   all.q      0     1     2     aquila.dacya.ucm.es
  4   default    0     1     1     orion.dacya.ucm.es
  5   all.q      0     1     6     ursa.dacya.ucm.es
Now, you can check the evolution of the job with the gwps command.

  $ gwps 
  USER  JID DM   EM   RWS START    END      EXEC    XFER    EXIT NAME  HOST
  user  0   pend ---- 000 10:42:09 --:--:-- 0:00:00 0:00:00 --   jt    --
and in succesive states
  user  0   prol ---- 000 10:42:09 --:--:-- 0:00:00 0:00:01 --   jt    cygnus.dacya.ucm.es/Fork
  user  0   wrap ---- 000 10:42:09 --:--:-- 0:00:27 0:00:04 --   jt    cygnus.dacya.ucm.es/Fork
  user  0   wrap pend 000 10:42:09 --:--:-- 0:00:27 0:00:04 --   jt    cygnus.dacya.ucm.es/Fork
  user  0   wrap actv 000 10:42:09 --:--:-- 0:00:27 0:00:04 --   jt    cygnus.dacya.ucm.es/Fork
  user  0   epil ---- 000 10:42:09 --:--:-- 0:00:31 0:00:05 --   jt    cygnus.dacya.ucm.es/Fork
  user  0   done ---- 000 10:42:09 10:43:01 0:00:31 0:00:08 0    jt    cygnus.dacya.ucm.es/Fork

At the beginning, the job is in pending state and not allocated to any resource. Then, the job is allocated to cygnus.dacya.ucm.es/Fork and begins the prolog stage.

Note
You can use option -c <delay> to refresh the output of the gwps command.

You can see the job history with the gwhistory command:

  $ gwhistory 0
  HID START    END      PROLOG  WRAPPER EPILOG  MIGR    REASON QUEUE    HOST
  0   10:42:22 10:43:01 0:00:04 0:00:31 0:00:04 0:00:00 ----   default  cygnus.dacya.ucm.es/Fork
Now it's time to retrieve the results. As you specified by default, the results of the execution of this job will be in the same folder, in a text file called sdtout_file.$JOB_ID.
  $ ls -lt example/
  total 8
  -rw-r--r-- 1 user staff  0 2007-02-20 10:42 stderr.0
  -rw-r--r-- 1 user staff 72 2007-02-20 10:42 stdout.0
  -rw-r--r-- 1 user staff 19 2007-02-20 10:33 jt
  $ cat example/stdout.0
  job.env
  stderr.execution
  stderr.wrapper
  stdout.execution
  stdout.wrapper
Done! You have done your first execution with GridWay!

Array jobs: Calculating the π number

Defining the problem

This is a well known exercise. For our purposes, we will calculate the integral of the following function:

pi1.jpg

Being f(x) = 4/(1+x2). So, π will be the integral of f(x) in the interval [0,1].

In order to calculate the whole integral, it's interesting to divide the function in several sections and compute them separately:

pi2.jpg

As you can see, the more sections you make, the more exact π will be:

pi3.jpg

So, you have a Grid with some nodes, you have GridWay… Why don't use them to calculate the π number by giving all the nodes a section to compute with only one command?

Note
You will find all the files needed to perform this example in the $GW_LOCATION/examples/pi directory.

The coding part

For this example, we have chosen the C Programming Language. Create a text file called pi.c and copy inside the following lines:

  #include <stdio.h>
  #include <string.h>
 
  int main (int argc, char** args)
  {
    int task_id;
    int total_tasks;
    long long int n;
    long long int i;
 
    double l_sum, x, h;
 
    task_id = atoi(args[1]);
    total_tasks = atoi(args[2]);
    n = atoll(args[3]);
 
    fprintf(stderr, "task_id=%d total_tasks=%d n=%lld\n", task_id, total_tasks, n);
 
    h = 1.0/n;
 
    l_sum = 0.0;
 
    for (i = task_id; i < n; i += total_tasks)
    {
      x = (i + 0.5)*h;
      l_sum += 4.0/(1.0 + x*x);
    }
 
    l_sum *= h;
 
    printf("%0.12g\n", l_sum);
 
    return 0;
  }

Now it's time to compile it:

$ gcc -O3 pi.c -o pi

after this, you should have an executable called pi. This command receives three parameters:

  • Task identifier: The identifier of the current task.
  • Total tasks: The number of tasks the computation should be divided into.
  • Number of intervals: The number of intervals over which the integral is being evaluated.

Defining the job

For making GridWay work with your program, you must create a Job Template. In this case, we will call it pi.jt. Copy the following lines inside:

  EXECUTABLE  = pi
  ARGUMENTS   = ${TASK_ID} ${TOTAL_TASKS} 100000
  STDOUT_FILE = stdout_file.${TASK_ID}
  STDERR_FILE = stderr_file.${TASK_ID}
  RANK        = CPU_MHZ

Submitting the jobs

This time, we will submit an array of jobs. This is done by issuing the following command:

  $ gwsubmit -v -t pi.jt -n 4
  ARRAY ID: 0

  TASK JOB
  0    0
  1    1
  2    2
  3    3
In order to wait for the jobs to complete, you can use the gwwait command.

The argument passed to gwwait is the array identifier given by gwsubmit when executed with the -v option. It could be also obtained through gwps

This command will block and return when all jobs have been executed:

  $ gwwait -v -A 0
  0   : 0
  1   : 0
  2   : 0
  3   : 0
This command, when issued with option -v shows the exit codes for each job in the array (usually, 0 means success).

Result post-processing

The execution of these jobs has returned some output files with the result of each execution:

  stdout_file.0
  stdout_file.1
  stdout_file.2
  stdout_file.3

Now, we will need something to sum the results inside each file. For this, you can use an awk script like the following:

  $ awk 'BEGIN {sum=0} {sum+=$1} END {printf "Pi is %0.12g\n", sum}' stdout_file.*
  Pi is 3.1415926536
Well, not much precision, right? You could try it again, but this time with a much higher number of intervals (e.g. 10,000,000,000). Would you increment also the number of tasks? Which would be the best compromise?

Do you imagine how easy would be to implement these steps in a shell script in order to perform them unattendedly? Here you are the prove:

  #!/bin/sh
 
  AID=`gwsubmit -v -t pi.jt -n 4 | head -1 | awk '{print $3}'`
 
  if [ $? -ne 0 ]
  then
      echo "Submission failed!"
      exit 1
  fi
 
  gwwait -v -A $AID
 
  if [ $? -eq 0 ]
  then
      awk 'BEGIN {sum=0} {sum+=$1} END {printf "Pi is %0.12g\n", sum}' stdout_file.*
  else
      echo "Some tasks failed!"
  fi

MPI jobs: Calculating the π number again

When applications show fine-grain parallelism, with small computation to communication ratio, and thus need lower latencies, MPI (Message Passing Interface) jobs give a better choice.

Following the π example, we will now perform its computation using MPI in a single job. Notice that MPI jobs can also be part of an array or complex job.

Note
You will find all the files needed to perform this example in the $GW_LOCATION/examples/mpi directory.

Create a text file called mpi.c and copy inside the following lines:

  #include "mpi.h"
  #include <stdio.h>
  #include <math.h>
 
  int main( int argc, char *argv[])
  {
      int done = 0, n, myid, numprocs, i;
      double PI25DT = 3.141592653589793238462643;
      double mypi, pi, h, sum, x;
      double startwtime = 0.0, endwtime;
      int  namelen;
      char processor_name[MPI_MAX_PROCESSOR_NAME];
 
      MPI_Init(&argc,&argv);
      MPI_Comm_size(MPI_COMM_WORLD,&numprocs);
      MPI_Comm_rank(MPI_COMM_WORLD,&myid);
      MPI_Get_processor_name(processor_name,&namelen);
 
      printf("Process %d on %s\n", myid, processor_name);
 
      n = 100000000;
 
      startwtime = MPI_Wtime();
 
      h   = 1.0 / (double) n;
      sum = 0.0;
      for (i = myid + 1; i <= n; i += numprocs)
      {
          x = h * ((double)i - 0.5);
          sum += 4.0 / (1.0 + x*x);
      }
      mypi = h * sum;
 
      MPI_Reduce(&mypi, &pi, 1, MPI_DOUBLE, MPI_SUM, 0, MPI_COMM_WORLD);
 
      if (myid == 0)
      {
          printf("pi is approximately %.16f, Error is %.16f\n",
                 pi, fabs(pi - PI25DT));
          endwtime = MPI_Wtime();
          printf("wall clock time = %f\n", endwtime-startwtime);
      }
 
      MPI_Finalize();
 
      return 0;
  }
Note
For more information about MPI, see http://www.mcs.anl.gov/mpi.

Notice that the above program already performs postprocessing in a single operand reduction operation MPI_Reduce which sums the partial results obtained by each processor.

Now it's time to compile it. Notice that you will need a compiler with MPI support like mpicc:

  $ mpicc -O3 mpi.c -o mpi
Now you must create a Job Template. In this case, we will call it “mpi.jt”:

  EXECUTABLE   = mpi
  STDOUT_FILE   = stdout.${JOB_ID}
  STDERR_FILE   = stderr.${JOB_ID}
  RANK          = CPU_MHZ
  TYPE          = "mpi"
  NP            = 2

Workflows

The powerful commands provided by GridWay to submit, control and synchronize jobs allow us to programmatically define complex jobs or workflows, where some jobs need data generated by other jobs. GridWay allows job submission to be dependent on the completion of other jobs. This new functionality provides support for the execution of workflows.

GridWay allows scientists and engineers to express their computational problems by using workflows. The capture of the job exit code allows users to define workflows, where each task depends on the output and exit code from the previous task. They may even involve branching, looping and spawning of subtasks, allowing the exploitation of the parallelism on the workflow of certain type of applications. The bash script flow control structures and the GridWay commands allow the development of workflows with the following functionality:

  • Sequence, parallelism, branching and looping structures
  • The workflow can be described in an abstract form without referring to specific resources for task execution
  • Quality of service constraints and fault tolerance are defined at task level

Job dependencies can be specified at submission by using the -d option of the gwsubmit command. A Job with dependencies will be submitted in the hold state, and once all the jobs on which it depends have successfully finished, it will be released. You can also release this job by hand with the gwkill.

A sample of DAG workflow

A DAG-based workflow consists of a temporal relationship between tasks, where the input, output or execution of one ore more tasks depends on one or more other tasks. For this example we have chosen a simple workflow.

Figure 2. Workflow example.

Workflow example.

In this example, job A generates a random number, jobs B and C add 1 to that number and, finally job D adds the result of these jobs. This is the final result is two times the number generated by A, plus two. In our case, the numbers are passed between jobs using the standard output files.

Job Template for job A (A.jt):

  EXECUTABLE=/bin/echo
  ARGUMENTS="$RANDOM"
  STDOUT_FILE=out.A

Job Template for jobs B and C (“B.jt” and “C.jt”):

  EXECUTABLE=/usr/bin/expr
  ARGUMENTS="`cat out.A`" + 1
  INPUT_FILES=out.A
  STDOUT_FILE=out.B #out.C for job C

Job Template for job D (“D.jt”):

  EXECUTABLE=/usr/bin/expr
  ARGUMENTS="`cat out.B`" + "`cat out.C`"
  INPUT_FILES=out.B, out.C
  STDOUT_FILE=out.workflow

Once you have set up the previous Job Templates, the workflow can be easily submitted with the following commands:

  $ gwsubmit -v -t A.jt
  JOB ID: 5

  $ gwsubmit -v -t B.jt -d "5"
  JOB ID: 6

  $ gwsubmit -v -t C.jt -d "5"
  JOB ID: 7

  $ gwsubmit -t C.jt -d "6 7"

Note
In the previous example, jobs B and C can be submitted as an array job using just one template with output, OUTPUT_FILES = out.${TASK_ID}. Therefore, input of job D will be INPUT_FILES = out.0, out.1.

The above steps can be easily implemented in a shell script.

  #!/bin/sh
 
  A_ID=`gwsubmit -v -t A.jt | cut -f2 -d':' | cut -f2 -d' '`
  B_ID=`gwsubmit -v -t B.jt -d "$A_ID" | cut -f2 -d':' | cut -f2 -d' '`
  C_ID=`gwsubmit -v -t C.jt -d "$A_ID" | cut -f2 -d':' | cut -f2 -d' '`
  D_ID=`gwsubmit -v -t D.jt -d "$B_ID $C_ID"| cut -f2 -d':' | cut -f2 -d' '`
 
  #Sync with last job of the workflow
  gwwait $D_ID
 
  echo "Random number `cat out.A`"
  echo "Workflow computation `cat out.workflow`"

Note that when input and output files vary depending on the iteration or job id number, you should generate Job Templates dynamically before submitting each job. This can be done programmatically by using the DRMAA API, or via shell scripting.

Using gwdag tool

Alternatively you can describe DAG workflows using a file similar to the one used by Condor DAGMAN. In this case the dependencies are not managed by GW but by the gwdag tool. You only have to substitute Condor job descriptions with GridWay job templates. Here is a file describing the same DAG as the previous example:

  JOB  A  A.jt
  JOB  B  B.jt
  JOB  C  C.jt
  JOB  D  D.jt
  PARENT A CHILD B C
  PARENT B C CHILD D

To submit this job you only have to specify the file describing this DAG to gwdag tool:

  $ gwdag <name of the file>
You can also get a DOT file for a DAG description that you can use later to generate a graph showing the flow using -d flag:
  $ gwdag -d <name of the file> > <name of the dot file>
Figure 3. Dag graph generated by the gwdag tool.

Dag graph generated by the gwdag tool.

Troubleshooting

Debugging job execution

GridWay reporting and accounting facilities provide information about overall performance and help debug job execution. GWD generates the following files under the $GW_LOCATION/var directory:

  • gwd.log: System level log. You can find log information of the activity of the middleware access drivers; and a coarse-grain log information about jobs.
  • $JOB_ID/job.log: Detailed log information for each job, it includes details of job state transitions, resource usage and performance.
  • $JOB_ID/stdout.wrapper: Standard output of the wrapper executable.
  • $JOB_ID/stderr.wrapper: Standard error output of the wrapper executable. By default, wrapper is executed with shell debugging options (-xv) active, so this is usually the best source of information in case of failure.

Frequent problems

Currently, many errors are handled silently and are only shown in the job.log file.

Also, there is a number of failures related to the underlying middleware (Globus in this case) that could make some jobs fail. It is a good idea to perform some basic testing of Globus when some jobs unexpectedly fail (check the GridWay 5.6: Configuration Guide to learn how to verify a Globus installation).

Use the GridWay lists, to submit your problems and they will be eventually appear in this guide.

Admin · Log In