Provided by: lam4-dev_7.1.2-1.4_i386 bug


       appschema - LAM application schema format


       # comments
       [<where>] [-np #] [-s <where>] [-wd <dir>] [-x <env>] <program> [<args>]
       [<where>] [-np #] [-s <where>] [-wd <dir>] [-x <env>] <program> [<args>]


       The application schema is an ASCII file containing a description of the
       programs which constitute an application.  It  is  used  by  mpirun(1),
       MPI_Comm_spawn, and MPI_Comm_spawn_multiple to start an MPI application
       (the MPI_Info key "file" can be  used  to  specify  an  app  schema  to
       MPI_Comm_spawn  and  MPI_Comm_spawn_multiple).   All  tokens  after the
       program name will be passed  as  command  line  arguments  to  the  new
       processes.   Ordering  of the other elements on the command line is not

       The meaning of the options is  the  same  as  in  mpirun(1).   See  the
       mpirun(1)  man  page  for a lengthy discussion of the nomenclature used
       for <where>.  Note, however, that if -wd is  used  in  the  application
       schema  file,  it  will override any -wd value specified on the command

       For each program line, processes will be created on LAM nodes according
       to the presence of <where> and the process count option (-np).

       only <where>  One process is created on each node.

       only -np      The  specified  number  of processes are scheduled across
                     all LAM nodes/CPUs.

       both          The specified number of processes  are  scheduled  across
                     the specified nodes/CPUs.

       neither       One process is created on the local node.

   Program Transfer
       By  default,  LAM  searches  for executable programs on the target node
       where a particular instantiation will run.  If the file system  is  not
       shared, the target nodes are homogeneous, and the program is frequently
       recompiled, it can be convenient to have LAM transfer the program  from
       a  source  node  (usually  the local node) to each target node.  The -s
       option specifies this behaviour and identifies the single source  node.


       # Example application schema
       # Note that it may be necessary to specify the entire pathname for
       # "master" and "slave" if you get "File not found" errors from
       # mpirun(1).
       # This schema starts a "master" process on CPU 0 with the argument
       # "42.0", and then 10 "slave" processes (that are all sent from the
       # local node) scheduled across all available CPUs.
       c0 master 42.0
       C -np 10 -s h slave


       mpirun(1),        MPI_Comm_spawn(2),        MPI_Comm_Spawn_multiple(2),
       MPIL_Spawn(2), introu(1)