Provided by: ffindex_0.9.9.7-4_amd64 bug

NAME

       ffindex_apply_mpi - apply a program to each FFindex entry (mpi enhanced)

DESCRIPTION

       --------------------------------------------------------------------------  The  value  of
       the MCA parameter "plm_rsh_agent" was set to a path that could not be found:

              plm_rsh_agent: ssh : rsh

       Please   either   unset   the   parameter,   or   check   that   the   path   is   correct
       --------------------------------------------------------------------------
       [lcy01-05:11100] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to start a daemon on the local
       node  in  file  ess_singleton_module.c  at  line  582 [lcy01-05:11100] [[INVALID],INVALID]
       ORTE_ERROR_LOG: Unable to start a daemon on the local node in file  ess_singleton_module.c
       at  line 166 -------------------------------------------------------------------------- It
       looks like orte_init failed for some reason; your parallel process  is  likely  to  abort.
       There  are  many  reasons that a parallel process can fail during orte_init; some of which
       are due to configuration or environment problems.  This failure appears to be an  internal
       failure;  here's  some  additional  information (which may only be relevant to an Open MPI
       developer):

              orte_ess_init failed --> Returned value Unable to start a daemon on the local  node
              (-127) instead of ORTE_SUCCESS

       --------------------------------------------------------------------------
       -------------------------------------------------------------------------- It  looks  like
       MPI_INIT failed for some reason; your parallel process is likely to abort.  There are many
       reasons that a parallel process can fail  during  MPI_INIT;  some  of  which  are  due  to
       configuration  or  environment  problems.  This failure appears to be an internal failure;
       here's some additional information (which may only be relevant to an Open MPI developer):

              ompi_mpi_init: ompi_rte_init failed --> Returned "Unable to start a daemon  on  the
              local node" (-127) instead of "Success" (0)

       --------------------------------------------------------------------------  ***  An  error
       occurred in MPI_Init *** on a NULL communicator  ***  MPI_ERRORS_ARE_FATAL  (processes  in
       this  communicator  will  now abort, ***    and potentially your MPI job) [lcy01-05:11100]
       Local abort before MPI_INIT completed completed successfully, but am not able to aggregate
       error messages, and not able to guarantee that all other processes were killed!

       The value of the MCA parameter "plm_rsh_agent" was set to a path that could not be found:

              plm_rsh_agent: ssh : rsh

       Please   either   unset   the   parameter,   or   check   that   the   path   is   correct
       --------------------------------------------------------------------------
       [lcy01-05:11102] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to start a daemon on the local
       node in file  ess_singleton_module.c  at  line  582  [lcy01-05:11102]  [[INVALID],INVALID]
       ORTE_ERROR_LOG:  Unable to start a daemon on the local node in file ess_singleton_module.c
       at line 166 --------------------------------------------------------------------------  It
       looks  like  orte_init  failed  for some reason; your parallel process is likely to abort.
       There are many reasons that a parallel process can fail during orte_init;  some  of  which
       are  due to configuration or environment problems.  This failure appears to be an internal
       failure; here's some additional information (which may only be relevant  to  an  Open  MPI
       developer):

              orte_ess_init  failed --> Returned value Unable to start a daemon on the local node
              (-127) instead of ORTE_SUCCESS

       --------------------------------------------------------------------------
       --------------------------------------------------------------------------  It  looks like
       MPI_INIT failed for some reason; your parallel process is likely to abort.  There are many
       reasons  that  a  parallel  process  can  fail  during  MPI_INIT; some of which are due to
       configuration or environment problems.  This failure appears to be  an  internal  failure;
       here's some additional information (which may only be relevant to an Open MPI developer):

              ompi_mpi_init:  ompi_rte_init  failed --> Returned "Unable to start a daemon on the
              local node" (-127) instead of "Success" (0)

       --------------------------------------------------------------------------  ***  An  error
       occurred  in  MPI_Init  ***  on a NULL communicator *** MPI_ERRORS_ARE_FATAL (processes in
       this communicator will now abort, ***    and potentially your  MPI  job)  [lcy01-05:11102]
       Local abort before MPI_INIT completed completed successfully, but am not able to aggregate
       error messages, and not able to guarantee that all other processes were killed!

BUGS

       User feedback is welcome, especially bugs, performance  issues  and  last  but  not  least
       convenience of the programs and API.

       Email Andreas Hauser hauser@genzentrum.lmu.de.

ffindex_apply_mpi --------------------------June-2017------------------------FFINDEX-APPLY-MPI(1)