X-Git-Url: http://git.salome-platform.org/gitweb/?a=blobdiff_plain;f=src%2FParaMEDMEM%2FMPIProcessorGroup.cxx;h=3bf01b69d864b6197cbc5a2c003be28bee1e1403;hb=9f966e0ed55cc976b59a17318a7be82445529eda;hp=fc3f0188a949c5c4022fa190fbbd87fc744573ce;hpb=48782c06022ca2caa36f849cb5a29ea4fe2aaa83;p=tools%2Fmedcoupling.git diff --git a/src/ParaMEDMEM/MPIProcessorGroup.cxx b/src/ParaMEDMEM/MPIProcessorGroup.cxx index fc3f0188a..3bf01b69d 100644 --- a/src/ParaMEDMEM/MPIProcessorGroup.cxx +++ b/src/ParaMEDMEM/MPIProcessorGroup.cxx @@ -1,21 +1,22 @@ -// Copyright (C) 2007-2008 CEA/DEN, EDF R&D +// Copyright (C) 2007-2015 CEA/DEN, EDF R&D // -// This library is free software; you can redistribute it and/or -// modify it under the terms of the GNU Lesser General Public -// License as published by the Free Software Foundation; either -// version 2.1 of the License. +// This library is free software; you can redistribute it and/or +// modify it under the terms of the GNU Lesser General Public +// License as published by the Free Software Foundation; either +// version 2.1 of the License, or (at your option) any later version. // -// This library is distributed in the hope that it will be useful, -// but WITHOUT ANY WARRANTY; without even the implied warranty of -// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU -// Lesser General Public License for more details. +// This library is distributed in the hope that it will be useful, +// but WITHOUT ANY WARRANTY; without even the implied warranty of +// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU +// Lesser General Public License for more details. // -// You should have received a copy of the GNU Lesser General Public -// License along with this library; if not, write to the Free Software -// Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA +// You should have received a copy of the GNU Lesser General Public +// License along with this library; if not, write to the Free Software +// Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA // -// See http://www.salome-platform.org/ or email : webmaster.salome@opencascade.com +// See http://www.salome-platform.org/ or email : webmaster.salome@opencascade.com // + #include "ProcessorGroup.hxx" #include "MPIProcessorGroup.hxx" #include "CommInterface.hxx" @@ -28,39 +29,38 @@ using namespace std; -/*! \defgroup processor_group Processor Groups - * - * \section processor_group_overview Overview - * The MPIProcessorGroup class is used to set up processor groups that help to define - * the MPI topology of the couplings. They can be set up in various ways, the most common being - * the use of the \c MPIProcessorGroup(Comminterface, int pfirst, int plast) - * constructor. - * - * The following code excerpt creates two processor groups on respectively 3 and 2 processors. - \verbatim - int main() - { - MPI_Init(&argc,&argv); - CommInterface comm_interface; - MPIProcessorGroup codeA_group(comm_interface, 0, 2); - MPIProcessorGroup codeB_group(comm_interface, 3, 4); - - ... - } - \endverbatim -*/ - namespace ParaMEDMEM { - /*! - \addtogroup processor_group - @{ + /*! + \anchor MPIProcessorGroup-det + \class MPIProcessorGroup + + The MPIProcessorGroup class represents a set of distinct "processors" (computation nodes) + in a MPI code. It is used to define the MPI topology of code couplings. + + Groups can be set up in various ways, the most common being + the use of the \c MPIProcessorGroup(Comminterface, int pfirst, int plast) + constructor. + + The following code excerpt creates two processor groups on respectively 3 and 2 processors. + \verbatim + int main() + { + MPI_Init(&argc,&argv); + CommInterface comm_interface; + MPIProcessorGroup codeA_group(comm_interface, 0, 2); // groups processors 0, 1 and 2 + MPIProcessorGroup codeB_group(comm_interface, 3, 4); // groups processors 3 and 4 + + ... + } + \endverbatim */ + /*! * Creates a processor group that is based on all the - MPI_COMM_WORLD processor.This routine must be called by all processors in MPI_COMM_WORLD. + processors of MPI_COMM_WORLD .This routine must be called by all processors in MPI_COMM_WORLD. \param interface CommInterface object giving access to the MPI communication layer */ @@ -86,7 +86,13 @@ namespace ParaMEDMEM */ MPIProcessorGroup::MPIProcessorGroup(const CommInterface& interface, set proc_ids, const MPI_Comm& world_comm): - ProcessorGroup(interface, proc_ids),_world_comm(world_comm) + ProcessorGroup(interface, proc_ids), _world_comm(world_comm) + { + updateMPISpecificAttributes(); + } + + + void MPIProcessorGroup::updateMPISpecificAttributes() { //Creation of a communicator MPI_Group group_world; @@ -97,19 +103,27 @@ namespace ParaMEDMEM _comm_interface.commRank(_world_comm,&rank_world); _comm_interface.commGroup(_world_comm, &group_world); - int* ranks=new int[proc_ids.size()]; + int* ranks=new int[_proc_ids.size()]; // copying proc_ids in ranks - copy::const_iterator,int*> (proc_ids.begin(), proc_ids.end(), ranks); - for (int i=0; i< proc_ids.size();i++) + copy::const_iterator,int*> (_proc_ids.begin(), _proc_ids.end(), ranks); + for (int i=0; i< (int)_proc_ids.size();i++) if (ranks[i]>size_world-1) - throw INTERP_KERNEL::Exception("invalid rank in set argument of MPIProcessorGroup constructor"); + { + delete[] ranks; + _comm_interface.groupFree(&group_world); // MPI_Group is a C structure and won't get de-allocated automatically? + throw INTERP_KERNEL::Exception("invalid rank in set argument of MPIProcessorGroup constructor"); + } - _comm_interface.groupIncl(group_world, proc_ids.size(), ranks, &_group); + _comm_interface.groupIncl(group_world, _proc_ids.size(), ranks, &_group); _comm_interface.commCreate(_world_comm, _group, &_comm); + + // clean-up delete[] ranks; + _comm_interface.groupFree(&group_world); // MPI_Group is a C structure and won't get de-allocated automatically? } + /*! Creates a processor group that is based on the processors between \a pstart and \a pend. This routine must be called by all processors in MPI_COMM_WORLD. @@ -118,7 +132,7 @@ namespace ParaMEDMEM \param pstart id in MPI_COMM_WORLD of the first processor in the group \param pend id in MPI_COMM_WORLD of the last processor in the group */ - MPIProcessorGroup::MPIProcessorGroup (const CommInterface& comm_interface, int pstart, int pend): ProcessorGroup(comm_interface,pstart,pend),_world_comm(MPI_COMM_WORLD) + MPIProcessorGroup::MPIProcessorGroup (const CommInterface& comm_interface, int pstart, int pend, const MPI_Comm& world_comm): ProcessorGroup(comm_interface,pstart,pend),_world_comm(world_comm) { //Creation of a communicator MPI_Group group_world; @@ -130,7 +144,10 @@ namespace ParaMEDMEM _comm_interface.commGroup(_world_comm, &group_world); if (pend>size_world-1 || pend proc_ids) : - ProcessorGroup(proc_group.getCommInterface()),_world_comm(MPI_COMM_WORLD) + ProcessorGroup(proc_group.getCommInterface()), + _world_comm(MPI_COMM_WORLD), _group(MPI_GROUP_NULL), _comm(MPI_COMM_NULL) { cout << "MPIProcessorGroup (const ProcessorGroup& proc_group, set proc_ids)" < procs; for (set::const_iterator iter=_proc_ids.begin(); iter!= _proc_ids.end(); iter++) procs.insert(*iter); - return new MPIProcessorGroup(_comm_interface, procs); + return new MPIProcessorGroup(_comm_interface, procs, _world_comm); } }