Salome HOME
mv ExplicitMapping.cxx to medtool subdirectory agr/medcalc
authorCédric Aguerre <cedric.aguerre@edf.fr>
Wed, 2 Dec 2015 13:17:49 +0000 (14:17 +0100)
committerCédric Aguerre <cedric.aguerre@edf.fr>
Wed, 2 Dec 2015 13:17:49 +0000 (14:17 +0100)
medtool/src/ParaMEDMEM/ExplicitMapping.cxx [new file with mode: 0644]
medtool/src/ParaMEDMEM/README [new file with mode: 0644]
src/ParaMEDMEM/ExplicitMapping.cxx [deleted file]
src/ParaMEDMEM/README [deleted file]

diff --git a/medtool/src/ParaMEDMEM/ExplicitMapping.cxx b/medtool/src/ParaMEDMEM/ExplicitMapping.cxx
new file mode 100644 (file)
index 0000000..f4e7545
--- /dev/null
@@ -0,0 +1,159 @@
+// Copyright (C) 2007-2015  CEA/DEN, EDF R&D
+//
+// This library is free software; you can redistribute it and/or
+// modify it under the terms of the GNU Lesser General Public
+// License as published by the Free Software Foundation; either
+// version 2.1 of the License, or (at your option) any later version.
+//
+// This library is distributed in the hope that it will be useful,
+// but WITHOUT ANY WARRANTY; without even the implied warranty of
+// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
+// Lesser General Public License for more details.
+//
+// You should have received a copy of the GNU Lesser General Public
+// License along with this library; if not, write to the Free Software
+// Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA  02111-1307 USA
+//
+// See http://www.salome-platform.org/ or email : webmaster.salome@opencascade.com
+//
+
+#include "ExplicitMapping.hxx"
+
+namespace ParaMEDMEM
+{
+
+  ExplicitMapping::ExplicitMapping():
+        _mapping(), _distant_domains(),
+        _numbers(0), _domains(0), _comm_buffer(0),
+        _buffer_index(0), _send_counts(0)
+  { }
+
+  ExplicitMapping::~ExplicitMapping()
+  {
+    if (_domains!=0) delete[] _domains;
+    if (_numbers!=0) delete[] _numbers;
+    if (_comm_buffer!=0) delete[] _comm_buffer;
+  }
+
+  void ExplicitMapping::pushBackElem(std::pair<int,int> idistant)
+  {
+    _mapping.push_back(idistant);
+  }
+
+  void  ExplicitMapping::setDistantElem(int ilocal, std::pair<int,int> idistant)
+  {
+    _mapping[ilocal]=idistant;
+  }
+
+  int ExplicitMapping::nbDistantDomains()
+  {
+    if (_distant_domains.empty())
+      {
+        for (std::vector <std::pair<int,int> >::const_iterator iter= _mapping.begin();
+            iter!=_mapping.end();
+            iter++)
+          _distant_domains.insert(iter->first);
+      }
+    return _distant_domains.size();
+  }
+
+  std::pair <int,int> ExplicitMapping::getDistantNumbering(int ielem)const
+  {
+    return _mapping[ielem];
+  }
+
+  int ExplicitMapping::getDistantDomain(int i)
+  {
+    if (_domains==0)
+      computeNumbers();
+
+    return _domains[i];
+  }
+
+  int ExplicitMapping::getNbDistantElems(int i)
+  {
+    if (_numbers==0)
+      computeNumbers();
+    return _numbers[i];
+  }
+
+  int* ExplicitMapping::serialize(int idproc)
+  {
+    _comm_buffer=new int[_mapping.size()*2];
+    std::vector<int> offsets(_distant_domains.size());
+    offsets[0]=0;
+    for (int i=1; i<(int)_distant_domains.size();i++)
+      offsets[i]=offsets[i-1]+_numbers[i-1];
+
+    for (int i=0; i<(int)_mapping.size(); i++)
+      {
+        int offset= offsets[_mapping[i].first];
+        _comm_buffer[offset*2]=idproc;
+        _comm_buffer[offset*2+1]=_mapping[i].second;
+        offsets[_mapping[i].first]++;
+      }
+    return _comm_buffer;
+  }
+
+  void ExplicitMapping::unserialize(int nbprocs, int* sizes,int nbtarget, int* targetrank, int* commbuffer)
+  {
+    int total_size=0;
+    for (int i=0; i< nbprocs; i++)
+      total_size+=sizes[i];
+
+    _mapping.resize(total_size);
+    _buffer_index=new int[total_size];
+    int indmap=0;
+    for (int i=0; i<nbprocs; i++)
+      for (int ielem=0; ielem<sizes[i]; ielem++)
+        {
+          _mapping[indmap].first=i;
+          _mapping[indmap].second=commbuffer[indmap*2+1];
+          _buffer_index[indmap]=commbuffer[indmap*2+1];
+          indmap++;
+        }
+    _numbers=new int [nbtarget];
+    _domains=new int [nbtarget];
+
+    int index=0;
+    for (int i=0; i<nbtarget; i++)
+      {
+        if (sizes[targetrank[i]]>0)
+          {
+            _numbers[index]=sizes[targetrank[i]];
+            _domains[index]=i;
+            index++;
+          }
+      }
+    _send_counts=new int[nbprocs];
+    for (int i=0; i<nbprocs; i++)
+      _send_counts[i]=sizes[i];
+  }
+
+  void ExplicitMapping::computeNumbers()
+  {
+    std::map <int,int> counts;
+    if (_numbers==0)
+      {
+        _numbers=new int[nbDistantDomains()];
+        _domains=new int[nbDistantDomains()];
+        for (int i=0; i<(int)_mapping.size(); i++)
+          {
+            if ( counts.find(_mapping[i].first) == counts.end())
+              counts.insert(std::make_pair(_mapping[i].first,1));
+            else
+              (counts[_mapping[i].first])++;
+          }
+        int counter=0;
+        for (std::map<int,int>::const_iterator iter=counts.begin();
+            iter!=counts.end();
+            iter++)
+          {
+            _numbers[counter]=iter->second;
+            _domains[counter]=iter->first;
+            counter++;
+          }
+      }
+  }
+
+}
diff --git a/medtool/src/ParaMEDMEM/README b/medtool/src/ParaMEDMEM/README
new file mode 100644 (file)
index 0000000..8a72f39
--- /dev/null
@@ -0,0 +1,194 @@
+Tests avec CPPUNIT 
+==================
+
+L'appel a MPI_Init n'est fait qu'une seule fois pour chaque suite de test
+(voir TestMPIAccessDEC.cxx, TestMPIAccess.cxx, TestParaMEDMEM.cxx qui inclut
+tous MPIMainTest.hxx)
+Il est suivi par l'execution de toute la suite des tests regroupes
+dans les trois headers MPIAccess.hxx, MPIAccessDEC.hxx, ParaMEDMEMTest.hxx
+On a enfin un seul appel a MPI_Finalize.
+
+Si un des tests d'une suite de tests comporte une anomalie cela
+peut avoir des implications sur l'execution des tests suivants.
+
+[ABN-2015] Lancement des tests avec valgrind
+============================================
+
+Exemple avec MPICH :
+
+rm tmp_log* ; mpirun -np 3 valgrind --leak-check=full --show-reachable=yes --log-file=tmp_log%p  src/ParaMEDMEMTest/TestParaMEDMEM && cat tmp_log*
+
+[ABN-2015] Lancement des tests avec gdb
+=======================================
+
+Exemple avec MPICH :
+
+mpirun -np 2 xterm -e gdb src/ParaMEDMEMTest/TestMPIAccess
+
+
+
+
+
+*************************************************************************************************************
+*********************** ce qui suit est obsolète ??                    **************************************
+*************************************************************************************************************
+
+Lancement des tests de TestParaMEDMEM avec CPPUNIT et TotalView (option -tv) :
+==============================================================================
+
+mpirun -np 5 -ssi rpi tcp C -tv -v -x PATH=${PATH},LD_LIBRARY_PATH=${LD_LIBRARY_PATH} TestParaMEDMEM
+
+Il arrive qu'on ne puisse pas utiliser totalview par manque de
+license.
+
+
+
+Lancement des tests de TestParaMEDMEM avec CPPUNIT et Valgrind avec "memory leaks" :
+====================================================================================
+
+mpirun -np 5 -ssi rpi tcp C -v -x PATH=${PATH},LD_LIBRARY_PATH=${LD_LIBRARY_PATH} valgrind --leak-check=full TestParaMEDMEM
+
+
+Lancement des tests fonctionnels de MPI_AccessDEC avec CPPUNIT :
+================================================================
+
+mpirun -np 11 -ssi rpi tcp C -v -x PATH=${PATH},LD_LIBRARY_PATH=${LD_LIBRARY_PATH} valgrind --leak-check=full TestMPIAccessDEC
+
+
+Lancement des tests unitaires de MPI_Access avec CPPUNIT :
+==========================================================
+
+mpirun -np 3 -ssi rpi tcp C -v -x PATH=${PATH},LD_LIBRARY_PATH=${LD_LIBRARY_PATH} valgrind --leak-check=full TestMPIAccess
+
+
+TestMPIAccess/TestMPIAccessDEC/TestParaMEDMEM et gcov :
+=======================================================
+
+Les resultats sont dans les repertoires suivants de $MED_BUILD_DIR/src/ParaMEDMEM/Test :
+-------------
+
+    TestParaMEDMEM-gcov/
+    TestMPIAccessDEC-gcov/
+    TestMPIAccess-gcov/
+
+Je n'y ai pas trouve d'anomalies.
+
+compilation : -fprofile-arcs -ftest-coverage
+-------------
+
+$MED_BUILD_DIR/src/ParaMEDMEM/makefile.in : LIB=libparamedar.a \
+-------------------------------------------     libparamed.la
+
+$MED_BUILD_DIR/src/ParaMEDMEM/Test/makefile.in : LIB = libParaMEDMEMTestar.a \
+------------------------------------------------       libParaMEDMEMTest.la
+
+execution et gcov :
+-------------------
+
+Pour pouvoir traiter les .cxx de ${MED_BUILD_DIR}/src/ParaMEDMEM et de
+${MED_BUILD_DIR}/src/ParaMEDMEM/Test, on execute deux fois gcov.
+
+cd ${MED_BUILD_DIR}/src/ParaMEDMEM/Test
+
+mpirun -np 3 -ssi rpi tcp C -v -x PATH=${PATH},LD_LIBRARY_PATH=${LD_LIBRARY_PATH} TestMPIAccess
+
+gcov TestMPIAccess.cxx test_MPI_Access_Send_Recv.cxx \
+                       test_MPI_Access_Cyclic_Send_Recv.cxx \
+                       test_MPI_Access_SendRecv.cxx \
+                       test_MPI_Access_ISend_IRecv.cxx \
+                       test_MPI_Access_Cyclic_ISend_IRecv.cxx \
+                       test_MPI_Access_ISendRecv.cxx \
+                       test_MPI_Access_Probe.cxx \
+                       test_MPI_Access_IProbe.cxx \
+                       test_MPI_Access_Cancel.cxx \
+                       test_MPI_Access_Send_Recv_Length.cxx \
+                       test_MPI_Access_ISend_IRecv_Length.cxx \
+                       test_MPI_Access_ISend_IRecv_Length_1.cxx \
+                       test_MPI_Access_Time.cxx \
+                       test_MPI_Access_Time_0.cxx \
+                       test_MPI_Access_ISend_IRecv_BottleNeck.cxx \
+                       ../MPI_Access.cxx
+gcov -o ../ TestMPIAccess.cxx test_MPI_Access_Send_Recv.cxx \
+                              test_MPI_Access_Cyclic_Send_Recv.cxx \
+                              test_MPI_Access_SendRecv.cxx \
+                              test_MPI_Access_ISend_IRecv.cxx \
+                              test_MPI_Access_Cyclic_ISend_IRecv.cxx \
+                              test_MPI_Access_ISendRecv.cxx \
+                              test_MPI_Access_Probe.cxx \
+                              test_MPI_Access_IProbe.cxx \
+                              test_MPI_Access_Cancel.cxx \
+                              test_MPI_Access_Send_Recv_Length.cxx \
+                              test_MPI_Access_ISend_IRecv_Length.cxx \
+                              test_MPI_Access_ISend_IRecv_Length_1.cxx \
+                              test_MPI_Access_Time.cxx \
+                              test_MPI_Access_Time_0.cxx \
+                              test_MPI_Access_ISend_IRecv_BottleNeck.cxx \
+                              ../MPI_Access.cxx
+
+
+cd ${MED_BUILD_DIR}/src/ParaMEDMEM/Test
+mpirun -np 11 -ssi rpi tcp C -v -x PATH=${PATH},LD_LIBRARY_PATH=${LD_LIBRARY_PATH} TestMPIAccessDEC
+
+gcov TestMPIAccessDEC.cxx test_AllToAllDEC.cxx \
+                          test_AllToAllvDEC.cxx \
+                          test_AllToAllTimeDEC.cxx \
+                          test_AllToAllvTimeDEC.cxx \
+                          test_AllToAllvTimeDoubleDEC.cxx \
+                          ../TimeInterpolator.cxx \
+                          ../LinearTimeInterpolator.cxx \
+                          ../MPI_Access.cxx \
+                          ../MPI_AccessDEC.cxx
+gcov -o ../ TestMPIAccessDEC.cxx test_AllToAllDEC.cxx \
+                                 test_AllToAllvDEC.cxx \
+                                 test_AllToAllTimeDEC.cxx \
+                                 test_AllToAllvTimeDEC.cxx \
+                                 test_AllToAllvTimeDoubleDEC.cxx \
+                                 ../TimeInterpolator.cxx \
+                                 ../LinearTimeInterpolator.cxx \
+                                 ../MPI_Access.cxx \
+                                 ../MPI_AccessDEC.cxx
+
+cd ${MED_BUILD_DIR}/src/ParaMEDMEM/Test
+mpirun -np 5 -ssi rpi tcp C -v -x PATH=${PATH},LD_LIBRARY_PATH=${LD_LIBRARY_PATH} TestParaMEDMEM
+
+gcov TestParaMEDMEM.cxx ParaMEDMEMTest.cxx \
+                             ParaMEDMEMTest_MPIProcessorGroup.cxx \
+                             ParaMEDMEMTest_BlockTopology.cxx \
+                             ParaMEDMEMTest_InterpKernelDEC.cxx \
+                        ../BlockTopology.cxx \
+                        ../ComponentTopology.cxx \
+                        ../DEC.cxx \
+                        ../ElementLocator.cxx \
+                        ../InterpolationMatrix.cxx \
+                        ../InterpKernelDEC.cxx \
+                        ../MPIProcessorGroup.cxx \
+                        ../MxN_Mapping.cxx \
+                        ../ParaFIELD.cxx \
+                        ../ParaMESH.cxx \
+                        ../ParaSUPPORT.cxx \
+                        ../ProcessorGroup.cxx \
+                        ../TimeInterpolator.cxx \
+                        ../LinearTimeInterpolator.cxx \
+                        ../MPI_Access.cxx \
+                        ../MPI_AccessDEC.cxx
+
+gcov -o ../ TestParaMEDMEM.cxx ParaMEDMEMTest.cxx \
+                                    ParaMEDMEMTest_MPIProcessorGroup.cxx \
+                                    ParaMEDMEMTest_BlockTopology.cxx \
+                                    ParaMEDMEMTest_InterpKernelDEC.cxx \
+                               ../BlockTopology.cxx \
+                               ../ComponentTopology.cxx \
+                               ../DEC.cxx \
+                               ../ElementLocator.cxx \
+                               ../InterpolationMatrix.cxx \
+                               ../InterpKernelDEC.cxx \
+                               ../MPIProcessorGroup.cxx \
+                               ../MxN_Mapping.cxx \
+                               ../ParaFIELD.cxx \
+                               ../ParaMESH.cxx \
+                               ../ParaSUPPORT.cxx \
+                               ../ProcessorGroup.cxx \
+                               ../TimeInterpolator.cxx \
+                               ../LinearTimeInterpolator.cxx \
+                               ../MPI_Access.cxx \
+                               ../MPI_AccessDEC.cxx
diff --git a/src/ParaMEDMEM/ExplicitMapping.cxx b/src/ParaMEDMEM/ExplicitMapping.cxx
deleted file mode 100644 (file)
index f4e7545..0000000
+++ /dev/null
@@ -1,159 +0,0 @@
-// Copyright (C) 2007-2015  CEA/DEN, EDF R&D
-//
-// This library is free software; you can redistribute it and/or
-// modify it under the terms of the GNU Lesser General Public
-// License as published by the Free Software Foundation; either
-// version 2.1 of the License, or (at your option) any later version.
-//
-// This library is distributed in the hope that it will be useful,
-// but WITHOUT ANY WARRANTY; without even the implied warranty of
-// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
-// Lesser General Public License for more details.
-//
-// You should have received a copy of the GNU Lesser General Public
-// License along with this library; if not, write to the Free Software
-// Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA  02111-1307 USA
-//
-// See http://www.salome-platform.org/ or email : webmaster.salome@opencascade.com
-//
-
-#include "ExplicitMapping.hxx"
-
-namespace ParaMEDMEM
-{
-
-  ExplicitMapping::ExplicitMapping():
-        _mapping(), _distant_domains(),
-        _numbers(0), _domains(0), _comm_buffer(0),
-        _buffer_index(0), _send_counts(0)
-  { }
-
-  ExplicitMapping::~ExplicitMapping()
-  {
-    if (_domains!=0) delete[] _domains;
-    if (_numbers!=0) delete[] _numbers;
-    if (_comm_buffer!=0) delete[] _comm_buffer;
-  }
-
-  void ExplicitMapping::pushBackElem(std::pair<int,int> idistant)
-  {
-    _mapping.push_back(idistant);
-  }
-
-  void  ExplicitMapping::setDistantElem(int ilocal, std::pair<int,int> idistant)
-  {
-    _mapping[ilocal]=idistant;
-  }
-
-  int ExplicitMapping::nbDistantDomains()
-  {
-    if (_distant_domains.empty())
-      {
-        for (std::vector <std::pair<int,int> >::const_iterator iter= _mapping.begin();
-            iter!=_mapping.end();
-            iter++)
-          _distant_domains.insert(iter->first);
-      }
-    return _distant_domains.size();
-  }
-
-  std::pair <int,int> ExplicitMapping::getDistantNumbering(int ielem)const
-  {
-    return _mapping[ielem];
-  }
-
-  int ExplicitMapping::getDistantDomain(int i)
-  {
-    if (_domains==0)
-      computeNumbers();
-
-    return _domains[i];
-  }
-
-  int ExplicitMapping::getNbDistantElems(int i)
-  {
-    if (_numbers==0)
-      computeNumbers();
-    return _numbers[i];
-  }
-
-  int* ExplicitMapping::serialize(int idproc)
-  {
-    _comm_buffer=new int[_mapping.size()*2];
-    std::vector<int> offsets(_distant_domains.size());
-    offsets[0]=0;
-    for (int i=1; i<(int)_distant_domains.size();i++)
-      offsets[i]=offsets[i-1]+_numbers[i-1];
-
-    for (int i=0; i<(int)_mapping.size(); i++)
-      {
-        int offset= offsets[_mapping[i].first];
-        _comm_buffer[offset*2]=idproc;
-        _comm_buffer[offset*2+1]=_mapping[i].second;
-        offsets[_mapping[i].first]++;
-      }
-    return _comm_buffer;
-  }
-
-  void ExplicitMapping::unserialize(int nbprocs, int* sizes,int nbtarget, int* targetrank, int* commbuffer)
-  {
-    int total_size=0;
-    for (int i=0; i< nbprocs; i++)
-      total_size+=sizes[i];
-
-    _mapping.resize(total_size);
-    _buffer_index=new int[total_size];
-    int indmap=0;
-    for (int i=0; i<nbprocs; i++)
-      for (int ielem=0; ielem<sizes[i]; ielem++)
-        {
-          _mapping[indmap].first=i;
-          _mapping[indmap].second=commbuffer[indmap*2+1];
-          _buffer_index[indmap]=commbuffer[indmap*2+1];
-          indmap++;
-        }
-    _numbers=new int [nbtarget];
-    _domains=new int [nbtarget];
-
-    int index=0;
-    for (int i=0; i<nbtarget; i++)
-      {
-        if (sizes[targetrank[i]]>0)
-          {
-            _numbers[index]=sizes[targetrank[i]];
-            _domains[index]=i;
-            index++;
-          }
-      }
-    _send_counts=new int[nbprocs];
-    for (int i=0; i<nbprocs; i++)
-      _send_counts[i]=sizes[i];
-  }
-
-  void ExplicitMapping::computeNumbers()
-  {
-    std::map <int,int> counts;
-    if (_numbers==0)
-      {
-        _numbers=new int[nbDistantDomains()];
-        _domains=new int[nbDistantDomains()];
-        for (int i=0; i<(int)_mapping.size(); i++)
-          {
-            if ( counts.find(_mapping[i].first) == counts.end())
-              counts.insert(std::make_pair(_mapping[i].first,1));
-            else
-              (counts[_mapping[i].first])++;
-          }
-        int counter=0;
-        for (std::map<int,int>::const_iterator iter=counts.begin();
-            iter!=counts.end();
-            iter++)
-          {
-            _numbers[counter]=iter->second;
-            _domains[counter]=iter->first;
-            counter++;
-          }
-      }
-  }
-
-}
diff --git a/src/ParaMEDMEM/README b/src/ParaMEDMEM/README
deleted file mode 100644 (file)
index 8a72f39..0000000
+++ /dev/null
@@ -1,194 +0,0 @@
-Tests avec CPPUNIT 
-==================
-
-L'appel a MPI_Init n'est fait qu'une seule fois pour chaque suite de test
-(voir TestMPIAccessDEC.cxx, TestMPIAccess.cxx, TestParaMEDMEM.cxx qui inclut
-tous MPIMainTest.hxx)
-Il est suivi par l'execution de toute la suite des tests regroupes
-dans les trois headers MPIAccess.hxx, MPIAccessDEC.hxx, ParaMEDMEMTest.hxx
-On a enfin un seul appel a MPI_Finalize.
-
-Si un des tests d'une suite de tests comporte une anomalie cela
-peut avoir des implications sur l'execution des tests suivants.
-
-[ABN-2015] Lancement des tests avec valgrind
-============================================
-
-Exemple avec MPICH :
-
-rm tmp_log* ; mpirun -np 3 valgrind --leak-check=full --show-reachable=yes --log-file=tmp_log%p  src/ParaMEDMEMTest/TestParaMEDMEM && cat tmp_log*
-
-[ABN-2015] Lancement des tests avec gdb
-=======================================
-
-Exemple avec MPICH :
-
-mpirun -np 2 xterm -e gdb src/ParaMEDMEMTest/TestMPIAccess
-
-
-
-
-
-*************************************************************************************************************
-*********************** ce qui suit est obsolète ??                    **************************************
-*************************************************************************************************************
-
-Lancement des tests de TestParaMEDMEM avec CPPUNIT et TotalView (option -tv) :
-==============================================================================
-
-mpirun -np 5 -ssi rpi tcp C -tv -v -x PATH=${PATH},LD_LIBRARY_PATH=${LD_LIBRARY_PATH} TestParaMEDMEM
-
-Il arrive qu'on ne puisse pas utiliser totalview par manque de
-license.
-
-
-
-Lancement des tests de TestParaMEDMEM avec CPPUNIT et Valgrind avec "memory leaks" :
-====================================================================================
-
-mpirun -np 5 -ssi rpi tcp C -v -x PATH=${PATH},LD_LIBRARY_PATH=${LD_LIBRARY_PATH} valgrind --leak-check=full TestParaMEDMEM
-
-
-Lancement des tests fonctionnels de MPI_AccessDEC avec CPPUNIT :
-================================================================
-
-mpirun -np 11 -ssi rpi tcp C -v -x PATH=${PATH},LD_LIBRARY_PATH=${LD_LIBRARY_PATH} valgrind --leak-check=full TestMPIAccessDEC
-
-
-Lancement des tests unitaires de MPI_Access avec CPPUNIT :
-==========================================================
-
-mpirun -np 3 -ssi rpi tcp C -v -x PATH=${PATH},LD_LIBRARY_PATH=${LD_LIBRARY_PATH} valgrind --leak-check=full TestMPIAccess
-
-
-TestMPIAccess/TestMPIAccessDEC/TestParaMEDMEM et gcov :
-=======================================================
-
-Les resultats sont dans les repertoires suivants de $MED_BUILD_DIR/src/ParaMEDMEM/Test :
--------------
-
-    TestParaMEDMEM-gcov/
-    TestMPIAccessDEC-gcov/
-    TestMPIAccess-gcov/
-
-Je n'y ai pas trouve d'anomalies.
-
-compilation : -fprofile-arcs -ftest-coverage
--------------
-
-$MED_BUILD_DIR/src/ParaMEDMEM/makefile.in : LIB=libparamedar.a \
--------------------------------------------     libparamed.la
-
-$MED_BUILD_DIR/src/ParaMEDMEM/Test/makefile.in : LIB = libParaMEDMEMTestar.a \
-------------------------------------------------       libParaMEDMEMTest.la
-
-execution et gcov :
--------------------
-
-Pour pouvoir traiter les .cxx de ${MED_BUILD_DIR}/src/ParaMEDMEM et de
-${MED_BUILD_DIR}/src/ParaMEDMEM/Test, on execute deux fois gcov.
-
-cd ${MED_BUILD_DIR}/src/ParaMEDMEM/Test
-
-mpirun -np 3 -ssi rpi tcp C -v -x PATH=${PATH},LD_LIBRARY_PATH=${LD_LIBRARY_PATH} TestMPIAccess
-
-gcov TestMPIAccess.cxx test_MPI_Access_Send_Recv.cxx \
-                       test_MPI_Access_Cyclic_Send_Recv.cxx \
-                       test_MPI_Access_SendRecv.cxx \
-                       test_MPI_Access_ISend_IRecv.cxx \
-                       test_MPI_Access_Cyclic_ISend_IRecv.cxx \
-                       test_MPI_Access_ISendRecv.cxx \
-                       test_MPI_Access_Probe.cxx \
-                       test_MPI_Access_IProbe.cxx \
-                       test_MPI_Access_Cancel.cxx \
-                       test_MPI_Access_Send_Recv_Length.cxx \
-                       test_MPI_Access_ISend_IRecv_Length.cxx \
-                       test_MPI_Access_ISend_IRecv_Length_1.cxx \
-                       test_MPI_Access_Time.cxx \
-                       test_MPI_Access_Time_0.cxx \
-                       test_MPI_Access_ISend_IRecv_BottleNeck.cxx \
-                       ../MPI_Access.cxx
-gcov -o ../ TestMPIAccess.cxx test_MPI_Access_Send_Recv.cxx \
-                              test_MPI_Access_Cyclic_Send_Recv.cxx \
-                              test_MPI_Access_SendRecv.cxx \
-                              test_MPI_Access_ISend_IRecv.cxx \
-                              test_MPI_Access_Cyclic_ISend_IRecv.cxx \
-                              test_MPI_Access_ISendRecv.cxx \
-                              test_MPI_Access_Probe.cxx \
-                              test_MPI_Access_IProbe.cxx \
-                              test_MPI_Access_Cancel.cxx \
-                              test_MPI_Access_Send_Recv_Length.cxx \
-                              test_MPI_Access_ISend_IRecv_Length.cxx \
-                              test_MPI_Access_ISend_IRecv_Length_1.cxx \
-                              test_MPI_Access_Time.cxx \
-                              test_MPI_Access_Time_0.cxx \
-                              test_MPI_Access_ISend_IRecv_BottleNeck.cxx \
-                              ../MPI_Access.cxx
-
-
-cd ${MED_BUILD_DIR}/src/ParaMEDMEM/Test
-mpirun -np 11 -ssi rpi tcp C -v -x PATH=${PATH},LD_LIBRARY_PATH=${LD_LIBRARY_PATH} TestMPIAccessDEC
-
-gcov TestMPIAccessDEC.cxx test_AllToAllDEC.cxx \
-                          test_AllToAllvDEC.cxx \
-                          test_AllToAllTimeDEC.cxx \
-                          test_AllToAllvTimeDEC.cxx \
-                          test_AllToAllvTimeDoubleDEC.cxx \
-                          ../TimeInterpolator.cxx \
-                          ../LinearTimeInterpolator.cxx \
-                          ../MPI_Access.cxx \
-                          ../MPI_AccessDEC.cxx
-gcov -o ../ TestMPIAccessDEC.cxx test_AllToAllDEC.cxx \
-                                 test_AllToAllvDEC.cxx \
-                                 test_AllToAllTimeDEC.cxx \
-                                 test_AllToAllvTimeDEC.cxx \
-                                 test_AllToAllvTimeDoubleDEC.cxx \
-                                 ../TimeInterpolator.cxx \
-                                 ../LinearTimeInterpolator.cxx \
-                                 ../MPI_Access.cxx \
-                                 ../MPI_AccessDEC.cxx
-
-cd ${MED_BUILD_DIR}/src/ParaMEDMEM/Test
-mpirun -np 5 -ssi rpi tcp C -v -x PATH=${PATH},LD_LIBRARY_PATH=${LD_LIBRARY_PATH} TestParaMEDMEM
-
-gcov TestParaMEDMEM.cxx ParaMEDMEMTest.cxx \
-                             ParaMEDMEMTest_MPIProcessorGroup.cxx \
-                             ParaMEDMEMTest_BlockTopology.cxx \
-                             ParaMEDMEMTest_InterpKernelDEC.cxx \
-                        ../BlockTopology.cxx \
-                        ../ComponentTopology.cxx \
-                        ../DEC.cxx \
-                        ../ElementLocator.cxx \
-                        ../InterpolationMatrix.cxx \
-                        ../InterpKernelDEC.cxx \
-                        ../MPIProcessorGroup.cxx \
-                        ../MxN_Mapping.cxx \
-                        ../ParaFIELD.cxx \
-                        ../ParaMESH.cxx \
-                        ../ParaSUPPORT.cxx \
-                        ../ProcessorGroup.cxx \
-                        ../TimeInterpolator.cxx \
-                        ../LinearTimeInterpolator.cxx \
-                        ../MPI_Access.cxx \
-                        ../MPI_AccessDEC.cxx
-
-gcov -o ../ TestParaMEDMEM.cxx ParaMEDMEMTest.cxx \
-                                    ParaMEDMEMTest_MPIProcessorGroup.cxx \
-                                    ParaMEDMEMTest_BlockTopology.cxx \
-                                    ParaMEDMEMTest_InterpKernelDEC.cxx \
-                               ../BlockTopology.cxx \
-                               ../ComponentTopology.cxx \
-                               ../DEC.cxx \
-                               ../ElementLocator.cxx \
-                               ../InterpolationMatrix.cxx \
-                               ../InterpKernelDEC.cxx \
-                               ../MPIProcessorGroup.cxx \
-                               ../MxN_Mapping.cxx \
-                               ../ParaFIELD.cxx \
-                               ../ParaMESH.cxx \
-                               ../ParaSUPPORT.cxx \
-                               ../ProcessorGroup.cxx \
-                               ../TimeInterpolator.cxx \
-                               ../LinearTimeInterpolator.cxx \
-                               ../MPI_Access.cxx \
-                               ../MPI_AccessDEC.cxx