X-Git-Url: http://git.salome-platform.org/gitweb/?a=blobdiff_plain;f=src%2FParaMEDMEMTest%2Ftest_MPI_Access_SendRecv.cxx;h=9fcb752da4819fdfb4eb9c404b2283925d3c4261;hb=c8d33262ed6d2041da8512b02f8c8b4ffadf733d;hp=49dec84e3bf99053f5f686be0d8c9034db58fa4d;hpb=75943f980f7b908052ef03c2c0154508f4b0a039;p=tools%2Fmedcoupling.git diff --git a/src/ParaMEDMEMTest/test_MPI_Access_SendRecv.cxx b/src/ParaMEDMEMTest/test_MPI_Access_SendRecv.cxx index 49dec84e3..9fcb752da 100644 --- a/src/ParaMEDMEMTest/test_MPI_Access_SendRecv.cxx +++ b/src/ParaMEDMEMTest/test_MPI_Access_SendRecv.cxx @@ -1,4 +1,4 @@ -// Copyright (C) 2007-2014 CEA/DEN, EDF R&D +// Copyright (C) 2007-2015 CEA/DEN, EDF R&D // // This library is free software; you can redistribute it and/or // modify it under the terms of the GNU Lesser General Public @@ -38,11 +38,11 @@ #define ENABLE_FORCED_FAILURES using namespace std; -using namespace ParaMEDMEM; +using namespace MEDCoupling; void MPIAccessTest::test_MPI_Access_SendRecv() { - cout << "MPIAccessTest::test_MPI_Access_SendRecv" << endl ; + debugStream << "MPIAccessTest::test_MPI_Access_SendRecv" << endl ; // MPI_Init(&argc, &argv) ; @@ -52,17 +52,18 @@ void MPIAccessTest::test_MPI_Access_SendRecv() { MPI_Comm_rank(MPI_COMM_WORLD,&myrank) ; if ( size < 2 ) { - cout << "MPIAccessTest::test_MPI_Access_SendRecv must be runned with 2 procs" << endl ; - CPPUNIT_FAIL("test_MPI_Access_SendRecv must be runned with 2 procs") ; + cerr << "MPIAccessTest::test_MPI_Access_SendRecv must be runned with 2 procs" << endl ; + //CPPUNIT_FAIL("test_MPI_Access_SendRecv must be runned with 2 procs") ; + return; } - cout << "MPIAccessTest::test_MPI_Access_SendRecv" << myrank << endl ; + debugStream << "MPIAccessTest::test_MPI_Access_SendRecv" << myrank << endl ; - ParaMEDMEM::CommInterface interface ; + MEDCoupling::CommInterface interface ; - ParaMEDMEM::MPIProcessorGroup* group = new ParaMEDMEM::MPIProcessorGroup(interface) ; + MEDCoupling::MPIProcessorGroup* group = new MEDCoupling::MPIProcessorGroup(interface) ; - ParaMEDMEM::MPIAccess mpi_access( group ) ; + MEDCoupling::MPIAccess mpi_access( group ) ; if ( myrank >= 2 ) { mpi_access.barrier() ; @@ -90,7 +91,7 @@ void MPIAccessTest::test_MPI_Access_SendRecv() { // outcount = mpi_access.MPIOutCount( recvRequestId[i] ) ; outcount = 1 ; } - cout << "test" << myrank << " Send sendRequestId " << sendRequestId[i] + debugStream << "test" << myrank << " Send sendRequestId " << sendRequestId[i] << " tag " << mpi_access.sendMPITag(target) << " recvRequestId " << recvRequestId[i] << " tag " << mpi_access.recvMPITag(target) @@ -103,13 +104,13 @@ void MPIAccessTest::test_MPI_Access_SendRecv() { << " recvbuf " << recvbuf << " KO" << "===========================================================" << endl ; - cout << strstream.str() << endl ; + debugStream << strstream.str() << endl ; CPPUNIT_FAIL( strstream.str() ) ; } char msgerr[MPI_MAX_ERROR_STRING] ; int lenerr ; mpi_access.errorString(sts, msgerr, &lenerr) ; - cout << "test" << myrank << " lenerr " << lenerr << " " + debugStream << "test" << myrank << " lenerr " << lenerr << " " << msgerr << endl ; if ( sts != MPI_SUCCESS ) { @@ -118,10 +119,10 @@ void MPIAccessTest::test_MPI_Access_SendRecv() { << "test" << myrank << " KO" << "===========================================================" << endl ; - cout << strstream.str() << endl ; + debugStream << strstream.str() << endl ; CPPUNIT_FAIL( strstream.str() ) ; } - mpi_access.check() ; + if(MPI_ACCESS_VERBOSE) mpi_access.check() ; } int flag ; @@ -129,7 +130,7 @@ void MPIAccessTest::test_MPI_Access_SendRecv() { if ( !flag ) { ostringstream strstream ; strstream << "test" << myrank << " flag " << flag << " KO" << endl ; - cout << strstream.str() << endl ; + debugStream << strstream.str() << endl ; CPPUNIT_FAIL( strstream.str() ) ; } mpi_access.waitAll(10,sendRequestId) ; @@ -137,11 +138,11 @@ void MPIAccessTest::test_MPI_Access_SendRecv() { if ( !flag ) { ostringstream strstream ; strstream << "test" << myrank << " flag " << flag << " KO" << endl ; - cout << strstream.str() << endl ; + debugStream << strstream.str() << endl ; CPPUNIT_FAIL( strstream.str() ) ; } mpi_access.waitAll(10,recvRequestId) ; - mpi_access.check() ; + if(MPI_ACCESS_VERBOSE) mpi_access.check() ; int sendrequests[10] ; int sendreqsize = mpi_access.sendRequestIds( target , 10 , sendrequests ) ; @@ -150,7 +151,7 @@ void MPIAccessTest::test_MPI_Access_SendRecv() { strstream << "=========================================================" << endl << "test" << myrank << " sendreqsize " << sendreqsize << " KO" << endl << "=========================================================" << endl ; - cout << strstream.str() << endl ; + debugStream << strstream.str() << endl ; CPPUNIT_FAIL( strstream.str() ) ; } int recvrequests[10] ; @@ -160,7 +161,7 @@ void MPIAccessTest::test_MPI_Access_SendRecv() { strstream << "=========================================================" << endl << "test" << myrank << " recvreqsize " << recvreqsize << " KO" << endl << "=========================================================" << endl ; - cout << strstream.str() << endl ; + debugStream << strstream.str() << endl ; CPPUNIT_FAIL( strstream.str() ) ; } @@ -170,7 +171,7 @@ void MPIAccessTest::test_MPI_Access_SendRecv() { // MPI_Finalize(); - cout << "test" << myrank << " OK" << endl ; + debugStream << "test" << myrank << " OK" << endl ; return ; }