Salome HOME
projects
/
tools
/
medcoupling.git
/ blobdiff
commit
grep
author
committer
pickaxe
?
search:
re
summary
|
shortlog
|
log
|
commit
|
commitdiff
|
tree
raw
|
inline
| side by side
Adapt for C++17
[tools/medcoupling.git]
/
src
/
ParaMEDMEMTest
/
test_MPI_Access_ISend_IRecv.cxx
diff --git
a/src/ParaMEDMEMTest/test_MPI_Access_ISend_IRecv.cxx
b/src/ParaMEDMEMTest/test_MPI_Access_ISend_IRecv.cxx
index baa3572c1fc98d34cf272b1d0fa89f17e76de857..99aaf640b626acaf81cb71aecf3181a953158266 100644
(file)
--- a/
src/ParaMEDMEMTest/test_MPI_Access_ISend_IRecv.cxx
+++ b/
src/ParaMEDMEMTest/test_MPI_Access_ISend_IRecv.cxx
@@
-1,4
+1,4
@@
-// Copyright (C) 2007-20
15
CEA/DEN, EDF R&D
+// Copyright (C) 2007-20
20
CEA/DEN, EDF R&D
//
// This library is free software; you can redistribute it and/or
// modify it under the terms of the GNU Lesser General Public
//
// This library is free software; you can redistribute it and/or
// modify it under the terms of the GNU Lesser General Public
@@
-38,7
+38,7
@@
#define ENABLE_FORCED_FAILURES
using namespace std;
#define ENABLE_FORCED_FAILURES
using namespace std;
-using namespace
ParaMEDMEM
;
+using namespace
MEDCoupling
;
void MPIAccessTest::test_MPI_Access_ISend_IRecv() {
void MPIAccessTest::test_MPI_Access_ISend_IRecv() {
@@
-52,18
+52,18
@@
void MPIAccessTest::test_MPI_Access_ISend_IRecv() {
MPI_Comm_rank(MPI_COMM_WORLD,&myrank) ;
if ( size < 2 ) {
MPI_Comm_rank(MPI_COMM_WORLD,&myrank) ;
if ( size < 2 ) {
- cerr << "test_MPI_Access_ISend_IRecv must be run
ned
with 2 procs" << endl ;
- //CPPUNIT_FAIL("test_MPI_Access_ISend_IRecv must be run
ned
with 2 procs") ;
+ cerr << "test_MPI_Access_ISend_IRecv must be run with 2 procs" << endl ;
+ //CPPUNIT_FAIL("test_MPI_Access_ISend_IRecv must be run with 2 procs") ;
return;
}
debugStream << "test_MPI_Access_ISend_IRecv" << myrank << endl ;
return;
}
debugStream << "test_MPI_Access_ISend_IRecv" << myrank << endl ;
-
ParaMEDMEM
::CommInterface interface ;
+
MEDCoupling
::CommInterface interface ;
-
ParaMEDMEM::MPIProcessorGroup* group = new ParaMEDMEM
::MPIProcessorGroup(interface) ;
+
MEDCoupling::MPIProcessorGroup* group = new MEDCoupling
::MPIProcessorGroup(interface) ;
-
ParaMEDMEM
::MPIAccess mpi_access( group ) ;
+
MEDCoupling
::MPIAccess mpi_access( group ) ;
#define maxreq 100
#define maxreq 100
@@
-102,7
+102,7
@@
void MPIAccessTest::test_MPI_Access_ISend_IRecv() {
mpi_access.test( RecvRequestId[j], flag ) ;
}
if ( flag ) {
mpi_access.test( RecvRequestId[j], flag ) ;
}
if ( flag ) {
- int
target,
source, tag, error, outcount ;
+ int source, tag, error, outcount ;
if ( myrank == 0 ) {
mpi_access.status( SendRequestId[j], target, tag, error, outcount,
true ) ;
if ( myrank == 0 ) {
mpi_access.status( SendRequestId[j], target, tag, error, outcount,
true ) ;
@@
-169,7
+169,6
@@
void MPIAccessTest::test_MPI_Access_ISend_IRecv() {
if ( myrank == 0 ) {
int sendrequests[maxreq] ;
int sendreqsize = mpi_access.sendRequestIds( target , maxreq , sendrequests ) ;
if ( myrank == 0 ) {
int sendrequests[maxreq] ;
int sendreqsize = mpi_access.sendRequestIds( target , maxreq , sendrequests ) ;
- int i ;
if ( sendreqsize != 0 ) {
ostringstream strstream ;
strstream << "=========================================================" << endl
if ( sendreqsize != 0 ) {
ostringstream strstream ;
strstream << "=========================================================" << endl