| [0b990d] | 1 | // | 
|---|
|  | 2 | // memmtmpi.cc | 
|---|
|  | 3 | // based on memmpi.cc | 
|---|
|  | 4 | // | 
|---|
|  | 5 | // Copyright (C) 1996 Limit Point Systems, Inc. | 
|---|
|  | 6 | // | 
|---|
|  | 7 | // Author: Curtis Janssen <cljanss@limitpt.com> | 
|---|
|  | 8 | // Maintainer: LPS | 
|---|
|  | 9 | // | 
|---|
|  | 10 | // This file is part of the SC Toolkit. | 
|---|
|  | 11 | // | 
|---|
|  | 12 | // The SC Toolkit is free software; you can redistribute it and/or modify | 
|---|
|  | 13 | // it under the terms of the GNU Library General Public License as published by | 
|---|
|  | 14 | // the Free Software Foundation; either version 2, or (at your option) | 
|---|
|  | 15 | // any later version. | 
|---|
|  | 16 | // | 
|---|
|  | 17 | // The SC Toolkit is distributed in the hope that it will be useful, | 
|---|
|  | 18 | // but WITHOUT ANY WARRANTY; without even the implied warranty of | 
|---|
|  | 19 | // MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the | 
|---|
|  | 20 | // GNU Library General Public License for more details. | 
|---|
|  | 21 | // | 
|---|
|  | 22 | // You should have received a copy of the GNU Library General Public License | 
|---|
|  | 23 | // along with the SC Toolkit; see the file COPYING.LIB.  If not, write to | 
|---|
|  | 24 | // the Free Software Foundation, 675 Mass Ave, Cambridge, MA 02139, USA. | 
|---|
|  | 25 | // | 
|---|
|  | 26 | // The U.S. Government is granted a limited license as per AL 91-7. | 
|---|
|  | 27 | // | 
|---|
|  | 28 |  | 
|---|
|  | 29 | #ifndef _util_group_memmtmpi_cc | 
|---|
|  | 30 | #define _util_group_memmtmpi_cc | 
|---|
|  | 31 |  | 
|---|
|  | 32 | #ifdef __GNUC__ | 
|---|
|  | 33 | #pragma implementation | 
|---|
|  | 34 | #endif | 
|---|
|  | 35 |  | 
|---|
|  | 36 | #include <assert.h> | 
|---|
|  | 37 |  | 
|---|
|  | 38 | #include <util/misc/formio.h> | 
|---|
|  | 39 | #include <util/group/messmpi.h> | 
|---|
|  | 40 | #include <util/group/memmtmpi.h> | 
|---|
|  | 41 |  | 
|---|
|  | 42 | #define MPICH_SKIP_MPICXX | 
|---|
|  | 43 | #include <mpi.h> | 
|---|
|  | 44 |  | 
|---|
|  | 45 | using namespace std; | 
|---|
|  | 46 |  | 
|---|
|  | 47 | // Define this to use immediate mode.  This was added added to work | 
|---|
|  | 48 | // around bugs in non-immediate mode optimizations in an MPI impl. | 
|---|
|  | 49 | #undef USE_IMMEDIATE_MODE | 
|---|
|  | 50 |  | 
|---|
|  | 51 | namespace sc { | 
|---|
|  | 52 |  | 
|---|
|  | 53 | static const int dbufsize = 32768; | 
|---|
|  | 54 |  | 
|---|
|  | 55 | /////////////////////////////////////////////////////////////////////// | 
|---|
|  | 56 | // The MTMPIThread class | 
|---|
|  | 57 |  | 
|---|
|  | 58 | class MTMPIThread: public Thread { | 
|---|
|  | 59 | private: | 
|---|
|  | 60 | MTMPIMemoryGrp *mem_; | 
|---|
|  | 61 | int req_tag_; | 
|---|
|  | 62 | int tag_; | 
|---|
|  | 63 | unsigned int nreq_recd_; | 
|---|
|  | 64 | double chunk[dbufsize]; | 
|---|
|  | 65 | public: | 
|---|
|  | 66 | MTMPIThread(MTMPIMemoryGrp *, int reqtype, int tag); | 
|---|
|  | 67 | void run(); | 
|---|
|  | 68 | int run_one(); | 
|---|
|  | 69 | unsigned int nreq_recd() { return nreq_recd_; } | 
|---|
|  | 70 | void set_nreq_recd(unsigned int val) { nreq_recd_ = val; } | 
|---|
|  | 71 | }; | 
|---|
|  | 72 |  | 
|---|
|  | 73 | MTMPIThread::MTMPIThread(MTMPIMemoryGrp *mem, | 
|---|
|  | 74 | int reqtype, int tag) | 
|---|
|  | 75 | { | 
|---|
|  | 76 | mem_ = mem; | 
|---|
|  | 77 | req_tag_ = reqtype; | 
|---|
|  | 78 | tag_ = tag; | 
|---|
|  | 79 | nreq_recd_ = 0; | 
|---|
|  | 80 | } | 
|---|
|  | 81 |  | 
|---|
|  | 82 | void | 
|---|
|  | 83 | MTMPIThread::run() | 
|---|
|  | 84 | { | 
|---|
|  | 85 | while(run_one()); | 
|---|
|  | 86 | } | 
|---|
|  | 87 |  | 
|---|
|  | 88 | int | 
|---|
|  | 89 | MTMPIThread::run_one() | 
|---|
|  | 90 | { | 
|---|
|  | 91 | int i; | 
|---|
|  | 92 | int dsize; | 
|---|
|  | 93 | int dremain; | 
|---|
|  | 94 | int doffset; | 
|---|
|  | 95 | long l; | 
|---|
|  | 96 | MemoryDataRequest req; | 
|---|
|  | 97 | MPI_Status status; | 
|---|
|  | 98 | #ifndef USE_IMMEDIATE_MODE | 
|---|
|  | 99 | MPI_Recv(req.data(),req.nbytes(),MPI_BYTE,MPI_ANY_SOURCE, | 
|---|
|  | 100 | req_tag_,mem_->comm_comm_,&status); | 
|---|
|  | 101 | #else | 
|---|
|  | 102 | MPI_Request mpireq; | 
|---|
|  | 103 | MPI_Irecv(req.data(),req.nbytes(),MPI_BYTE,MPI_ANY_SOURCE, | 
|---|
|  | 104 | req_tag_,mem_->comm_comm_,&mpireq); | 
|---|
|  | 105 | MPI_Wait(&mpireq,&status); | 
|---|
|  | 106 | #endif // USE_IMMEDIATE_MODE | 
|---|
|  | 107 | int rtag = req.serial_number(); | 
|---|
|  | 108 | if (mem_->debug_) { | 
|---|
|  | 109 | mem_->print_lock_->lock(); | 
|---|
|  | 110 | req.print("RECV",mem_->hout); | 
|---|
|  | 111 | mem_->print_lock_->unlock(); | 
|---|
|  | 112 | } | 
|---|
|  | 113 | if (req.touches_data()) { | 
|---|
|  | 114 | assert(req.size() >= 0); | 
|---|
|  | 115 | if (req.offset()+req.size() > mem_->localsize()) { | 
|---|
|  | 116 | mem_->print_lock_->lock(); | 
|---|
|  | 117 | req.print("BAD RECV"); | 
|---|
|  | 118 | ExEnv::outn() << "mem_->localsize() = " << mem_->localsize() << endl; | 
|---|
|  | 119 | mem_->print_lock_->lock(); | 
|---|
|  | 120 | } | 
|---|
|  | 121 | assert(req.offset()+req.size() <= mem_->localsize()); | 
|---|
|  | 122 | } | 
|---|
|  | 123 | switch (req.request()) { | 
|---|
|  | 124 | case MemoryDataRequest::Deactivate: | 
|---|
|  | 125 | return 0; | 
|---|
|  | 126 | case MemoryDataRequest::Retrieve: | 
|---|
|  | 127 | nreq_recd_++; | 
|---|
|  | 128 | if (req.lock()) | 
|---|
|  | 129 | mem_->obtain_local_lock(req.offset(), req.offset()+req.size()); | 
|---|
|  | 130 | MPI_Send(&mem_->data_[req.offset()],req.size(),MPI_BYTE, | 
|---|
|  | 131 | req.node(),rtag,mem_->comp_comm_); | 
|---|
|  | 132 | break; | 
|---|
|  | 133 | case MemoryDataRequest::Replace: | 
|---|
|  | 134 | nreq_recd_++; | 
|---|
|  | 135 | // May be able to get rid of this MPI_Send - MLL | 
|---|
|  | 136 | MPI_Send(&tag_,1,MPI_INT,req.node(),rtag,mem_->comp_comm_); | 
|---|
|  | 137 | MPI_Recv(&mem_->data_[req.offset()],req.size(),MPI_BYTE, | 
|---|
|  | 138 | req.node(),tag_,mem_->comm_comm_,&status); | 
|---|
|  | 139 | if (req.lock()) | 
|---|
|  | 140 | mem_->release_local_lock(req.offset(), req.offset()+req.size()); | 
|---|
|  | 141 | break; | 
|---|
|  | 142 | case MemoryDataRequest::DoubleSum: | 
|---|
|  | 143 | nreq_recd_++; | 
|---|
|  | 144 | //      MPI_Send(&tag_,1,MPI_INT,req.node(),rtag,mem_->comm_); | 
|---|
|  | 145 | dsize = req.size()/sizeof(double); | 
|---|
|  | 146 | dremain = dsize; | 
|---|
|  | 147 | doffset = req.offset()/sizeof(double); | 
|---|
|  | 148 | mem_->obtain_local_lock(req.offset(), req.offset()+req.size()); | 
|---|
|  | 149 | while(dremain>0) { | 
|---|
|  | 150 | int dchunksize = dbufsize; | 
|---|
|  | 151 | if (dremain < dchunksize) dchunksize = dremain; | 
|---|
|  | 152 | #ifndef USE_IMMEDIATE_MODE | 
|---|
|  | 153 | MPI_Recv(chunk,dchunksize,MPI_DOUBLE, | 
|---|
|  | 154 | req.node(),rtag ,mem_->comm_comm_,&status); | 
|---|
|  | 155 | #else | 
|---|
|  | 156 | MPI_Request mpireq; | 
|---|
|  | 157 | MPI_Irecv(chunk,dchunksize,MPI_DOUBLE, | 
|---|
|  | 158 | req.node(),rtag ,mem_->comm_comm_,&mpireq); | 
|---|
|  | 159 | MPI_Wait(&mpireq,&status); | 
|---|
|  | 160 | #endif // USE_IMMEDIATE_MODE | 
|---|
|  | 161 | double *source_data = &((double*)mem_->data_)[doffset]; | 
|---|
|  | 162 | for (i=0; i<dchunksize; i++) { | 
|---|
|  | 163 | source_data[i] += chunk[i]; | 
|---|
|  | 164 | } | 
|---|
|  | 165 | dremain -= dchunksize; | 
|---|
|  | 166 | doffset += dchunksize; | 
|---|
|  | 167 | } | 
|---|
|  | 168 | mem_->release_local_lock(req.offset(), req.offset()+req.size()); | 
|---|
|  | 169 | break; | 
|---|
|  | 170 | default: | 
|---|
|  | 171 | mem_->print_lock_->lock(); | 
|---|
|  | 172 | ExEnv::outn() << "MTMPIThread: bad memory data request" << endl; | 
|---|
|  | 173 | mem_->print_lock_->unlock(); | 
|---|
|  | 174 | abort(); | 
|---|
|  | 175 | } | 
|---|
|  | 176 | return 1; | 
|---|
|  | 177 | } | 
|---|
|  | 178 |  | 
|---|
|  | 179 | /////////////////////////////////////////////////////////////////////// | 
|---|
|  | 180 | // The MTMPIMemoryGrp class | 
|---|
|  | 181 |  | 
|---|
|  | 182 | static ClassDesc MTMPIMemoryGrp_cd( | 
|---|
|  | 183 | typeid(MTMPIMemoryGrp),"MTMPIMemoryGrp",1,"public ActiveMsgMemoryGrp", | 
|---|
|  | 184 | 0, create<MTMPIMemoryGrp>, 0); | 
|---|
|  | 185 |  | 
|---|
|  | 186 | MTMPIMemoryGrp::MTMPIMemoryGrp(const Ref<MessageGrp>& msg, | 
|---|
|  | 187 | const Ref<ThreadGrp>& th, | 
|---|
|  | 188 | MPI_Comm comm): | 
|---|
|  | 189 | ActiveMsgMemoryGrp(msg) | 
|---|
|  | 190 | { | 
|---|
|  | 191 | if (debug_) ExEnv::outn() << "MTMPIMemoryGrp CTOR entered" << endl; | 
|---|
|  | 192 |  | 
|---|
|  | 193 | th_ = th; | 
|---|
|  | 194 |  | 
|---|
|  | 195 | init_mtmpimg(comm,th_->nthread()); | 
|---|
|  | 196 | } | 
|---|
|  | 197 |  | 
|---|
|  | 198 | MTMPIMemoryGrp::MTMPIMemoryGrp(const Ref<KeyVal>& keyval): | 
|---|
|  | 199 | ActiveMsgMemoryGrp(keyval) | 
|---|
|  | 200 | { | 
|---|
|  | 201 | if (debug_) ExEnv::outn() << "MTMPIMemoryGrp keyval CTOR entered" << endl; | 
|---|
|  | 202 |  | 
|---|
|  | 203 | th_ = ThreadGrp::get_default_threadgrp(); | 
|---|
|  | 204 |  | 
|---|
|  | 205 | KeyValValueint nthreaddef(th_->nthread()); | 
|---|
|  | 206 | int nthread = keyval->intvalue("num_threads",nthreaddef); | 
|---|
|  | 207 | ExEnv::out0() << indent << "MTMPIMemoryGrp: num_threads = " << nthread << endl; | 
|---|
|  | 208 |  | 
|---|
|  | 209 | init_mtmpimg(MPI_COMM_WORLD,nthread); | 
|---|
|  | 210 | } | 
|---|
|  | 211 |  | 
|---|
|  | 212 | MTMPIMemoryGrp::~MTMPIMemoryGrp() | 
|---|
|  | 213 | { | 
|---|
|  | 214 | deactivate(); | 
|---|
|  | 215 | for (int i=0; i<th_->nthread()-1; i++) { | 
|---|
|  | 216 | delete thread_[i]; | 
|---|
|  | 217 | } | 
|---|
|  | 218 | delete[] thread_; | 
|---|
|  | 219 | delete[] nreq_sent_; | 
|---|
|  | 220 | delete[] nreq_sent_buf_; | 
|---|
|  | 221 | } | 
|---|
|  | 222 |  | 
|---|
|  | 223 | void | 
|---|
|  | 224 | MTMPIMemoryGrp::init_mtmpimg(MPI_Comm comm, int nthread) | 
|---|
|  | 225 | { | 
|---|
|  | 226 | int i; | 
|---|
|  | 227 | active_ = 0; | 
|---|
|  | 228 |  | 
|---|
|  | 229 | if (nthread < 2) nthread = 2; | 
|---|
|  | 230 | th_ = th_->clone(nthread); | 
|---|
|  | 231 | nthread = th_->nthread(); | 
|---|
|  | 232 |  | 
|---|
|  | 233 | if (debug_) { | 
|---|
|  | 234 | char name[256]; | 
|---|
|  | 235 | sprintf(name, "mpqc.hand.%d", me()); | 
|---|
|  | 236 | hout.open(name); | 
|---|
|  | 237 | sprintf(name, "mpqc.main.%d", me()); | 
|---|
|  | 238 | mout.open(name); | 
|---|
|  | 239 | } | 
|---|
|  | 240 |  | 
|---|
|  | 241 | MPI_Comm_dup(comm, &comp_comm_); | 
|---|
|  | 242 | MPI_Comm_dup(comm, &comm_comm_); | 
|---|
|  | 243 |  | 
|---|
|  | 244 | MPI_Errhandler_set(comp_comm_, MPI_ERRORS_ARE_FATAL); | 
|---|
|  | 245 | MPI_Errhandler_set(comm_comm_, MPI_ERRORS_ARE_FATAL); | 
|---|
|  | 246 |  | 
|---|
|  | 247 | serial_ = 0; | 
|---|
|  | 248 | req_tag_ = 15001; | 
|---|
|  | 249 |  | 
|---|
|  | 250 | serial_lock_ = th_->new_lock(); | 
|---|
|  | 251 |  | 
|---|
|  | 252 | thread_ = new MTMPIThread*[nthread-1]; | 
|---|
|  | 253 | th_->add_thread(0,0); | 
|---|
|  | 254 | for (i=1; i<nthread; i++) { | 
|---|
|  | 255 | thread_[i-1] = new MTMPIThread(this,req_tag_,req_tag_ + i); | 
|---|
|  | 256 | th_->add_thread(i,thread_[i-1]); | 
|---|
|  | 257 | } | 
|---|
|  | 258 | print_lock_ = th_->new_lock(); | 
|---|
|  | 259 |  | 
|---|
|  | 260 | nreq_sent_ = new unsigned int[n()]; | 
|---|
|  | 261 | memset(nreq_sent_, 0, sizeof(unsigned int)*n()); | 
|---|
|  | 262 | nreq_sent_buf_ = new unsigned int[n()]; | 
|---|
|  | 263 | } | 
|---|
|  | 264 |  | 
|---|
|  | 265 | int | 
|---|
|  | 266 | MTMPIMemoryGrp::serial(int node) | 
|---|
|  | 267 | { | 
|---|
|  | 268 | serial_lock_->lock(); | 
|---|
|  | 269 | nreq_sent_[node]++; | 
|---|
|  | 270 | int r = serial_; | 
|---|
|  | 271 | serial_++; | 
|---|
|  | 272 | if (serial_ == req_tag_) serial_ = 0; | 
|---|
|  | 273 | serial_lock_->unlock(); | 
|---|
|  | 274 | return r; | 
|---|
|  | 275 | } | 
|---|
|  | 276 |  | 
|---|
|  | 277 | void | 
|---|
|  | 278 | MTMPIMemoryGrp::retrieve_data(void *data, int node, int offset, int size, | 
|---|
|  | 279 | int lock) | 
|---|
|  | 280 | { | 
|---|
|  | 281 | MemoryDataRequest req(MemoryDataRequest::Retrieve,me(),offset,size,lock, | 
|---|
|  | 282 | serial(node)); | 
|---|
|  | 283 | int tag = req.serial_number(); | 
|---|
|  | 284 |  | 
|---|
|  | 285 | // send the request | 
|---|
|  | 286 | if (debug_) { | 
|---|
|  | 287 | print_lock_->lock(); | 
|---|
|  | 288 | req.print("SEND",mout); | 
|---|
|  | 289 | print_lock_->unlock(); | 
|---|
|  | 290 | } | 
|---|
|  | 291 | MPI_Send(req.data(),req.nbytes(),MPI_BYTE,node,req_tag_,comm_comm_); | 
|---|
|  | 292 |  | 
|---|
|  | 293 | // receive the data | 
|---|
|  | 294 | MPI_Status status; | 
|---|
|  | 295 | MPI_Recv(data,size,MPI_BYTE,node,tag,comp_comm_,&status); | 
|---|
|  | 296 | } | 
|---|
|  | 297 |  | 
|---|
|  | 298 | void | 
|---|
|  | 299 | MTMPIMemoryGrp::replace_data(void *data, int node, int offset, int size, | 
|---|
|  | 300 | int unlock) | 
|---|
|  | 301 | { | 
|---|
|  | 302 | MemoryDataRequest req(MemoryDataRequest::Replace,me(),offset,size,unlock, | 
|---|
|  | 303 | serial(node)); | 
|---|
|  | 304 | int tag = req.serial_number(); | 
|---|
|  | 305 |  | 
|---|
|  | 306 | if (debug_) { | 
|---|
|  | 307 | print_lock_->lock(); | 
|---|
|  | 308 | req.print("SEND",mout); | 
|---|
|  | 309 | print_lock_->unlock(); | 
|---|
|  | 310 | } | 
|---|
|  | 311 | MPI_Send(req.data(),req.nbytes(),MPI_BYTE,node,req_tag_,comm_comm_); | 
|---|
|  | 312 |  | 
|---|
|  | 313 | // wait for the go ahead message | 
|---|
|  | 314 | MPI_Status status; | 
|---|
|  | 315 | int rtag; | 
|---|
|  | 316 | MPI_Recv(&rtag,1,MPI_INT,node,tag,comp_comm_,&status); | 
|---|
|  | 317 |  | 
|---|
|  | 318 | // send the data | 
|---|
|  | 319 | MPI_Send(data,size,MPI_BYTE,node,rtag,comm_comm_); | 
|---|
|  | 320 | } | 
|---|
|  | 321 |  | 
|---|
|  | 322 | void | 
|---|
|  | 323 | MTMPIMemoryGrp::sum_data(double *data, int node, int offset, int size) | 
|---|
|  | 324 | { | 
|---|
|  | 325 | MemoryDataRequest req(MemoryDataRequest::DoubleSum,me(),offset,size, | 
|---|
|  | 326 | 0, serial(node)); | 
|---|
|  | 327 | int tag = req.serial_number(); | 
|---|
|  | 328 |  | 
|---|
|  | 329 | // send the request | 
|---|
|  | 330 | if (debug_) { | 
|---|
|  | 331 | print_lock_->lock(); | 
|---|
|  | 332 | req.print("SEND",mout); | 
|---|
|  | 333 | print_lock_->unlock(); | 
|---|
|  | 334 | } | 
|---|
|  | 335 | #ifndef USE_IMMEDIATE_MODE | 
|---|
|  | 336 | MPI_Send(req.data(),req.nbytes(),MPI_BYTE,node,req_tag_,comm_comm_); | 
|---|
|  | 337 | #else | 
|---|
|  | 338 | MPI_Status status; | 
|---|
|  | 339 | MPI_Request mpireq; | 
|---|
|  | 340 | MPI_Isend(req.data(),req.nbytes(),MPI_BYTE,node,req_tag_,comm_comm_,&mpireq); | 
|---|
|  | 341 | MPI_Wait(&mpireq,&status); | 
|---|
|  | 342 | #endif // USE_IMMEDIATE_MODE | 
|---|
|  | 343 |  | 
|---|
|  | 344 | // wait for the go ahead message | 
|---|
|  | 345 | //  MPI_Status status; | 
|---|
|  | 346 | //  int rtag; | 
|---|
|  | 347 | //  MPI_Recv(&rtag,1,MPI_INT,node,tag,comm_,&status); | 
|---|
|  | 348 |  | 
|---|
|  | 349 | int dsize = size/sizeof(double); | 
|---|
|  | 350 | int dremain = dsize; | 
|---|
|  | 351 | int dcurrent = 0; | 
|---|
|  | 352 | while(dremain>0) { | 
|---|
|  | 353 | int dchunksize = dbufsize; | 
|---|
|  | 354 | if (dremain < dchunksize) dchunksize = dremain; | 
|---|
|  | 355 | // send the data | 
|---|
|  | 356 | #ifndef USE_IMMEDIATE_MODE | 
|---|
|  | 357 | MPI_Send(&data[dcurrent],dchunksize,MPI_DOUBLE, | 
|---|
|  | 358 | node,tag,comm_comm_); | 
|---|
|  | 359 | #else | 
|---|
|  | 360 | MPI_Request mpireq; | 
|---|
|  | 361 | MPI_Isend(&data[dcurrent],dchunksize,MPI_DOUBLE, | 
|---|
|  | 362 | node,tag,comm_comm_,&mpireq); | 
|---|
|  | 363 | MPI_Wait(&mpireq,&status); | 
|---|
|  | 364 | #endif // USE_IMMEDIATE_MODE | 
|---|
|  | 365 | dcurrent += dchunksize; | 
|---|
|  | 366 | dremain -= dchunksize; | 
|---|
|  | 367 | } | 
|---|
|  | 368 | } | 
|---|
|  | 369 |  | 
|---|
|  | 370 | void | 
|---|
|  | 371 | MTMPIMemoryGrp::activate() | 
|---|
|  | 372 | { | 
|---|
|  | 373 | // Only remote requests require the handler.  There are only remote | 
|---|
|  | 374 | // requests if there is more than one node. | 
|---|
|  | 375 | if (n() == 1) return; | 
|---|
|  | 376 |  | 
|---|
|  | 377 | if (th_->nthread() < 2) { | 
|---|
|  | 378 | ExEnv::outn() << "MTMPIMemoryGrp didn't get enough threads" << endl; | 
|---|
|  | 379 | abort(); | 
|---|
|  | 380 | } | 
|---|
|  | 381 |  | 
|---|
|  | 382 | if (active_) return; | 
|---|
|  | 383 | active_ = 1; | 
|---|
|  | 384 |  | 
|---|
|  | 385 | th_->start_threads(); | 
|---|
|  | 386 | } | 
|---|
|  | 387 |  | 
|---|
|  | 388 | void | 
|---|
|  | 389 | MTMPIMemoryGrp::deactivate() | 
|---|
|  | 390 | { | 
|---|
|  | 391 | if (!active_) return; | 
|---|
|  | 392 | active_ = 0; | 
|---|
|  | 393 |  | 
|---|
|  | 394 | // send a shutdown message | 
|---|
|  | 395 | MemoryDataRequest req(MemoryDataRequest::Deactivate); | 
|---|
|  | 396 | if (debug_) { | 
|---|
|  | 397 | print_lock_->lock(); | 
|---|
|  | 398 | req.print("SEND",mout); | 
|---|
|  | 399 | print_lock_->unlock(); | 
|---|
|  | 400 | } | 
|---|
|  | 401 | for (int i=1; i<th_->nthread(); i++) { | 
|---|
|  | 402 | #ifndef USE_IMMEDIATE_MODE | 
|---|
|  | 403 | MPI_Send(req.data(),req.nbytes(),MPI_BYTE,me(),req_tag_,comm_comm_); | 
|---|
|  | 404 | #else | 
|---|
|  | 405 | MPI_Request mpireq; | 
|---|
|  | 406 | MPI_Status status; | 
|---|
|  | 407 | MPI_Isend(req.data(),req.nbytes(),MPI_BYTE,me(),req_tag_,comm_comm_,&mpireq); | 
|---|
|  | 408 | MPI_Wait(&mpireq,&status); | 
|---|
|  | 409 | #endif // USE_IMMEDIATE_MODE | 
|---|
|  | 410 | } | 
|---|
|  | 411 |  | 
|---|
|  | 412 | // wait on the thread to shutdown | 
|---|
|  | 413 | th_->wait_threads(); | 
|---|
|  | 414 | } | 
|---|
|  | 415 |  | 
|---|
|  | 416 | void | 
|---|
|  | 417 | MTMPIMemoryGrp::sync() | 
|---|
|  | 418 | { | 
|---|
|  | 419 | if (active_) { | 
|---|
|  | 420 | MPI_Allreduce(nreq_sent_, nreq_sent_buf_, | 
|---|
|  | 421 | n(), MPI_UNSIGNED, MPI_SUM, comm_comm_); | 
|---|
|  | 422 | deactivate(); | 
|---|
|  | 423 | unsigned int nreq_recd = 0; | 
|---|
|  | 424 | for (int i=0; i<th_->nthread()-1; i++) { | 
|---|
|  | 425 | nreq_recd += thread_[i]->nreq_recd(); | 
|---|
|  | 426 | thread_[i]->set_nreq_recd(0); | 
|---|
|  | 427 | } | 
|---|
|  | 428 | int n_outstanding = nreq_sent_buf_[me()] - nreq_recd; | 
|---|
|  | 429 | for (int i=0; i<n_outstanding; i++) { | 
|---|
|  | 430 | thread_[0]->run_one(); | 
|---|
|  | 431 | } | 
|---|
|  | 432 | memset(nreq_sent_, 0, sizeof(unsigned int)*n()); | 
|---|
|  | 433 | // Make sure processing of all outstanding requests is finished | 
|---|
|  | 434 | // before starting the next phase. | 
|---|
|  | 435 | MPI_Barrier(comm_comm_); | 
|---|
|  | 436 | activate(); | 
|---|
|  | 437 | } | 
|---|
|  | 438 | else { | 
|---|
|  | 439 | MPI_Barrier(comm_comm_); | 
|---|
|  | 440 | } | 
|---|
|  | 441 | } | 
|---|
|  | 442 |  | 
|---|
|  | 443 | #endif | 
|---|
|  | 444 |  | 
|---|
|  | 445 | ///////////////////////////////////////////////////////////////////////////// | 
|---|
|  | 446 |  | 
|---|
|  | 447 | } | 
|---|
|  | 448 |  | 
|---|
|  | 449 | // Local Variables: | 
|---|
|  | 450 | // mode: c++ | 
|---|
|  | 451 | // c-file-style: "CLJ" | 
|---|
|  | 452 | // End: | 
|---|