PSCF v1.2
MpiSendRecv.h
Go to the documentation of this file.
1#ifdef UTIL_MPI
2#ifndef UTIL_MPI_SEND_RECV_H
3#define UTIL_MPI_SEND_RECV_H
4
5/*
6* Util Package - C++ Utilities for Scientific Computation
7*
8* Copyright 2010 - 2017, The Regents of the University of Minnesota
9* Distributed under the terms of the GNU General Public License.
10*/
11
76#include <util/global.h>
77
78#include <util/mpi/MpiTraits.h>
79#include <util/containers/DArray.h>
80#include <util/containers/DMatrix.h>
81
82namespace Util
83{
84
85 // Scalar parameters
86
98 template <typename T>
99 void send(MPI::Comm& comm, T& data, int dest, int tag)
100 {
102 UTIL_THROW("No committed MPI type in send<T>");
103 comm.Send(&data, 1, MpiTraits<T>::type, dest, tag);
104 }
105
117 template <typename T>
118 void recv(MPI::Comm& comm, T& data, int source, int tag)
119 {
121 UTIL_THROW("No committed MPI type in recv<T>");
122 comm.Recv(&data, 1, MpiTraits<T>::type, source, tag);
123 }
124
135 template <typename T>
136 void bcast(MPI::Intracomm& comm, T& data, int root)
137 {
139 UTIL_THROW("No committed MPI type in bcast<T>");
140 comm.Bcast(&data, 1, MpiTraits<T>::type, root);
141 }
142
143 // C Array partial specializations
144
157 template <typename T>
158 void send(MPI::Comm& comm, T* array, int count, int dest, int tag)
159 {
161 comm.Send(array, count, MpiTraits<T>::type, dest, tag);
162 } else {
163 // Try send<T> by element, in case of explicit specialization.
164 // If there is no specialization or type, send<T> throws.
165 for (int i = 0; i < count; ++i) {
166 send<T>(comm, array[i], dest, tag);
167 }
168 }
169 }
170
183 template <typename T>
184 void recv(MPI::Comm& comm, T* array, int count, int source, int tag)
185 {
187 comm.Recv(array, count, MpiTraits<T>::type, source, tag);
188 } else {
189 // Try recv<T> by element, in case of explicit specialization.
190 // If there is no specialization or type, recv<T> throws.
191 for (int i = 0; i < count; ++i) {
192 recv<T>(comm, array[i], source, tag);
193 }
194 }
195 }
196
208 template <typename T>
209 void bcast(MPI::Intracomm& comm, T* array, int count, int root)
210 {
212 comm.Bcast(array, count, MpiTraits<T>::type, root);
213 } else {
214 // Try bcast<T> by element, in case of explicit specialization.
215 // If there is no specialization or type, bcast<T> throws.
216 for (int i = 0; i < count; ++i) {
217 bcast<T>(comm, array[i], root);
218 }
219 }
220 }
221
222 // DArray container partial specializations
223
236 template <typename T>
237 void send(MPI::Comm& comm, DArray<T>& array, int count, int dest, int tag)
238 {
239 // Preconditions
240 if (!(array.isAllocated())) {
241 UTIL_THROW("Cannot read unallocated DArray");
242 }
243 if (count > array.capacity()) {
244 UTIL_THROW("Error: Logical size count > DArray capacity");
245 }
246
248 comm.Send(&array[0], count, MpiTraits<T>::type, dest, tag);
249 } else {
250 // Try send<T> by element, in case of explicit specialization.
251 // If there is no specialization or type, send<T> throws.
252 for (int i = 0; i < count; ++i) {
253 send<T>(comm, array[i], dest, tag);
254 }
255 }
256 }
257
270 template <typename T>
271 void recv(MPI::Comm& comm, DArray<T>& array, int count, int source, int tag)
272 {
273 // Preconditions
274 if (!(array.isAllocated())) {
275 UTIL_THROW("Cannot read unallocated DArray");
276 }
277 if (count > array.capacity()) {
278 UTIL_THROW("Error: Logical size count > DArray capacity");
279 }
280
282 comm.Recv(&array[0], count, MpiTraits<T>::type, source, tag);
283 } else {
284 // Try recv<T> by element, in case of explicit specialization.
285 // If there is no specialization or type, recv<T> throws.
286 for (int i = 0; i < count; ++i) {
287 recv<T>(comm, array[i], source, tag);
288 }
289 }
290 }
291
303 template <typename T>
304 void bcast(MPI::Intracomm& comm, DArray<T>& array, int count, int root)
305 {
306 // Preconditions
307 if (!(array.isAllocated())) {
308 UTIL_THROW("Cannot read unallocated DArray");
309 }
310 if (count > array.capacity()) {
311 UTIL_THROW("Error: Logical size count > DArray capacity");
312 }
313
315 comm.Bcast(&array[0], count, MpiTraits<T>::type, root);
316 } else {
317 // try bcast<T> by element, in case of explicit specialization.
318 // If there is no specialization or type, bcast<T> throws.
319 for (int i = 0; i < count; ++i) {
320 bcast<T>(comm, array[i], root);
321 }
322 }
323 }
324
325 // DMatrix container partial specializations
326
340 template <typename T>
341 void send(MPI::Comm& comm, DMatrix<T>& matrix, int m, int n, int dest, int tag)
342 {
343 // Preconditions
344 if (!(matrix.isAllocated())) {
345 UTIL_THROW("Cannot read unallocated DMatrix");
346 }
347 if (m > matrix.capacity1()) {
348 UTIL_THROW("Error: Logical size m > DMatrix<T>::capacity1()");
349 }
350 if (n > matrix.capacity2()) {
351 UTIL_THROW("Error: Logical size n > DMatrix<T>::capacity2()");
352 }
353
355 int mp = matrix.capacity1();
356 int np = matrix.capacity2();
357 comm.Send(&matrix(0, 0), mp*np, MpiTraits<T>::type, dest, tag);
358 // Note: This method sends the entire physical memory block.
359 } else {
360 // try send<T> by element, in case of explicit specialization.
361 // If there is no specialization or type, send<T> throws.
362 int i, j;
363 for (i = 0; i < m; ++i) {
364 for (j = 0; j < n; ++j) {
365 send<T>(comm, matrix(i, j), dest, tag);
366 }
367 }
368 }
369 }
370
384 template <typename T>
385 void recv(MPI::Comm& comm, DMatrix<T>& matrix, int m, int n,
386 int source, int tag)
387 {
388 // Preconditions
389 if (!(matrix.isAllocated())) {
390 UTIL_THROW("Cannot recv unallocated DMatrix");
391 }
392 if (m > matrix.capacity1()) {
393 UTIL_THROW("Error: Logical size m > DMatrix<T>::capacity1()");
394 }
395 if (n > matrix.capacity2()) {
396 UTIL_THROW("Error: Logical size n > DMatrix<T>::capacity2()");
397 }
398
400 int mp = matrix.capacity1();
401 int np = matrix.capacity2();
402 comm.Recv(&matrix(0, 0), mp*np, MpiTraits<T>::type, source, tag);
403 // Note: This method receives the entire physical memory block.
404 } else {
405 // try recv<T> by element, in case of explicit specialization.
406 // If there is no specialization or type, recv<T> throws.
407 int i, j;
408 for (i = 0; i < m; ++i) {
409 for (j = 0; j < n; ++j) {
410 recv<T>(comm, matrix(i, j), source, tag);
411 }
412 }
413 }
414 }
415
428 template <typename T>
429 void bcast(MPI::Intracomm& comm, DMatrix<T>& matrix, int m, int n, int root)
430 {
431 // Preconditions
432 if (!(matrix.isAllocated())) {
433 UTIL_THROW("Cannot bcast unallocated DMatrix");
434 }
435 if (m > matrix.capacity1()) {
436 UTIL_THROW("Error: Logical size m > DMatrix<T>::capacity1()");
437 }
438 if (n > matrix.capacity2()) {
439 UTIL_THROW("Error: Logical size n > DMatrix<T>::capacity2()");
440 }
441
443 int mp = matrix.capacity1();
444 int np = matrix.capacity2();
445 comm.Bcast(&matrix(0, 0), mp*np, MpiTraits<T>::type, root);
446 // Note: This method receives the entire physical memory block.
447 } else {
448 // Try bcast<T> by element, in case of explicit specialization.
449 // If there is no specialization or type, bcast<T> throws.
450 int i, j;
451 for (i = 0; i < m; ++i) {
452 for (j = 0; j < n; ++j) {
453 bcast<T>(comm, matrix(i, j), root);
454 }
455 }
456 }
457 }
458
459 // bool (explicit specializations)
460
464 template <>
465 void send<bool>(MPI::Comm& comm, bool& data, int dest, int tag);
466
470 template <>
471 void recv<bool>(MPI::Comm& comm, bool& data, int source, int tag);
472
476 template <>
477 void bcast<bool>(MPI::Intracomm& comm, bool& data, int root);
478
479 // std::string (explicit specializations)
480
484 template <> void
485 send<std::string>(MPI::Comm& comm, std::string& data, int dest, int tag);
486
490 template <> void
491 recv<std::string>(MPI::Comm& comm, std::string& data, int source, int tag);
492
496 template <>
497 void bcast<std::string>(MPI::Intracomm& comm, std::string& data, int root);
498
499}
500#endif
501#endif
int capacity() const
Return allocated size.
Definition Array.h:159
Dynamically allocatable contiguous array template.
bool isAllocated() const
Return true if this DArray has been allocated, false otherwise.
Definition DArray.h:247
Dynamically allocated Matrix.
Definition DMatrix.h:25
bool isAllocated() const
Return true if the DMatrix has been allocated, false otherwise.
Definition DMatrix.h:202
int capacity2() const
Get number of columns (range of the second array index).
Definition Matrix.h:143
int capacity1() const
Get number of rows (range of the first array index).
Definition Matrix.h:136
Default MpiTraits class.
Definition MpiTraits.h:44
File containing preprocessor macros for error handling.
#define UTIL_THROW(msg)
Macro for throwing an Exception, reporting function, file and line number.
Definition global.h:51
Utility classes for scientific computation.
void send(MPI::Comm &comm, T &data, int dest, int tag)
Send a single T value.
Definition MpiSendRecv.h:99
void recv< std::string >(MPI::Comm &comm, std::string &data, int source, int tag)
Explicit specialization of recv for std::string data.
void send< bool >(MPI::Comm &comm, bool &data, int dest, int tag)
Explicit specialization of send for bool data.
void recv< bool >(MPI::Comm &comm, bool &data, int source, int tag)
Explicit specialization of recv for bool data.
void bcast< std::string >(MPI::Intracomm &comm, std::string &data, int root)
Explicit specialization of bcast for std::string data.
void bcast(MPI::Intracomm &comm, T &data, int root)
Broadcast a single T value.
void bcast< bool >(MPI::Intracomm &comm, bool &data, int root)
Explicit specialization of bcast for bool data.
void recv(MPI::Comm &comm, T &data, int source, int tag)
Receive a single T value.
void send< std::string >(MPI::Comm &comm, std::string &data, int dest, int tag)
Explicit specialization of send for std::string data.