PSCF v1.2
AmIteratorGrid.tpp
1#ifndef RPG_AM_ITERATOR_GRID_TPP
2#define RPG_AM_ITERATOR_GRID_TPP
3
4/*
5* PSCF - Polymer Self-Consistent Field Theory
6*
7* Copyright 2016 - 2022, The Regents of the University of Minnesota
8* Distributed under the terms of the GNU General Public License.
9*/
10
11#include "AmIteratorGrid.h"
12#include <rpg/System.h>
13#include <prdc/crystal/UnitCell.h>
14#include <prdc/cuda/resources.h>
15#include <pscf/inter/Interaction.h>
16#include <pscf/iterator/NanException.h>
17
18#include <util/global.h>
19#include <cmath>
20
21namespace Pscf {
22namespace Rpg {
23
24 using namespace Util;
25 using namespace Prdc;
26 using namespace Prdc::Cuda;
27
28 // Constructor
29 template <int D>
31 : Iterator<D>(system),
32 imposedFields_(system)
33 {
34 setClassName("AmIteratorGrid");
35 isSymmetric_ = false;
36 }
37
38 // Destructor
39 template <int D>
42
43 // Read parameter file block
44 template <int D>
45 void AmIteratorGrid<D>::readParameters(std::istream& in)
46 {
47 // Call parent class readParameters
48 AmIteratorTmpl<Iterator<D>,FieldCUDA>::readParameters(in);
49 AmIteratorTmpl<Iterator<D>,FieldCUDA>::readErrorType(in);
50
51 // Allocate local modified copy of Interaction class
52 interaction_.setNMonomer(system().mixture().nMonomer());
53
54 // Default parameter values
55 isFlexible_ = 1;
56 scaleStress_ = 10.0;
57
58 int np = system().domain().unitCell().nParameter();
59 UTIL_CHECK(np > 0);
60 UTIL_CHECK(np <= 6);
61 UTIL_CHECK(system().domain().unitCell().lattice() != UnitCell<D>::Null);
62
63 // Read in optional isFlexible value
64 readOptional(in, "isFlexible", isFlexible_);
65
66 // Populate flexibleParams_ based on isFlexible_ (all 0s or all 1s),
67 // then optionally overwrite with user input from param file
68 if (isFlexible_) {
69 flexibleParams_.clear();
70 for (int i = 0; i < np; i++) {
71 flexibleParams_.append(true); // Set all values to true
72 }
73 // Read optional flexibleParams_ array to overwrite current array
74 readOptionalFSArray(in, "flexibleParams", flexibleParams_, np);
75 if (nFlexibleParams() == 0) isFlexible_ = false;
76 } else { // isFlexible_ = false
77 flexibleParams_.clear();
78 for (int i = 0; i < np; i++) {
79 flexibleParams_.append(false); // Set all values to false
80 }
81 }
82
83 // Read optional scaleStress value
84 readOptional(in, "scaleStress", scaleStress_);
85
86 // Read optional ImposedFieldsGenerator object
87 readParamCompositeOptional(in, imposedFields_);
88 }
89
90 // Output timing results to log file.
91 template<int D>
92 void AmIteratorGrid<D>::outputTimers(std::ostream& out)
93 {
94 // Output timing results, if requested.
95 out << "\n";
96 out << "Iterator times contributions:\n";
97 AmIteratorTmpl<Iterator<D>, FieldCUDA >::outputTimers(out);
98 }
99
100 // Protected virtual function
101
102 // Setup before entering iteration loop
103 template <int D>
104 void AmIteratorGrid<D>::setup(bool isContinuation)
105 {
106 if (imposedFields_.isActive()) {
107 imposedFields_.setup();
108 }
109
110 AmIteratorTmpl<Iterator<D>, FieldCUDA>::setup(isContinuation);
111 interaction_.update(system().interaction());
112 }
113
114 // Private virtual functions used to implement AM algorithm
115
116 // Set vector a equal to vector b (a = b).
117 template <int D>
119 {
120 UTIL_CHECK(b.capacity() == a.capacity());
121 VecOp::eqV(a, b);
122 }
123
124 // Compute and return inner product of two real fields.
125 template <int D>
126 double AmIteratorGrid<D>::dotProduct(FieldCUDA const & a,
127 FieldCUDA const & b)
128 {
129 UTIL_CHECK(a.capacity() == b.capacity());
130 double val = Reduce::innerProduct(a, b);
131 if (std::isnan(val)) { // if value is NaN, throw NanException
132 throw NanException("AmIteratorGrid::dotProduct", __FILE__,
133 __LINE__, 0);
134 }
135 return val;
136 }
137
138 // Find the maximum magnitude element of a residual vector.
139 template <int D>
140 double AmIteratorGrid<D>::maxAbs(FieldCUDA const & a)
141 {
142 double val = Reduce::maxAbs(a);
143 if (std::isnan(val)) { // if value is NaN, throw NanException
144 throw NanException("AmIteratorGrid::maxAbs", __FILE__, __LINE__, 0);
145 }
146 return val;
147 }
148
149 // Update the series of residual vectors.
150 template <int D>
151 void AmIteratorGrid<D>::updateBasis(RingBuffer<FieldCUDA> & basis,
152 RingBuffer<FieldCUDA> const & hists)
153 {
154 // Make sure at least two histories are stored
155 UTIL_CHECK(hists.size() >= 2);
156
157 // Set up array to store new basis
158 basis.advance();
159 if (basis[0].isAllocated()) {
160 UTIL_CHECK(basis[0].capacity() == hists[0].capacity());
161 } else {
162 basis[0].allocate(hists[0].capacity());
163 }
164
165 // New basis vector is difference between two most recent states
166 VecOp::subVV(basis[0], hists[0], hists[1]);
167 }
168
169 // Compute trial field so as to minimize L2 norm of residual.
170 template <int D>
171 void
172 AmIteratorGrid<D>::addHistories(FieldCUDA& trial,
173 RingBuffer<FieldCUDA> const & basis,
174 DArray<double> coeffs, int nHist)
175 {
176 for (int i = 0; i < nHist; i++) {
177 VecOp::addEqVc(trial, basis[i], -1.0 * coeffs[i]);
178 }
179 }
180
181 // Add predicted error to the trial field.
182 template <int D>
183 void AmIteratorGrid<D>::addPredictedError(FieldCUDA& fieldTrial,
184 FieldCUDA const & resTrial,
185 double lambda)
186 {
187 VecOp::addEqVc(fieldTrial, resTrial, lambda);
188 }
189
190 // Checks if the system has an initial guess
191 template <int D>
192 bool AmIteratorGrid<D>::hasInitialGuess()
193 { return system().w().hasData(); }
194
195 // Compute the number of elements in the residual vector.
196 template <int D>
197 int AmIteratorGrid<D>::nElements()
198 {
199 const int nMonomer = system().mixture().nMonomer();
200 const int nMesh = system().mesh().size();
201
202 int nEle = nMonomer*nMesh;
203
204 if (isFlexible_) {
205 nEle += nFlexibleParams();
206 }
207
208 return nEle;
209 }
210
211 // Get the current w fields and lattice parameters.
212 template <int D>
213 void AmIteratorGrid<D>::getCurrent(FieldCUDA& curr)
214 {
215 const int nMonomer = system().mixture().nMonomer();
216 const int nMesh = system().mesh().size();
217 const int n = nElements();
218 UTIL_CHECK(curr.capacity() == n);
219
220 // Loop to unfold the system fields and store them in one long array
221 for (int i = 0; i < nMonomer; i++) {
222 VecOp::eqV(curr, system().w().rgrid(i), i*nMesh, 0, nMesh);
223 }
224
225 // If flexible unit cell, also store unit cell parameters
226 if (isFlexible_) {
227 const int nParam = system().unitCell().nParameter();
228 FSArray<double,6> const & parameters
229 = system().unitCell().parameters();
230
231 // convert into a cudaReal array
232 HostDArray<cudaReal> tempH(nFlexibleParams());
233 int counter = 0;
234 for (int i = 0; i < nParam; i++) {
235 if (flexibleParams_[i]) {
236 tempH[counter] = scaleStress_ * parameters[i];
237 counter++;
238 }
239 }
240 UTIL_CHECK(counter == tempH.capacity());
241
242 // Copy parameters to the end of the curr array
243 FieldCUDA tempD;
244 tempD.associate(curr, nMonomer*nMesh, tempH.capacity());
245 tempD = tempH; // copy from host to device
246 }
247 }
248
249 // Solve MDE for current state of system.
250 template <int D>
251 void AmIteratorGrid<D>::evaluate()
252 {
253 // Solve MDEs for current omega field
254 system().compute(isFlexible_);
255 }
256
257 // Gets the residual vector from system.
258 template <int D>
259 void AmIteratorGrid<D>::getResidual(FieldCUDA& resid)
260 {
261 const int n = nElements();
262 const int nMonomer = system().mixture().nMonomer();
263 const int nMesh = system().mesh().size();
264
265 // Initialize residuals to zero. Kernel will take care of potential
266 // additional elements (n vs nMesh).
267 VecOp::eqS(resid, 0.0);
268
269 // Array of FieldCUDA arrays associated with slices of resid.
270 // one FieldCUDA array per monomer species, each of size nMesh.
271 DArray<FieldCUDA> residSlices;
272 residSlices.allocate(nMonomer);
273 for (int i = 0; i < nMonomer; i++) {
274 residSlices[i].associate(resid, i*nMesh, nMesh);
275 }
276
277 // Compute SCF residuals
278 for (int i = 0; i < nMonomer; i++) {
279 for (int j = 0; j < nMonomer; j++) {
280 VecOp::addVcVcVc(residSlices[i], residSlices[i], 1.0,
281 system().c().rgrid(j), interaction_.chi(i, j),
282 system().w().rgrid(j), -interaction_.p(i, j));
283 }
284 }
285
286 // If iterator has mask, account for it in residual values
287 if (system().hasMask()) {
288 double coeff = -1.0 / interaction_.sumChiInverse();
289 for (int i = 0; i < nMonomer; ++i) {
290 VecOp::addEqVc(residSlices[i], system().mask().rgrid(), coeff);
291 }
292 }
293
294 // If iterator has external fields, account for them in the values
295 // of the residuals
296 if (system().hasExternalFields()) {
297 for (int i = 0; i < nMonomer; ++i) {
298 for (int j = 0; j < nMonomer; ++j) {
299 double p = interaction_.p(i,j);
300 VecOp::addEqVc(residSlices[i], system().h().rgrid(j), p);
301 }
302 }
303 }
304
305 // If ensemble is not canonical, account for incompressibility.
306 if (!system().mixture().isCanonical()) {
307 cudaReal factor = 1.0 / interaction_.sumChiInverse();
308 VecOp::subEqS(resid, factor, 0, nMonomer*nMesh);
309 } else {
310 for (int i = 0; i < nMonomer; i++) {
311 // Find current average
312 cudaReal average = findAverage(residSlices[i]);
313 // subtract out average to set residual average to zero
314 VecOp::subEqS(residSlices[i], average);
315 }
316 }
317
318 // If variable unit cell, compute stress residuals
319 if (isFlexible_) {
320 const int nParam = system().unitCell().nParameter();
321 HostDArray<cudaReal> stressH(nFlexibleParams());
322
323 int counter = 0;
324 for (int i = 0; i < nParam; i++) {
325 if (flexibleParams_[i]) {
326 double stress = system().mixture().stress(i);
327
328 // Correct stress to account for effect of imposed fields
329 if (imposedFields_.isActive()) {
330 stress = imposedFields_.correctedStress(i,stress);
331 }
332
333 stressH[counter] = -1 * scaleStress_ * stress;
334 counter++;
335 }
336 }
337 UTIL_CHECK(counter == stressH.capacity());
338
339 FieldCUDA stressD;
340 stressD.associate(resid, nMonomer*nMesh, stressH.capacity());
341 stressD = stressH; // copy from host to device
342 }
343 }
344
345 // Update the system with a new trial field vector.
346 template <int D>
347 void AmIteratorGrid<D>::update(FieldCUDA& newGuess)
348 {
349 const int nMonomer = system().mixture().nMonomer();
350 const int nMesh = system().mesh().size();
351
352 // If canonical, explicitly set homogeneous field components
353 if (system().mixture().isCanonical()) {
354 cudaReal average, wAverage, cAverage;
355 for (int i = 0; i < nMonomer; i++) {
356 // Define array associated with a slice of newGuess
357 FieldCUDA ngSlice;
358 ngSlice.associate(newGuess, i*nMesh, nMesh);
359
360 // Find current spatial average
361 average = findAverage(ngSlice);
362
363 // Subtract average from field, setting average to zero
364 VecOp::subEqS(ngSlice, average);
365
366 // Compute the new average omega value
367 wAverage = 0;
368 for (int j = 0; j < nMonomer; j++) {
369 // Find average concentration for j monomers
370 if (system().w().isSymmetric()) { // c().basis() has data
371 cAverage = system().c().basis(j)[0];
372 } else { // average must be calculated
373 cAverage = findAverage(system().c().rgrid(j));
374 }
375 wAverage += interaction_.chi(i,j) * cAverage;
376 }
377
378 // If system has external fields, include them in homogeneous field
379 if (system().hasExternalFields()) {
380 if (system().h().isSymmetric()) { // h().basis() has data
381 wAverage += system().h().basis(i)[0];
382 } else { // average must be calculated
383 wAverage += findAverage(system().h().rgrid(i));
384 }
385 }
386
387 // Add new average omega value to the field
388 VecOp::addEqS(ngSlice, wAverage);
389 }
390 }
391
392 system().setWRGrid(newGuess);
393 system().symmetrizeWFields();
394
395 // If flexible unit cell, update cell parameters
396 if (isFlexible_) {
397 FSArray<double,6> parameters;
398 parameters = system().domain().unitCell().parameters();
399 const int nParam = system().unitCell().nParameter();
400
401 // transfer from device to host
402 HostDArray<cudaReal> tempH(nFlexibleParams());
403 tempH.copySlice(newGuess, nMonomer*nMesh);
404
405 const double coeff = 1.0 / scaleStress_;
406 int counter = 0;
407 for (int i = 0; i < nParam; i++) {
408 if (flexibleParams_[i]) {
409 parameters[i] = coeff * tempH[i];
410 counter++;
411 }
412 }
413 UTIL_CHECK(counter == tempH.capacity());
414
415 system().setUnitCell(parameters);
416 }
417
418 // Update imposed fields if needed
419 if (imposedFields_.isActive()) {
420 imposedFields_.update();
421 }
422 }
423
424 // Output relevant system details to the iteration log file.
425 template<int D>
426 void AmIteratorGrid<D>::outputToLog()
427 {
428 if (isFlexible_ && verbose() > 1) {
429 const int nParam = system().domain().unitCell().nParameter();
430 const int nMonomer = system().mixture().nMonomer();
431 const int nMesh = system().mesh().size();
432
433 // transfer stress residuals from device to host
434 HostDArray<cudaReal> tempH(nFlexibleParams());
435 tempH.copySlice(residual(), nMonomer*nMesh);
436
437 int counter = 0;
438 for (int i = 0; i < nParam; i++) {
439 if (flexibleParams_[i]) {
440 double stress = tempH[counter] / (-1.0 * scaleStress_);
441 Log::file()
442 << " Cell Param " << i << " = "
443 << Dbl(system().domain().unitCell().parameters()[i], 15)
444 << " , stress = "
445 << Dbl(stress, 15)
446 << "\n";
447 counter++;
448 }
449 }
450 }
451 }
452
453 // Return specialized sweep parameter types to add to the Sweep object
454 template<int D>
456 {
458 if (imposedFields_.isActive()) {
459 arr = imposedFields_.getParameterTypes();
460 }
461 return arr;
462 }
463 // Set the value of a specialized sweep parameter
464 template<int D>
465 void AmIteratorGrid<D>::setParameter(std::string name, DArray<int> ids,
466 double value, bool& success)
467 {
468 if (imposedFields_.isActive()) {
469 imposedFields_.setParameter(name, ids, value, success);
470 } else {
471 success = false;
472 }
473 }
474 // Get the value of a specialized sweep parameter
475 template<int D>
476 double AmIteratorGrid<D>::getParameter(std::string name,
477 DArray<int> ids, bool& success)
478 const
479 {
480 if (imposedFields_.isActive()) {
481 return imposedFields_.getParameter(name, ids, success);
482 } else {
483 success = false;
484 return 0.0;
485 }
486 }
487
488 // --- Private member functions specific to this implementation ---
489
490 // Calculate the average value of an array.
491 template<int D>
492 cudaReal AmIteratorGrid<D>::findAverage(FieldCUDA const & field)
493 {
494 return Reduce::sum(field) / field.capacity();
495 }
496
497}
498}
499#endif
Template for Anderson mixing iterator algorithm.
Dynamic array on the GPU device with aligned data.
Definition rpg/System.h:32
int capacity() const
Return allocated capacity.
Exception thrown when not-a-number (NaN) is encountered.
Base template for UnitCell<D> classes, D=1, 2 or 3.
Definition rpg/System.h:34
Rpg implementation of the Anderson Mixing iterator.
AmIteratorGrid(System< D > &system)
Constructor.
void readParameters(std::istream &in)
Read all parameters and initialize.
void setClassName(const char *className)
Set class name string.
void setParameter(std::string name, DArray< int > ids, double value, bool &success)
Set the value of a specialized sweep parameter.
double getParameter(std::string name, DArray< int > ids, bool &success) const
Get the value of a specialized sweep parameter.
void setup(bool isContinuation)
Setup iterator just before entering iteration loop.
void outputTimers(std::ostream &out)
Output timing results to log file.
GArray< ParameterType > getParameterTypes()
Return specialized sweep parameter types to add to the Sweep object.
Base class for iterative solvers for SCF equations.
Definition rpg/System.h:37
Main class for calculations that represent one system.
Definition rpg/System.h:107
Dynamically allocatable contiguous array template.
void allocate(int capacity)
Allocate the underlying C array.
Definition DArray.h:199
Wrapper for a double precision number, for formatted ostream output.
Definition Dbl.h:40
A fixed capacity (static) contiguous array with a variable logical size.
Definition rpg/System.h:28
An automatically growable array, analogous to a std::vector.
Definition GArray.h:34
static std::ostream & file()
Get log ostream by reference.
Definition Log.cpp:57
Class for storing history of previous values in an array.
Definition RingBuffer.h:27
void allocate(int capacity)
Allocate a new empty buffer.
Definition RingBuffer.h:257
void advance()
Advances the pointer to an added element without assigning a value.
Definition RingBuffer.h:306
int size() const
Return number of values currently in the buffer.
Definition RingBuffer.h:325
File containing preprocessor macros for error handling.
#define UTIL_CHECK(condition)
Assertion macro suitable for serial or parallel production code.
Definition global.h:68
cudaReal innerProduct(DeviceArray< cudaReal > const &a, DeviceArray< cudaReal > const &b)
Compute inner product of two real arrays (GPU kernel wrapper).
Definition Reduce.cu:891
cudaReal maxAbs(DeviceArray< cudaReal > const &in)
Get maximum absolute magnitude of array elements (GPU kernel wrapper).
Definition Reduce.cu:648
cudaReal sum(DeviceArray< cudaReal > const &in)
Compute sum of array elements (GPU kernel wrapper).
Definition Reduce.cu:480
void eqV(DeviceArray< cudaReal > &a, DeviceArray< cudaReal > const &b, const int beginIdA, const int beginIdB, const int n)
Vector assignment, a[i] = b[i], kernel wrapper (cudaReal).
Definition VecOp.cu:1020
void subEqS(DeviceArray< cudaReal > &a, cudaReal const b, const int beginIdA, const int n)
Vector subtraction in-place, a[i] -= b, kernel wrapper (cudaReal).
Definition VecOp.cu:1779
void addEqS(DeviceArray< cudaReal > &a, cudaReal const b, const int beginIdA, const int n)
Vector addition in-place, a[i] += b, kernel wrapper (cudaReal).
Definition VecOp.cu:1683
void addVcVcVc(DeviceArray< cudaReal > &a, DeviceArray< cudaReal > const &b, cudaReal const c, DeviceArray< cudaReal > const &d, cudaReal const e, DeviceArray< cudaReal > const &f, cudaReal const g)
3-vec addition w coeff, a[i] = (b[i]*c) + (d[i]*e) + (f[i]*g), kernel wrapper.
Definition VecOpMisc.cu:322
void subVV(DeviceArray< cudaReal > &a, DeviceArray< cudaReal > const &b, DeviceArray< cudaReal > const &c, const int beginIdA, const int beginIdB, const int beginIdC, const int n)
Vector subtraction, a[i] = b[i] - c[i], kernel wrapper (cudaReal).
Definition VecOp.cu:1228
void eqS(DeviceArray< cudaReal > &a, cudaReal const b, const int beginIdA, const int n)
Vector assignment, a[i] = b, kernel wrapper (cudaReal).
Definition VecOp.cu:1054
void addEqVc(DeviceArray< cudaReal > &a, DeviceArray< cudaReal > const &b, cudaReal const c)
Vector addition in-place w/ coefficient, a[i] += b[i] * c, kernel wrapper.
Definition VecOpMisc.cu:343
PSCF package top-level namespace.
Definition param_pc.dox:1
Utility classes for scientific computation.