682 *
const unsigned int component)
const 694 * <a name=
"ThecodeHeatEquationcodeimplementation"></a>
695 * <h3>The <code>HeatEquation</code> implementation</h3>
699 * It is time now
for the implementation of the main
class. Let
's 700 * start with the constructor which selects a linear element, a time 701 * step constant at 1/500 (remember that one period of the source 702 * on the right hand side was set to 0.2 above, so we resolve each 703 * period with 100 time steps) and chooses the Crank Nicolson method 704 * by setting @f$\theta=1/2@f$. 708 * HeatEquation<dim>::HeatEquation() 710 * , dof_handler(triangulation) 712 * , time_step(1. / 500) 713 * , timestep_number(0) 722 * <a name="codeHeatEquationsetup_systemcode"></a> 723 * <h4><code>HeatEquation::setup_system</code></h4> 727 * The next function is the one that sets up the DoFHandler object, 728 * computes the constraints, and sets the linear algebra objects 729 * to their correct sizes. We also compute the mass and Laplace 730 * matrix here by simply calling two functions in the library. 734 * Note that we do not take the hanging node constraints into account when 735 * assembling the matrices (both functions have an AffineConstraints argument 736 * that defaults to an empty object). This is because we are going to 737 * condense the constraints in run() after combining the matrices for the 742 * void HeatEquation<dim>::setup_system() 744 * dof_handler.distribute_dofs(fe); 746 * std::cout << std::endl 747 * << "===========================================" << std::endl 748 * << "Number of active cells: " << triangulation.n_active_cells() 750 * << "Number of degrees of freedom: " << dof_handler.n_dofs() 754 * constraints.clear(); 755 * DoFTools::make_hanging_node_constraints(dof_handler, constraints); 756 * constraints.close(); 758 * DynamicSparsityPattern dsp(dof_handler.n_dofs()); 759 * DoFTools::make_sparsity_pattern(dof_handler, 762 * /*keep_constrained_dofs = */ true); 763 * sparsity_pattern.copy_from(dsp); 765 * mass_matrix.reinit(sparsity_pattern); 766 * laplace_matrix.reinit(sparsity_pattern); 767 * system_matrix.reinit(sparsity_pattern); 769 * MatrixCreator::create_mass_matrix(dof_handler, 770 * QGauss<dim>(fe.degree + 1), 772 * MatrixCreator::create_laplace_matrix(dof_handler, 773 * QGauss<dim>(fe.degree + 1), 776 * solution.reinit(dof_handler.n_dofs()); 777 * old_solution.reinit(dof_handler.n_dofs()); 778 * system_rhs.reinit(dof_handler.n_dofs()); 785 * <a name="codeHeatEquationsolve_time_stepcode"></a> 786 * <h4><code>HeatEquation::solve_time_step</code></h4> 790 * The next function is the one that solves the actual linear system 791 * for a single time step. There is nothing surprising here: 795 * void HeatEquation<dim>::solve_time_step() 797 * SolverControl solver_control(1000, 1e-8 * system_rhs.l2_norm()); 798 * SolverCG<Vector<double>> cg(solver_control); 800 * PreconditionSSOR<SparseMatrix<double>> preconditioner; 801 * preconditioner.initialize(system_matrix, 1.0); 803 * cg.solve(system_matrix, solution, system_rhs, preconditioner); 805 * constraints.distribute(solution); 807 * std::cout << " " << solver_control.last_step() << " CG iterations." 816 * <a name="codeHeatEquationoutput_resultscode"></a> 817 * <h4><code>HeatEquation::output_results</code></h4> 821 * Neither is there anything new in generating graphical output other than the 822 * fact that we tell the DataOut object what the current time and time step 823 * number is, so that this can be written into the output file : 827 * void HeatEquation<dim>::output_results() const 829 * DataOut<dim> data_out; 831 * data_out.attach_dof_handler(dof_handler); 832 * data_out.add_data_vector(solution, "U"); 834 * data_out.build_patches(); 836 * data_out.set_flags(DataOutBase::VtkFlags(time, timestep_number)); 838 * const std::string filename = 839 * "solution-" + Utilities::int_to_string(timestep_number, 3) + ".vtk"; 840 * std::ofstream output(filename); 841 * data_out.write_vtk(output); 848 * <a name="codeHeatEquationrefine_meshcode"></a> 849 * <h4><code>HeatEquation::refine_mesh</code></h4> 853 * This function is the interesting part of the program. It takes care of 854 * the adaptive mesh refinement. The three tasks 855 * this function performs is to first find out which cells to 856 * refine/coarsen, then to actually do the refinement and eventually 857 * transfer the solution vectors between the two different grids. The first 858 * task is simply achieved by using the well-established Kelly error 859 * estimator on the solution. The second task is to actually do the 860 * remeshing. That involves only basic functions as well, such as the 861 * <code>refine_and_coarsen_fixed_fraction</code> that refines those cells 862 * with the largest estimated error that together make up 60 per cent of the 863 * error, and coarsens those cells with the smallest error that make up for 864 * a combined 40 per cent of the error. Note that for problems such as the 865 * current one where the areas where something is going on are shifting 866 * around, we want to aggressively coarsen so that we can move cells 867 * around to where it is necessary. 871 * As already discussed in the introduction, too small a mesh leads to 872 * too small a time step, whereas too large a mesh leads to too little 873 * resolution. Consequently, after the first two steps, we have two 874 * loops that limit refinement and coarsening to an allowable range of 879 * void HeatEquation<dim>::refine_mesh(const unsigned int min_grid_level, 880 * const unsigned int max_grid_level) 882 * Vector<float> estimated_error_per_cell(triangulation.n_active_cells()); 884 * KellyErrorEstimator<dim>::estimate( 886 * QGauss<dim - 1>(fe.degree + 1), 887 * std::map<types::boundary_id, const Function<dim> *>(), 889 * estimated_error_per_cell); 891 * GridRefinement::refine_and_coarsen_fixed_fraction(triangulation, 892 * estimated_error_per_cell, 896 * if (triangulation.n_levels() > max_grid_level) 897 * for (const auto &cell : 898 * triangulation.active_cell_iterators_on_level(max_grid_level)) 899 * cell->clear_refine_flag(); 900 * for (const auto &cell : 901 * triangulation.active_cell_iterators_on_level(min_grid_level)) 902 * cell->clear_coarsen_flag(); 905 * These two loops above are slightly different but this is easily 906 * explained. In the first loop, instead of calling 907 * <code>triangulation.end()</code> we may as well have called 908 * <code>triangulation.end_active(max_grid_level)</code>. The two 909 * calls should yield the same iterator since iterators are sorted 910 * by level and there should not be any cells on levels higher than 911 * on level <code>max_grid_level</code>. In fact, this very piece 912 * of code makes sure that this is the case. 916 * As part of mesh refinement we need to transfer the solution vectors 917 * from the old mesh to the new one. To this end we use the 918 * SolutionTransfer class and we have to prepare the solution vectors that 919 * should be transferred to the new grid (we will lose the old grid once 920 * we have done the refinement so the transfer has to happen concurrently 921 * with refinement). At the point where we call this function, we will 922 * have just computed the solution, so we no longer need the old_solution 923 * variable (it will be overwritten by the solution just after the mesh 924 * may have been refined, i.e., at the end of the time step; see below). 925 * In other words, we only need the one solution vector, and we copy it 926 * to a temporary object where it is safe from being reset when we further 927 * down below call <code>setup_system()</code>. 931 * Consequently, we initialize a SolutionTransfer object by attaching 932 * it to the old DoF handler. We then prepare the triangulation and the 933 * data vector for refinement (in this order). 936 * SolutionTransfer<dim> solution_trans(dof_handler); 938 * Vector<double> previous_solution; 939 * previous_solution = solution; 940 * triangulation.prepare_coarsening_and_refinement(); 941 * solution_trans.prepare_for_coarsening_and_refinement(previous_solution); 945 * Now everything is ready, so do the refinement and recreate the DoF 946 * structure on the new grid, and finally initialize the matrix structures 947 * and the new vectors in the <code>setup_system</code> function. Next, we 948 * actually perform the interpolation of the solution from old to new 949 * grid. The final step is to apply the hanging node constraints to the 950 * solution vector, i.e., to make sure that the values of degrees of 951 * freedom located on hanging nodes are so that the solution is 952 * continuous. This is necessary since SolutionTransfer only operates on 953 * cells locally, without regard to the neighborhoof. 956 * triangulation.execute_coarsening_and_refinement(); 959 * solution_trans.interpolate(previous_solution, solution); 960 * constraints.distribute(solution); 968 * <a name="codeHeatEquationruncode"></a> 969 * <h4><code>HeatEquation::run</code></h4> 973 * This is the main driver of the program, where we loop over all 974 * time steps. At the top of the function, we set the number of 975 * initial global mesh refinements and the number of initial cycles of 976 * adaptive mesh refinement by repeating the first time step a few 977 * times. Then we create a mesh, initialize the various objects we will 978 * work with, set a label for where we should start when re-running 979 * the first time step, and interpolate the initial solution onto 980 * out mesh (we choose the zero function here, which of course we could 981 * do in a simpler way by just setting the solution vector to zero). We 982 * also output the initial time step once. 986 * @note If you're an experienced programmer, you may be surprised
987 * that we use a <code>
goto</code> statement in
this piece of code!
988 * <code>
goto</code> statements are not particularly well liked any
989 * more since Edsgar Dijkstra,
one of the greats of computer science,
990 * wrote a letter in 1968 called
"Go To Statement considered harmful" 991 * (see <a href=
"http://en.wikipedia.org/wiki/Considered_harmful">here</a>).
992 * The author of
this code subscribes to
this notion whole-heartedly:
993 * <code>goto</code> is hard to understand. In fact, deal.II contains
994 * virtually no occurrences: excluding code that was essentially
995 * transcribed from books and not counting duplicated code pieces,
996 * there are 3 locations in about 600,000 lines of code; we also
997 * use it in 4 tutorial programs, in exactly the same context
998 * as here. Instead of trying to justify the occurrence here,
999 * let
's first look at the code and we'll come back to the issue
1000 * at the
end of
function.
1003 * template <int dim>
1006 *
const unsigned int initial_global_refinement = 2;
1007 *
const unsigned int n_adaptive_pre_refinement_steps = 4;
1014 *
unsigned int pre_refinement_step = 0;
1019 * start_time_iteration:
1021 * tmp.
reinit(solution.size());
1022 * forcing_terms.
reinit(solution.size());
1028 * solution = old_solution;
1034 * Then we start the main
loop until the computed time exceeds our
1035 *
end time of 0.5. The
first task is to build the right hand
1036 * side of the linear system we need to solve in each time step.
1037 * Recall that it contains the term @f$MU^{n-1}-(1-\theta)k_n AU^{n-1}@f$.
1038 * We put these terms into the variable system_rhs, with the
1039 * help of a temporary vector:
1042 *
while (time <= 0.5)
1044 * time += time_step;
1045 * ++timestep_number;
1047 * std::cout <<
"Time step " << timestep_number <<
" at t=" << time
1052 * laplace_matrix.vmult(tmp, old_solution);
1053 * system_rhs.add(-(1 - theta) * time_step, tmp);
1057 * The
second piece is to compute the contributions of the source
1058 * terms. This corresponds to the term @f$k_n
1059 * \left[ (1-\theta)
F^{n-1} + \theta
F^n \right]@f$. The following
1061 * vectors @f$F@f$, where we
set the time of the right hand side
1062 * (source)
function before we evaluate it. The result of
this 1063 * all ends up in the forcing_terms variable:
1066 * RightHandSide<dim> rhs_function;
1067 * rhs_function.set_time(time);
1072 * forcing_terms = tmp;
1073 * forcing_terms *= time_step * theta;
1075 * rhs_function.set_time(time - time_step);
1081 * forcing_terms.
add(time_step * (1 - theta), tmp);
1085 * Next, we add the forcing terms to the ones that
1086 * come from the time stepping, and also build the
matrix 1087 * @f$M+k_n\theta
A@f$ that we have to
invert in each time step.
1088 * The
final piece of these operations is to eliminate
1089 * hanging node constrained degrees of freedom from the
1093 * system_rhs += forcing_terms;
1096 * system_matrix.
add(theta * time_step, laplace_matrix);
1098 * constraints.condense(system_matrix, system_rhs);
1102 * There is
one more operation we need to
do before we
1103 * can solve it: boundary values. To
this end, we create
1104 * a boundary
value object,
set the proper time to the
one 1105 * of the current time step, and evaluate it as we have
1106 * done many times before. The result is used to also
1107 *
set the correct boundary values in the linear system:
1111 * BoundaryValues<dim> boundary_values_function;
1112 * boundary_values_function.set_time(time);
1114 * std::map<types::global_dof_index, double> boundary_values;
1117 * boundary_values_function,
1128 * With
this out of the way, all we have to
do is solve the
1129 * system, generate graphical data, and...
1132 * solve_time_step();
1138 * ...take care of mesh refinement. Here, what we want to
do is
1139 * (i)
refine the requested number of times at the very beginning
1140 * of the solution procedure, after which we jump to the top to
1141 * restart the time iteration, (ii)
refine every fifth time
1146 * The time
loop and, indeed, the main part of the program ends
1147 * with starting into the next time step by setting old_solution
1148 * to the solution we have just computed.
1151 *
if ((timestep_number == 1) &&
1152 * (pre_refinement_step < n_adaptive_pre_refinement_steps))
1154 * refine_mesh(initial_global_refinement,
1155 * initial_global_refinement +
1156 * n_adaptive_pre_refinement_steps);
1157 * ++pre_refinement_step;
1159 * tmp.
reinit(solution.size());
1160 * forcing_terms.
reinit(solution.size());
1162 * std::cout << std::endl;
1164 *
goto start_time_iteration;
1166 *
else if ((timestep_number > 0) && (timestep_number % 5 == 0))
1168 * refine_mesh(initial_global_refinement,
1169 * initial_global_refinement +
1170 * n_adaptive_pre_refinement_steps);
1171 * tmp.
reinit(solution.size());
1172 * forcing_terms.
reinit(solution.size());
1175 * old_solution = solution;
1181 * Now that you have seen what the
function does, let us come back to the issue
1182 * of the <code>
goto</code>. In essence, what the code does is
1183 * something like
this:
1184 * <div
class=CodeFragmentInTutorialComment>
1189 * start_time_iteration:
1190 *
for (timestep=1...)
1193 *
if (timestep==1 && not happy with the result)
1195 * adjust some data structures;
1196 *
goto start_time_iteration;
1203 * Here, the condition
"happy with the result" is whether we
'd like to keep 1204 * the current mesh or would rather refine the mesh and start over on the 1205 * new mesh. We could of course replace the use of the <code>goto</code> 1207 * <div class=CodeFragmentInTutorialComment> 1215 * if (not happy with the result) 1216 * adjust some data structures; 1223 * for (timestep=2...) 1231 * This has the advantage of getting rid of the <code>goto</code> 1232 * but the disadvantage of having to duplicate the code that implements 1233 * the "solve timestep" and "postprocess" operations in two different 1234 * places. This could be countered by putting these parts of the code 1235 * (sizable chunks in the actual implementation above) into their 1236 * own functions, but a <code>while(true)</code> loop with a 1237 * <code>break</code> statement is not really all that much easier 1238 * to read or understand than a <code>goto</code>. 1242 * In the end, one might simply agree that <i>in general</i> 1243 * <code>goto</code> statements are a bad idea but be pragmatic 1244 * and state that there may be occasions where they can help avoid 1245 * code duplication and awkward control flow. This may be one of these 1254 * <a name="Thecodemaincodefunction"></a> 1255 * <h3>The <code>main</code> function</h3> 1259 * Having made it this far, there is, again, nothing 1260 * much to discuss for the main function of this 1261 * program: it looks like all such functions since @ref step_6 "step-6". 1268 * using namespace Step26; 1270 * HeatEquation<2> heat_equation_solver; 1271 * heat_equation_solver.run(); 1273 * catch (std::exception &exc) 1275 * std::cerr << std::endl 1277 * << "----------------------------------------------------" 1279 * std::cerr << "Exception on processing: " << std::endl 1280 * << exc.what() << std::endl 1281 * << "Aborting!" << std::endl 1282 * << "----------------------------------------------------" 1289 * std::cerr << std::endl 1291 * << "----------------------------------------------------" 1293 * std::cerr << "Unknown exception!" << std::endl 1294 * << "Aborting!" << std::endl 1295 * << "----------------------------------------------------" 1303 <a name="Results"></a><h1>Results</h1> 1306 As in many of the tutorials, the actual output of the program matters less 1307 than how we arrived there. Nonetheless, here it is: 1309 =========================================== 1310 Number of active cells: 48 1311 Number of degrees of freedom: 65 1313 Time step 1 at t=0.002 1316 =========================================== 1317 Number of active cells: 60 1318 Number of degrees of freedom: 81 1321 Time step 1 at t=0.002 1324 =========================================== 1325 Number of active cells: 105 1326 Number of degrees of freedom: 136 1329 Time step 1 at t=0.002 1334 Time step 249 at t=0.498 1336 Time step 250 at t=0.5 1339 =========================================== 1340 Number of active cells: 1803 1341 Number of degrees of freedom: 2109 1344 Maybe of more interest is a visualization of the solution and the mesh on which 1347 <img src="https://www.dealii.org/images/steps/developer/step-26.movie.gif" alt="Animation of the solution of step 26."> 1349 The movie shows how the two sources switch on and off and how the mesh reacts 1350 to this. It is quite obvious that the mesh as is is probably not the best we 1351 could come up with. We'll
get back to
this in the next section.
1354 <a name=
"extensions"></a>
1355 <a name=
"Possibilitiesforextensions"></a><h3>Possibilities
for extensions</h3>
1358 There are at least two areas where
one can improve
this program significantly:
1359 adaptive time stepping and a better choice of the mesh.
1361 <a name=
"Adaptivetimestepping"></a><h4>Adaptive time stepping</h4>
1364 Having chosen an implicit time stepping scheme, we are not bound by any
1365 CFL-like condition on the time step. Furthermore, because the time scales on
1366 which change happens on a given cell in the heat equation are not bound to the
1367 cells
diameter (unlike the
case with the wave equation, where we had a fixed
1368 speed of information transport that couples the temporal and spatial scales),
1369 we can choose the time step as we please. Or, better, choose it as we deem
1370 necessary
for accuracy.
1372 Looking at the solution, it is clear that the action does not happen uniformly
1373 over time: a lot is changing around the time we
switch on a source, things
1374 become less dramatic once a source is on
for a little
while, and we enter a
1375 long phase of decline when both sources are off. During these times, we could
1376 surely
get away with a larger time step than before without sacrificing too
1379 The literature has many suggestions on how to choose the time step size
1380 adaptively. Much can be learned,
for example, from the way ODE solvers choose
1381 their time steps. One can also be inspired by a posteriori error estimators
1382 that can, ideally, be written in a way that the consist of a temporal and a
1383 spatial contribution to the overall error. If the temporal
one is too large,
1384 we should choose a smaller time step. Ideas in
this direction can be found,
1385 for example, in the PhD thesis of a former principal developer of deal.II,
1386 Ralf Hartmann, published by the University of Heidelberg, Germany, in 2002.
1389 <a name=
"Bettertimesteppingmethods"></a><h4>Better time stepping methods</h4>
1392 We here use
one of the simpler time stepping methods, namely the
second order
1393 in time Crank-Nicolson method. However, more accurate methods such as
1394 Runge-Kutta methods are available and should be used as they
do not represent
1395 much additional effort. It is not difficult to implement
this for the current
1396 program, but a more systematic treatment is also given in @ref step_52
"step-52".
1399 <a name=
"Betterrefinementcriteria"></a><h4>Better refinement criteria</h4>
1402 If you look at the meshes in the movie above, it is clear that they are not
1403 particularly well suited to the task at hand. In fact, they look rather
1406 There are two factors at play. First, there are some islands where cells
1407 have been refined but that are surrounded by non-refined cells (and there
1408 are probably also a few occasional coarsened islands). These are not terrible,
1409 as they most of the time
do not affect the approximation quality of the mesh,
1410 but they also don
't help because so many of their additional degrees of 1411 freedom are in fact constrained by hanging node constraints. That said, 1412 this is easy to fix: the Triangulation class takes an argument to its 1413 constructor indicating a level of "mesh smoothing". Passing one of many 1414 possible flags, this instructs the triangulation to refine some additional 1415 cells, or not to refine some cells, so that the resulting mesh does not have 1418 The second problem is more severe: the mesh appears to lag the solution. 1419 The underlying reason is that we only adapt the mesh once every fifth 1420 time step, and only allow for a single refinement in these cases. Whenever a 1421 source switches on, the solution had been very smooth in this area before and 1422 the mesh was consequently rather coarse. This implies that the next time step 1423 when we refine the mesh, we will get one refinement level more in this area, 1424 and five time steps later another level, etc. But this is not enough: first, 1425 we should refine immediately when a source switches on (after all, in the 1426 current context we at least know what the right hand side is), and we should 1427 allow for more than one refinement level. Of course, all of this can be done 1428 using deal.II, it just requires a bit of algorithmic thinking in how to make 1432 <a name="Positivitypreservation"></a><h4>Positivity preservation</h4> 1435 To increase the accuracy and resolution of your simulation in time, one 1436 typically decreases the time step size @f$k_n@f$. If you start playing around 1437 with the time step in this particular example, you will notice that the 1438 solution becomes partly negative, if @f$k_n@f$ is below a certain threshold. 1439 This is not what we would expect to happen (in nature). 1441 To get an idea of this behavior mathematically, let us consider a general, 1442 fully discrete problem: 1444 A u^{n} = B u^{n-1}. 1446 The general form of the @f$i@f$th equation then reads: 1448 a_{ii} u^{n}_i &= b_{ii} u^{n-1}_i + 1449 \sum\limits_{j \in S_i} \left( b_{ij} u^{n-1}_j - a_{ij} u^{n}_j \right), 1451 where @f$S_i@f$ is the set of degrees of freedom that DoF @f$i@f$ couples with (i.e., 1452 for which either the matrix @f$A@f$ or matrix @f$B@f$ has a nonzero entry at position 1453 @f$(i,j)@f$). If all coefficients 1454 fulfill the following conditions: 1456 a_{ii} &> 0, & b_{ii} &\geq 0, & a_{ij} &\leq 0, & b_{ij} &\geq 0, 1460 all solutions @f$u^{n}@f$ keep their sign from the previous ones @f$u^{n-1}@f$, and 1461 consequently from the initial values @f$u^0@f$. See e.g. 1462 <a href="http://bookstore.siam.org/cs14/">Kuzmin, Hämäläinen</a> 1463 for more information on positivity preservation. 1465 Depending on the PDE to solve and the time integration scheme used, one is 1466 able to deduce conditions for the time step @f$k_n@f$. For the heat equation with 1467 the Crank-Nicolson scheme, 1468 <a href="https://doi.org/10.2478/cmam-2010-0025">Schatz et. al.</a> have 1469 translated it to the following ones: 1471 (1 - \theta) k a_{ii} &\leq m_{ii},\qquad \forall i, 1473 \theta k \left| a_{ij} \right| &\geq m_{ij},\qquad j \neq i, 1475 where @f$M = m_{ij}@f$ denotes the mass matrix and @f$A = a_{ij}@f$ the stiffness 1476 matrix with @f$a_{ij} \leq 0@f$ for @f$j \neq i@f$, respectively. With 1477 @f$a_{ij} \leq 0@f$, we can formulate bounds for the global time step @f$k@f$ as 1480 k_{\text{max}} &= \frac{ 1 }{ 1 - \theta } 1481 \min\left( \frac{ m_{ii} }{ a_{ii} } \right),~ \forall i, 1483 k_{\text{min}} &= \frac{ 1 }{ \theta } 1484 \max\left( \frac{ m_{ij} }{ \left|a_{ij}\right| } \right),~ j \neq i. 1486 In other words, the time step is constrained by <i>both a lower 1487 and upper bound</i> in case of a Crank-Nicolson scheme. These bounds should be 1488 considered along with the CFL condition to ensure significance of the performed 1491 Being unable to make the time step as small as we want to get more 1492 accuracy without losing the positivity property is annoying. It raises 1493 the question of whether we can at least <i>compute</i> the minimal time step 1494 we can choose to ensure positivity preservation in this particular tutorial. 1496 the SparseMatrix objects for both mass and stiffness that are created via 1497 the MatrixCreator functions. Iterating through each entry via SparseMatrixIterators 1498 lets us check for diagonal and off-diagonal entries to set a proper time step 1499 dynamically. For quadratic matrices, the diagonal element is stored as the 1500 first member of a row (see SparseMatrix documentation). An exemplary code 1501 snippet on how to grab the entries of interest from the <code>mass_matrix</code> 1505 Assert (mass_matrix.m() == mass_matrix.n(), ExcNotQuadratic()); 1506 const unsigned int num_rows = mass_matrix.m(); 1507 double mass_matrix_min_diag = std::numeric_limits<double>::max(), 1508 mass_matrix_max_offdiag = 0.; 1510 SparseMatrixIterators::Iterator<double,true> row_it (&mass_matrix, 0); 1512 for(unsigned int m = 0; m<num_rows; ++m) 1514 // check the diagonal element 1515 row_it = mass_matrix.begin(m); 1516 mass_matrix_min_diag = std::min(row_it->value(), mass_matrix_min_diag); 1519 // check the off-diagonal elements 1520 for(; row_it != mass_matrix.end(m); ++row_it) 1521 mass_matrix_max_offdiag = std::max(row_it->value(), mass_matrix_max_offdiag); 1525 Using the information so computed, we can bound the time step via the formulas 1529 <a name="PlainProg"></a> 1530 <h1> The plain program</h1> 1531 @include "step-26.cc" void loop(ITERATOR begin, typename identity< ITERATOR >::type end, DOFINFO &dinfo, INFOBOX &info, const std::function< void(DOFINFO &, typename INFOBOX::CellInfo &)> &cell_worker, const std::function< void(DOFINFO &, typename INFOBOX::CellInfo &)> &boundary_worker, const std::function< void(DOFINFO &, DOFINFO &, typename INFOBOX::CellInfo &, typename INFOBOX::CellInfo &)> &face_worker, ASSEMBLER &assembler, const LoopControl &lctrl=LoopControl())
Contents is actually a matrix.
static const types::blas_int one
static ::ExceptionBase & ExcIndexRange(int arg1, int arg2, int arg3)
Tensor< 2, dim, Number > F(const Tensor< 2, dim, Number > &Grad_u)
void hyper_L(Triangulation< dim > &tria, const double left=-1., const double right=1., const bool colorize=false)
constexpr SymmetricTensor< 2, dim, Number > invert(const SymmetricTensor< 2, dim, Number > &t)
void add(const std::vector< size_type > &indices, const std::vector< OtherNumber > &values)
void random(DoFHandlerType &dof_handler)
void mass_matrix(FullMatrix< double > &M, const FEValuesBase< dim > &fe, const double factor=1.)
#define Assert(cond, exc)
VectorType::value_type * end(VectorType &V)
virtual void reinit(const size_type N, const bool omit_zeroing_entries=false)
void run(const std::vector< std::vector< Iterator >> &colored_iterators, Worker worker, Copier copier, const ScratchData &sample_scratch_data, const CopyData &sample_copy_data, const unsigned int queue_length=2 *MultithreadInfo::n_threads(), const unsigned int chunk_size=8)
const ::parallel::distributed::Triangulation< dim, spacedim > * triangulation
void refine(Triangulation< dim, spacedim > &tria, const Vector< Number > &criteria, const double threshold, const unsigned int max_to_mark=numbers::invalid_unsigned_int)