v0.15.5
Loading...
Searching...
No Matches
SetUpSchurImpl.cpp
Go to the documentation of this file.
1/** @file SetUpSchurImpl
2 * @brief
3 * @date 2023-05-13
4 *
5 * @license{This project is released under the MIT License.}
6 *
7 */
8
10
12 : SetUpSchur(), mField(m_field), epCorePtr(ep_core_ptr) {}
13 virtual ~SetUpSchurImpl() {}
14
18
19private:
22
25
26 boost::shared_ptr<std::vector<boost::weak_ptr<NumeredDofEntity>>>
27 piolaZeroDofsVec; //< Dofs on crack surface
28 boost::shared_ptr<std::vector<unsigned char>>
29 piolaZeroDofsMarker; //< marker for crack dofs on surface
30
50
51 boost::shared_ptr<P_MultiGridData> pMGPtr;
52
54 std::vector<std::string> schur_field_list{epCorePtr->hybridSpatialDisp,
56 std::vector<boost::shared_ptr<Range>> dm_range_list{nullptr, nullptr};
57 return std::make_pair(schur_field_list, dm_range_list);
58 };
59
60 auto getA00Fields() {
61 std::vector<std::string> a00_field_list{
62
64
66
68
70
72
73 };
74 std::vector<boost::shared_ptr<Range>> range_list_ptr(a00_field_list.size(),
75 nullptr);
76 return std::make_pair(a00_field_list, range_list_ptr);
77 }
78};
79
82
83 auto create_schur_dm = [&](SmartPetscObj<DM> &dm_sub) {
85
86 dm_sub = createDM(mField.get_comm(), "DMMOFEM_MG");
87 CHKERR DMMoFEMCreateSubDM(dm_sub, epCorePtr->dmElastic, "SUB_SCHUR");
88 CHKERR DMMoFEMSetSquareProblem(dm_sub, PETSC_TRUE);
89 CHKERR DMMoFEMSetIsPartitioned(dm_sub, PETSC_TRUE);
92
93 int r_idx = 0;
94 auto [schur_field_list, schur_range_list] = getSchurFields();
95 for (auto f : schur_field_list) {
96 MOFEM_LOG("EP", Sev::inform) << "Add schur field: " << f;
97 CHKERR DMMoFEMAddSubFieldRow(dm_sub, f, schur_range_list[r_idx]);
98 CHKERR DMMoFEMAddSubFieldCol(dm_sub, f, schur_range_list[r_idx]);
99 ++r_idx;
100 }
101 CHKERR DMSetUp(dm_sub);
103 };
104
105 auto create_a00_dm = [&](SmartPetscObj<DM> &dm_sub) {
107 dm_sub = createDM(mField.get_comm(), "DMMOFEM");
108 CHKERR DMMoFEMCreateSubDM(dm_sub, epCorePtr->dmElastic, "SUB_A00");
109 CHKERR DMMoFEMSetSquareProblem(dm_sub, PETSC_TRUE);
110 CHKERR DMMoFEMSetIsPartitioned(dm_sub, PETSC_TRUE);
114
115 int r_idx = 0;
116 auto [a00_field_list, a00_range_list] = getA00Fields();
117 for (auto f : a00_field_list) {
118 MOFEM_LOG("EP", Sev::inform) << "Add a00 field: " << f;
119 CHKERR DMMoFEMAddSubFieldRow(dm_sub, f, a00_range_list[r_idx]);
120 CHKERR DMMoFEMAddSubFieldCol(dm_sub, f, a00_range_list[r_idx]);
121 ++r_idx;
122 }
123 CHKERR DMSetUp(dm_sub);
125 };
126
127 auto get_snes = [&](TS ts) {
128 SNES snes;
129 CHKERR TSGetSNES(ts, &snes);
130 return snes;
131 };
132
133 auto get_ksp = [&](SNES snes) {
134 KSP ksp;
135 CHKERR SNESGetKSP(snes, &ksp);
136 CHKERR KSPSetFromOptions(ksp);
137 return ksp;
138 };
139
140 auto get_pc = [&](KSP ksp) {
141 PC pc;
142 CHKERR KSPGetPC(ksp, &pc);
143 return pc;
144 };
145
146 auto ksp = get_ksp(get_snes(ts));
147 auto pc = get_pc(ksp);
148
149 PetscBool is_pcfs = PETSC_FALSE;
150 PetscObjectTypeCompare((PetscObject)pc, PCFIELDSPLIT, &is_pcfs);
151 if (is_pcfs) {
152
153 MOFEM_LOG("EP", Sev::inform) << "SetUpSchurImpl::setUp: PCFIELDSPLIT";
154
155 SmartPetscObj<DM> schur_dm, a00_dm;
156 CHKERR create_schur_dm(schur_dm);
157 CHKERR create_a00_dm(a00_dm);
158
159 auto dm_elastic = epCorePtr->dmElastic;
160 auto vol_elem_name = epCorePtr->elementVolumeName;
161 auto skel_elem_name = epCorePtr->skeletonElement;
162 auto contact_elem_name = epCorePtr->contactElement;
163 auto natural_bc_element_name = epCorePtr->naturalBcElement;
164
165 std::vector<std::pair<std::string, std::string>> mat_block_list = {
166
173
181
182 };
183
184 if (epCorePtr->noStretch) {
185 mat_block_list.push_back(
187 mat_block_list.push_back(
189 mat_block_list.push_back(
191 }
192
193 mat_block_list.push_back({epCorePtr->rotAxis, epCorePtr->rotAxis});
194 mat_block_list.push_back({epCorePtr->stretchTensor, epCorePtr->rotAxis});
195 mat_block_list.push_back({epCorePtr->rotAxis, epCorePtr->stretchTensor});
196
197 auto get_nested_mat_data = [&](auto schur_dm, auto block_dm) {
198 auto block_mat_data = createBlockMatStructure(
199 dm_elastic,
200
201 {
202
203 {vol_elem_name, mat_block_list},
204
205 {skel_elem_name,
206
207 {
208
214
215 }},
216
217 {contact_elem_name,
218
219 {
220
226
227 }},
228
229 {natural_bc_element_name,
230
231 {
232
236
237 }}
238
239 }
240
241 );
242
243 auto [a00_field_list, a00_range_list] = getA00Fields();
244
246
247 {schur_dm, a00_dm}, block_mat_data,
248
249 a00_field_list,
250
251 a00_range_list,
252
253 false
254
255 );
256 };
257
258 auto nested_mat_data = get_nested_mat_data(schur_dm, a00_dm);
260 CHKERR DMSetMatType(epCorePtr->dmElastic, MATSHELL);
261
264
265 if (std::abs(epCorePtr->alphaRho) >
266 std::numeric_limits<double>::epsilon()) {
267 auto swap_assemble = [](TS ts, PetscReal t, Vec u, Vec u_t, Vec utt,
268 PetscReal a, PetscReal aa, Mat A, Mat B,
269 void *ctx) {
270 return TsSetI2Jacobian(ts, t, u, u_t, utt, a, aa, B, A, ctx);
271 };
272 auto ts_ctx_ptr = getDMTsCtx(epCorePtr->dmElastic);
273 CHKERR TSSetI2Jacobian(ts, m, p, swap_assemble, ts_ctx_ptr.get());
274 } else {
275 auto swap_assemble = [](TS ts, PetscReal t, Vec u, Vec u_t, PetscReal a,
276 Mat A, Mat B, void *ctx) {
277 return TsSetIJacobian(ts, t, u, u_t, a, B, A, ctx);
278 };
279 auto ts_ctx_ptr = getDMTsCtx(epCorePtr->dmElastic);
280 CHKERR TSSetIJacobian(ts, m, p, swap_assemble, ts_ctx_ptr.get());
281 }
282 CHKERR KSPSetOperators(ksp, m, p);
283
284 auto set_assembly = [&]() {
286
287 aoS = getDMSubData(schur_dm)->getSmartRowMap();
288 S = createDMHybridisedL2Matrix(schur_dm);
289 CHKERR MatSetBlockSize(S, SPACE_DIM *
291 epCorePtr->S = S;
292 epCorePtr->aoS = aoS;
293
294 auto set_assemble = [&]() {
296 auto schur_asmb_pre_proc_lhs = boost::make_shared<FEMethod>();
297 auto schur_asmb_pre_proc_rhs = boost::make_shared<FEMethod>();
298
299 schur_asmb_pre_proc_lhs->preProcessHook = [this]() {
301 CHKERR MatZeroEntries(S);
303 };
304
305 auto *schur_asmb_pre_proc_rhs_ptr = schur_asmb_pre_proc_rhs.get();
306 schur_asmb_pre_proc_rhs->preProcessHook =
307 [this, schur_asmb_pre_proc_rhs_ptr]() {
309 auto prb_ptr = schur_asmb_pre_proc_rhs_ptr->problemPtr;
310 auto dofs_prb = prb_ptr->getNumeredRowDofsPtr();
311
312 auto crack_faces = epCorePtr->crackFaces;
313 piolaZeroDofsVec->clear();
314 CHKERR mField.getInterface<ProblemsManager>()
315 ->getSideDofsOnBrokenSpaceEntities(
316 *piolaZeroDofsVec, prb_ptr->getName(), ROW,
317 epCorePtr->piolaStress, *crack_faces, SPACE_DIM, 0,
318 SPACE_DIM);
319
320 piolaZeroDofsMarker->resize(dofs_prb->size(), 0);
321 piolaZeroDofsMarker->clear();
322 for (auto &dof : *piolaZeroDofsVec) {
323 if (auto dof_ptr = dof.lock()) {
324 auto idx = dof_ptr->getPetscLocalDofIdx();
325 (*piolaZeroDofsMarker)[idx] = 1;
326 }
327 }
328 // FIXME: This needs investigation, true is required for fracture
329 constexpr bool hard_coded_set_bc_debug = true;
330 if constexpr (hard_coded_set_bc_debug) {
331
332 auto problem_name =
333 schur_asmb_pre_proc_rhs_ptr->problemPtr->getName();
334 auto crack_faces = epCorePtr->crackFaces;
335
336 SmartPetscObj<IS> crack_hybrid_is;
337 CHKERR epCorePtr->mField.getInterface<ISManager>()
338 ->isCreateProblemFieldAndRankLocal(
339 problem_name, ROW, epCorePtr->hybridSpatialDisp, 0,
340 SPACE_DIM, crack_hybrid_is, &*crack_faces);
341
342 SmartPetscObj<IS> crack_piola_is;
343 CHKERR epCorePtr->mField.getInterface<ISManager>()
344 ->isCreateProblemBrokenFieldAndRankLocal(*piolaZeroDofsVec,
345 crack_piola_is);
346
347 const double *a_x;
348 CHKERR VecGetArrayRead(schur_asmb_pre_proc_rhs_ptr->x, &a_x);
349 auto zero_by_is = [&](auto is) {
351 const PetscInt *is_array;
352 PetscInt is_size;
353 CHKERR ISGetLocalSize(is, &is_size);
354 CHKERR ISGetIndices(is, &is_array);
355 for (int i = 0; i != is_size; ++i) {
356 // FIXME: That is irregular, need investigate if needed
357 const_cast<double *>(a_x)[is_array[i]] = 0;
358 }
359 CHKERR ISRestoreIndices(is, &is_array);
361 };
362
363 CHKERR zero_by_is(crack_hybrid_is);
364 CHKERR zero_by_is(crack_piola_is);
365
366 CHKERR VecRestoreArrayRead(schur_asmb_pre_proc_rhs_ptr->x,
367 &a_x);
368
369 CHKERR epCorePtr->mField.getInterface<VecManager>()
370 ->setLocalGhostVector(problem_name, COL,
371 schur_asmb_pre_proc_rhs_ptr->x,
372 INSERT_VALUES, SCATTER_REVERSE);
373 }
374
376 };
377
378 auto schur_asmb_post_proc_lhs = boost::make_shared<FEMethod>();
379 auto schur_asmb_post_proc_rhs = boost::make_shared<FEMethod>();
380
381 auto schur_asmb_post_proc_rhs_weak_ptr =
382 boost::weak_ptr<FEMethod>(schur_asmb_post_proc_rhs);
383 schur_asmb_post_proc_rhs->postProcessHook =
384 [this, schur_asmb_post_proc_rhs_weak_ptr]() {
386
387 if (auto schur_asmb_post_proc_rhs_ptr =
388 schur_asmb_post_proc_rhs_weak_ptr.lock()) {
389
390 CHKERR VecGhostUpdateBegin(schur_asmb_post_proc_rhs_ptr->f,
391 ADD_VALUES, SCATTER_REVERSE);
392 CHKERR VecGhostUpdateEnd(schur_asmb_post_proc_rhs_ptr->f,
393 ADD_VALUES, SCATTER_REVERSE);
394 CHKERR VecAssemblyBegin(schur_asmb_post_proc_rhs_ptr->f);
395 CHKERR VecAssemblyEnd(schur_asmb_post_proc_rhs_ptr->f);
396 *(schur_asmb_post_proc_rhs_ptr->vecAssembleSwitch) = false;
397
398 {
399
400 auto problem_name =
401 schur_asmb_post_proc_rhs_ptr->problemPtr->getName();
402
403 auto crack_faces = epCorePtr->crackFaces;
404
405 SmartPetscObj<IS> crack_hybrid_is;
406 CHKERR epCorePtr->mField.getInterface<ISManager>()
407 ->isCreateProblemFieldAndRankLocal(
408 problem_name, ROW, epCorePtr->hybridSpatialDisp, 0,
409 SPACE_DIM, crack_hybrid_is, &*crack_faces);
410
411 SmartPetscObj<IS> crack_piola_is;
412 CHKERR epCorePtr->mField.getInterface<ISManager>()
413 ->isCreateProblemBrokenFieldAndRankLocal(
414 *piolaZeroDofsVec, crack_piola_is);
415
416 double *a_f;
417 CHKERR VecGetArray(schur_asmb_post_proc_rhs_ptr->f, &a_f);
418 const double *a_x;
419 CHKERR VecGetArrayRead(schur_asmb_post_proc_rhs_ptr->x, &a_x);
420 auto zero_by_is = [&](auto is) {
422 const PetscInt *is_array;
423 PetscInt is_size;
424 CHKERR ISGetLocalSize(is, &is_size);
425 CHKERR ISGetIndices(is, &is_array);
426 for (int i = 0; i != is_size; ++i) {
427 a_f[is_array[i]] = -a_x[is_array[i]];
428 }
429 CHKERR ISRestoreIndices(is, &is_array);
431 };
432
433 CHKERR zero_by_is(crack_hybrid_is);
434 CHKERR zero_by_is(crack_piola_is);
435
436 CHKERR VecRestoreArray(schur_asmb_post_proc_rhs_ptr->f, &a_f);
437 CHKERR VecRestoreArrayRead(schur_asmb_post_proc_rhs_ptr->x,
438 &a_x);
439 }
440
441 } else {
442 SETERRQ(PETSC_COMM_SELF, PETSC_ERR_PLIB,
443 "schur_asmb_post_proc_rhs is expired");
444 }
445
447 };
448
449 auto schur_asmb_post_proc_lhs_weak_ptr =
450 boost::weak_ptr<FEMethod>(schur_asmb_post_proc_lhs);
451 schur_asmb_post_proc_lhs->postProcessHook =
452 [this, schur_asmb_post_proc_lhs_weak_ptr]() {
454
455 if (auto schur_asmb_post_proc_lhs_ptr =
456 schur_asmb_post_proc_lhs_weak_ptr.lock()) {
457
458 if (pMGPtr) {
459 CHKERR pMGPtr->setUP();
460 pMGPtr.reset();
461 }
462
463 auto crack_faces = epCorePtr->crackFaces;
464
465 // Assemble matrix
466 CHKERR MatAssemblyBegin(schur_asmb_post_proc_lhs_ptr->B,
467 MAT_FINAL_ASSEMBLY);
468 CHKERR MatAssemblyEnd(schur_asmb_post_proc_lhs_ptr->B,
469 MAT_FINAL_ASSEMBLY);
470 *(schur_asmb_post_proc_lhs_ptr->matAssembleSwitch) = false;
471 {
472 SmartPetscObj<IS> crack_hybrid_is;
473 CHKERR epCorePtr->mField.getInterface<ISManager>()
474 ->isCreateProblemFieldAndRank(
475 "ELASTIC_PROBLEM", ROW, epCorePtr->hybridSpatialDisp,
476 0, SPACE_DIM, crack_hybrid_is, &*crack_faces);
477 CHKERR MatZeroRowsColumnsIS(schur_asmb_post_proc_lhs_ptr->B,
478 crack_hybrid_is, 1, PETSC_NULLPTR,
479 PETSC_NULLPTR);
480 }
481 {
482 SmartPetscObj<IS> crack_piola_is;
483 CHKERR epCorePtr->mField.getInterface<ISManager>()
484 ->isCreateProblemBrokenFieldAndRank(*piolaZeroDofsVec,
485 crack_piola_is);
486 CHKERR MatZeroRowsColumnsIS(schur_asmb_post_proc_lhs_ptr->B,
487 crack_piola_is, 1, PETSC_NULLPTR,
488 PETSC_NULLPTR);
489 }
490
491 auto [a00_field_list, a00_range_list] = getA00Fields();
493 epCorePtr->mField, schur_asmb_post_proc_lhs_ptr->B, S,
494 a00_field_list, a00_range_list, aoS);
495 epCorePtr->a00FieldList = a00_field_list;
496 epCorePtr->a00RangeList = a00_range_list;
497
498 // Apply essential constrains to Schur complement
499 CHKERR MatAssemblyBegin(S, MAT_FINAL_ASSEMBLY);
500 CHKERR MatAssemblyEnd(S, MAT_FINAL_ASSEMBLY);
501
502 SmartPetscObj<IS> crack_hybrid_is;
503 CHKERR epCorePtr->mField.getInterface<ISManager>()
504 ->isCreateProblemFieldAndRank(
505 "SUB_SCHUR", ROW, epCorePtr->hybridSpatialDisp, 0,
506 SPACE_DIM, crack_hybrid_is, &*crack_faces);
507 epCorePtr->crackHybridIs = crack_hybrid_is;
508 CHKERR MatZeroRowsColumnsIS(S, crack_hybrid_is, 1,
509 PETSC_NULLPTR, PETSC_NULLPTR);
510
511 } else {
512 SETERRQ(PETSC_COMM_SELF, PETSC_ERR_PLIB,
513 "schur_asmb_post_proc_lhs is expired");
514 }
515
517 };
518
519 auto ts_ctx_ptr = getDMTsCtx(epCorePtr->dmElastic);
520 ts_ctx_ptr->getPreProcessIFunction().push_front(
521 schur_asmb_pre_proc_rhs);
522 ts_ctx_ptr->getPostProcessIFunction().push_back(
523 schur_asmb_post_proc_rhs);
524 ts_ctx_ptr->getPreProcessIJacobian().push_front(
525 schur_asmb_pre_proc_lhs);
526 ts_ctx_ptr->getPostProcessIJacobian().push_back(
527 schur_asmb_post_proc_lhs);
529 };
530
532 boost::make_shared<std::vector<boost::weak_ptr<NumeredDofEntity>>>();
533 piolaZeroDofsMarker = boost::make_shared<std::vector<unsigned char>>();
534 CHKERR set_assemble();
535
537 };
538
539 auto set_pc = [&]() {
541 auto a00_is = getDMSubData(a00_dm)->getSmartRowIs();
542 auto schur_is = getDMSubData(schur_dm)->getSmartRowIs();
543 CHKERR PCFieldSplitSetIS(pc, NULL, a00_is);
544 CHKERR PCFieldSplitSetIS(pc, NULL, schur_is);
545 CHKERR PCFieldSplitSetSchurPre(pc, PC_FIELDSPLIT_SCHUR_PRE_USER, S);
547 };
548
549 auto set_diagonal_pc = [&]() {
551 KSP *subksp;
552 CHKERR PCFieldSplitSchurGetSubKSP(pc, PETSC_NULLPTR, &subksp);
553 auto get_pc = [](auto ksp) {
554 PC pc_raw;
555 CHKERR KSPGetPC(ksp, &pc_raw);
556 return SmartPetscObj<PC>(pc_raw, true); // bump reference
557 };
558 CHKERR setSchurA00MatSolvePC(get_pc(subksp[0]));
559
560 auto set_pc_p_mg = [&](auto dm, auto pc, auto S) {
562 CHKERR PCSetDM(pc, dm);
563 PetscBool same = PETSC_FALSE;
564 PetscObjectTypeCompare((PetscObject)pc, PCMG, &same);
565 if (same) {
566 auto smart_pc = SmartPetscObj<PC>(pc, true);
567 pMGPtr = boost::make_shared<P_MultiGridData>(dm, smart_pc, S);
568 }
569 PetscObjectTypeCompare((PetscObject)pc, PCKSP, &same);
570 if (same) {
571 MOFEM_LOG("EP", Sev::inform)
572 << "SetUpSchurImpl::setUp: fieldsplit 1 PCKSP";
573 CHKERR PCSetFromOptions(pc);
574 KSP ksp;
575 CHKERR PCKSPGetKSP(pc, &ksp);
576 CHKERR KSPSetFromOptions(ksp);
577 PC ksp_pc;
578 CHKERR KSPGetPC(ksp, &ksp_pc);
579 CHKERR PCSetFromOptions(ksp_pc);
580 PetscObjectTypeCompare((PetscObject)ksp_pc, PCMG, &same);
581 if (same) {
582 auto smart_pc = SmartPetscObj<PC>(ksp_pc, true);
583 pMGPtr = boost::make_shared<P_MultiGridData>(dm, smart_pc, S);
584 }
585 }
587 };
588
589 CHKERR set_pc_p_mg(schur_dm, get_pc(subksp[1]), S);
590
591 CHKERR PetscFree(subksp);
593 };
594
595 CHKERR set_assembly();
596 CHKERR set_pc();
597 CHKERR TSSetUp(ts);
598 CHKERR KSPSetUp(ksp);
599 CHKERR set_diagonal_pc();
600
601 } else {
602 MOFEM_LOG("EP", Sev::inform) << "SetUpSchurImpl::setUp: PCLU or other";
603
604 epCorePtr->elasticFeLhs->getOpPtrVector().push_front(
606 epCorePtr->elasticFeLhs->getOpPtrVector().push_back(
608 epCorePtr->elasticBcLhs->getOpPtrVector().push_front(
610 epCorePtr->elasticBcLhs->getOpPtrVector().push_back(
612 }
613
615}
616
617boost::shared_ptr<EshelbianCore::SetUpSchur>
619 EshelbianCore *ep_core_ptr) {
620 return boost::shared_ptr<SetUpSchur>(
621 new SetUpSchurImpl(m_field, ep_core_ptr));
622}
constexpr double a
@ COL
@ ROW
#define MoFEMFunctionBegin
First executable line of each MoFEM function, used for error handling. Final line of MoFEM functions ...
#define MoFEMFunctionReturn(a)
Last executable line of each PETSc function used for error handling. Replaces return()
#define CHKERR
Inline error check.
PetscErrorCode DMMoFEMSetIsPartitioned(DM dm, PetscBool is_partitioned)
Definition DMMoFEM.cpp:1113
PetscErrorCode DMMoFEMCreateSubDM(DM subdm, DM dm, const char problem_name[])
Must be called by user to set Sub DM MoFEM data structures.
Definition DMMoFEM.cpp:215
PetscErrorCode DMMoFEMAddElement(DM dm, std::string fe_name)
add element to dm
Definition DMMoFEM.cpp:488
PetscErrorCode DMMoFEMSetSquareProblem(DM dm, PetscBool square_problem)
set squared problem
Definition DMMoFEM.cpp:450
PetscErrorCode DMMoFEMAddSubFieldRow(DM dm, const char field_name[])
Definition DMMoFEM.cpp:238
PetscErrorCode DMMoFEMAddSubFieldCol(DM dm, const char field_name[])
Definition DMMoFEM.cpp:280
#define MOFEM_LOG(channel, severity)
Log.
#define NBFACETRI_L2(P)
Number of base functions on triangle for L2 space.
FTensor::Index< 'i', SPACE_DIM > i
const FTensor::Tensor2< T, Dim, Dim > Vec
PetscErrorCode MoFEMErrorCode
MoFEM/PETSc error code.
PetscErrorCode TsSetIJacobian(TS ts, PetscReal t, Vec u, Vec u_t, PetscReal a, Mat A, Mat B, void *ctx)
Set function evaluating jacobian in TS solver.
Definition TsCtx.cpp:169
auto getDMTsCtx(DM dm)
Get TS context data structure used by DM.
Definition DMMoFEM.hpp:1279
OpSchurAssembleBase * createOpSchurAssembleEnd(std::vector< std::string > fields_name, std::vector< boost::shared_ptr< Range > > field_ents, SmartPetscObj< AO > ao, SmartPetscObj< Mat > schur, bool sym_schur, bool symm_op)
Construct a new Op Schur Assemble End object.
Definition Schur.cpp:2663
MoFEMErrorCode setSchurA00MatSolvePC(SmartPetscObj< PC > pc)
Set PC for A00 block.
Definition Schur.cpp:2705
boost::shared_ptr< PCMGSetUpViaApproxOrdersCtx > createPCMGSetUpViaApproxOrdersCtx(DM dm, Mat A, bool use_shell_mat)
createPCMGSetUpViaApproxOrdersCtx
auto createDMHybridisedL2Matrix(DM dm)
Get smart hybridised L2 matrix from DM.
Definition DMMoFEM.hpp:1207
auto getDMSubData(DM dm)
Get sub problem data structure.
Definition DMMoFEM.hpp:1295
PetscErrorCode TsSetI2Jacobian(TS ts, PetscReal t, Vec u, Vec u_t, Vec u_tt, PetscReal a, PetscReal aa, Mat A, Mat B, void *ctx)
Calculation Jacobian for second order PDE in time.
Definition TsCtx.cpp:519
MoFEMErrorCode PCMGSetUpViaApproxOrders(PC pc, boost::shared_ptr< PCMGSetUpViaApproxOrdersCtx > ctx, int verb)
Function build MG structure.
boost::shared_ptr< BlockStructure > createBlockMatStructure(DM dm, SchurFEOpsFEandFields schur_fe_op_vec)
Create a Mat Diag Blocks object.
Definition Schur.cpp:1082
boost::shared_ptr< NestSchurData > createSchurNestedMatrixStruture(std::pair< SmartPetscObj< DM >, SmartPetscObj< DM > > dms, boost::shared_ptr< BlockStructure > block_mat_data_ptr, std::vector< std::string > fields_names, std::vector< boost::shared_ptr< Range > > field_ents, bool add_preconditioner_block)
Get the Schur Nest Mat Array object.
Definition Schur.cpp:2421
MoFEMErrorCode DMMoFEMSetNestSchurData(DM dm, boost::shared_ptr< NestSchurData >)
Definition DMMoFEM.cpp:1555
auto createDMNestSchurMat(DM dm)
Definition DMMoFEM.hpp:1221
auto createDM(MPI_Comm comm, const std::string dm_type_name)
Creates smart DM object.
MoFEMErrorCode assembleBlockMatSchur(MoFEM::Interface &m_field, Mat B, Mat S, std::vector< std::string > fields_name, std::vector< boost::shared_ptr< Range > > field_ents, SmartPetscObj< AO > ao)
Assemble Schur matrix.
Definition Schur.cpp:1895
OpSchurAssembleBase * createOpSchurAssembleBegin()
Definition Schur.cpp:2658
auto createDMBlockMat(DM dm)
Definition DMMoFEM.hpp:1214
constexpr double t
plate stiffness
Definition plate.cpp:58
FTensor::Index< 'm', 3 > m
static boost::shared_ptr< SetUpSchur > createSetUpSchur(MoFEM::Interface &m_field, EshelbianCore *ep_core_ptr)
std::vector< boost::shared_ptr< Range > > a00RangeList
const std::string skeletonElement
MoFEM::Interface & mField
const std::string spatialL2Disp
SmartPetscObj< IS > crackHybridIs
boost::shared_ptr< FaceElementForcesAndSourcesCore > elasticBcLhs
const std::string elementVolumeName
const std::string piolaStress
std::vector< std::string > a00FieldList
const std::string bubbleField
boost::shared_ptr< VolumeElementForcesAndSourcesCore > elasticFeLhs
static PetscBool noStretch
const std::string rotAxis
const std::string contactDisp
const std::string naturalBcElement
boost::shared_ptr< Range > crackFaces
const std::string hybridSpatialDisp
SmartPetscObj< DM > dmElastic
Elastic problem.
const std::string stretchTensor
const std::string contactElement
virtual MPI_Comm & get_comm() const =0
Deprecated interface functions.
intrusive_ptr for managing petsc objects
MoFEMErrorCode getInterface(IFACE *&iface) const
Get interface reference to pointer of interface.
P_MultiGridData(SmartPetscObj< DM > dm, SmartPetscObj< PC > pc, SmartPetscObj< Mat > S)
EshelbianCore * epCorePtr
boost::shared_ptr< std::vector< unsigned char > > piolaZeroDofsMarker
SmartPetscObj< Mat > S
boost::shared_ptr< std::vector< boost::weak_ptr< NumeredDofEntity > > > piolaZeroDofsVec
boost::shared_ptr< P_MultiGridData > pMGPtr
MoFEMErrorCode setUp(TS ts)
MoFEMErrorCode setUp(SmartPetscObj< KSP >)
virtual ~SetUpSchurImpl()
MoFEMErrorCode postProc()
SmartPetscObj< AO > aoS
MoFEMErrorCode preProc()
MoFEM::Interface & mField
SetUpSchurImpl(MoFEM::Interface &m_field, EshelbianCore *ep_core_ptr)
[Push operators to pipeline]
constexpr int SPACE_DIM
[Define dimension]
Definition elastic.cpp:18
constexpr AssemblyType A
[Define dimension]
Definition elastic.cpp:21