Szczegóły publikacji
Opis bibliograficzny
Fast solver for advection dominated diffusion using residual minimization and neural networks / Tomasz Służalec, Maciej PASZYŃSKI // W: Computational Science – ICCS 2023 : 23rd international conference : Prague, Czech Republic, July 3–5, 2023 : proceedings, Pt. 2 / eds. Jiří Mikyška [et al.]. — Cham, Switzerland : Springer, cop. 2023. — (Lecture Notes in Computer Science ; ISSN 0302-9743 ; LNCS 14074). — ISBN: 978-3-031-36020-6; e-ISBN: 978-3-031-36021-3. — S. 517–531. — Bibliogr., Abstr. — Publikacja dostępna online od: 2023-06-26. — T. Służalec - afiliacja: Jagiellonian University
Autorzy (2)
- Służalec Tomasz
- AGHPaszyński Maciej
Dane bibliometryczne
| ID BaDAP | 147735 |
|---|---|
| Data dodania do BaDAP | 2023-07-20 |
| DOI | 10.1007/978-3-031-36021-3_52 |
| Rok publikacji | 2023 |
| Typ publikacji | materiały konferencyjne (aut.) |
| Otwarty dostęp | |
| Wydawca | Springer |
| Konferencja | International Conference on Computational Science 2023 |
| Czasopismo/seria | Lecture Notes in Computer Science |
Abstract
Advection-dominated diffusion is a challenging computational problem that requires special stabilization efforts. Unfortunately, the numerical solution obtained with the commonly used Galerkin method delivers unexpected oscillation resulting in an inaccurate numerical solution. The theoretical background resulting from the famous inf-sup condition tells us that the finite-dimensional test space employed by the Galerkin method does not allow us to reach the supremum necessary for problem stability. We enlarge the test space to overcome this problem. We do it for a fixed trial space. The method that allows us to do so is the residual minimization method. This method, however, requires the solution to a much larger system of linear equations than the standard Galerkin method. We represent the larger test space by its set of optimal test functions, forming a basis of the same dimension as the trial space in the Galerkin method. The resulting Petrov-Galerkin method stabilizes our challenging advection-dominated problem. We train the optimal test functions offline with the neural network to speed up the computations. We also observe that the optimal test functions, usually global, can be approximated with local support functions, resulting in a low computational cost for the solver and a stable numerical solution.