Suivez

la liste

Context

High-fidelity simulations of turbulent compressible flows in aerodynamics typically require the numerical analysis of three-dimensional flows around complex geometries. In optimization, linear analysis or data assimilation area, typical Computational Fluid Dynamics (CFD) workflows, such as fixed-point iterations or adjoint-state methods, require the solution of large, sparse, non-symmetric and ill-conditioned linear systems. An active area of research focuses on developing parallel, robust and efficient solvers capable of delivering solutions to such systems within a prescribed error tolerance. A key challenge, especially as problem sizes approach billions of unknowns, is the design of effective preconditioning operators. Although the solve phase still account for a substantial portion of the total CPU time in parallel simulations, the number of computing cores is usually dictated by the requirements of the CFD study itself, rather than by the needs of the linear solver.

Objectives

Recent studies have demonstrated the strong performance of hybrid direct-iterative strategies. In these approaches, from the algebraic decomposition of the matrix, a flexible Krylov method is employed with a domain decomposition method as preconditioner and an approximate direct method as subdomain solver. The main objective is now to design a preconditioning operator that fully exploits the memory budget already allocated for the CFD simulation. The first MPI paradigm splits the study domain geometrically into well-balanced partitions. We aim to introduce a second MPI paradigm devoted to linear systems: enlarging subdomains for the local approximate direct solvers may significantly enhance the global numerical efficiency of the approach. To mitigate the computational and memory costs traditionally associated with direct solvers, we rely on the general-purpose multifrontal MUMPS solver that exploits variable precision and possible low-rank property of matrices. Initial numerical experiments using a fixed accuracy Block Low-Rank (BLR) multifrontal factorization as the subdomain solver have already shown promising CPU-time reductions, together with a significant memory compression of the L and U factors compared with a classical full LU factorization. The new HPC capabilities available in MUMPS, combined with variable subdomain sizes, may open further opportunities for performance improvements. In addition, adapting the accuracy of the subdomain solver to the subdomain stiffness may offer further benefits. To preserve scalability under such heterogeneous configurations, we will also investigate load-balancing techniques with weighted graph partitioning.

Desired candidate profile

Master's degree or equivalent with a background in Applied Mathematics or Computer Science. A keen interest in numerical linear algebra and programming (C++, Fortran, Python) would be welcome.

Collaboration

Team PEQUAN (Performance and Quality of Numerical Algorithms) from the laboratory LIP6 at Sorbonne University.

A more detailed description of the subject can be found in the attached PDF file. To apply, please follow the web link and fill the form.