No Arabic abstract
The Reynolds-Averaged Navier-Stokes equations and the Large-Eddy Simulation equations can be coupled using a transition function to switch from a set of equations applied in some areas of a domain to the other set in the other part of the domain. Following this idea, different time integration schemes can be coupled. In this context, we developed a hybrid time integration scheme that spatially couples the explicit scheme of Heun and the implicit scheme of Crank and Nicolson using a dedicated transition function. This scheme is linearly stable and second-order accurate. In this paper, an extension of this hybrid scheme is introduced to deal with a temporal adaptive procedure. The idea is to treat the time integration procedure with unstructured grids as it is performed with Cartesian grids with local mesh refinement. Depending on its characteristic size, each mesh cell is assigned a rank. And for two cells from two consecutive ranks, the ratio of the associated time steps for time marching the solutions is $2$. As a consequence, the cells with the lowest rank iterate more than the other ones to reach the same physical time. In a finite-volume context, a key ingredient is to keep the conservation property for the interfaces that separate two cells of different ranks. After introducing the different schemes, the paper recalls briefly the coupling procedure, and details the extension to the temporal adaptive procedure. The new time integration scheme is validated with the propagation of 1D wave packet, the Sods tube, and the transport of a bi-dimensional vortex in an uniform flow.
This paper addresses how two time integration schemes, the Heuns scheme for explicit time integration and the second-order Crank-Nicolson scheme for implicit time integration, can be coupled spatially. This coupling is the prerequisite to perform a coupled Large Eddy Simulation / Reynolds Averaged Navier-Stokes computation in an industrial context, using the implicit time procedure for the boundary layer (RANS) and the explicit time integration procedure in the LES region. The coupling procedure is designed in order to switch from explicit to implicit time integrations as fast as possible, while maintaining stability. After introducing the different schemes, the paper presents the initial coupling procedure adapted from a published reference and shows that it can amplify some numerical waves. An alternative procedure, studied in a coupled time/space framework, is shown to be stable and with spectral properties in agreement with the requirements of industrial applications. The coupling technique is validated with standard test cases, ranging from one-dimensional to three-dimensional flows.
Within the mode-coupling theory (MCT) of the glass transition, we reconsider the numerical schemes to evaluate the MCT functional. Here we propose nonuniform discretizations of the wave number, in contrast to the standard equidistant grid, in order to decrease the number of grid points without losing accuracy. We discuss in detail how the integration scheme on the new grids has to be modified from standard Riemann integration. We benchmark our approach by solving the MCT equations numerically for mono-disperse hard disks and hard spheres and by computing the critical packing fraction and the nonergodicity parameters. Our results show that significant improvements in performance can be obtained by employing a nonuniform grid.
We introduce a recent symplectic integration scheme derived for solving physically motivated systems with non-separable Hamiltonians. We show its relevance to Riemannian manifold Hamiltonian Monte Carlo (RMHMC) and provide an alternative to the currently used generalised leapfrog symplectic integrator, which relies on solving multiple fixed point iterations to convergence. Via this approach, we are able to reduce the number of higher-order derivative calculations per leapfrog step. We explore the implications of this integrator and demonstrate its efficacy in reducing the computational burden of RMHMC. Our code is provided in a new open-source Python package, hamiltorch.
In this paper, an efficient high-order gas-kinetic scheme (EHGKS) is proposed to solve the Euler equations for compressible flows. We re-investigate the underlying mechanism of the high-order gas-kinetic scheme (HGKS) and find a new strategy to improve its efficiency. The main idea of the new scheme contains two parts. Firstly, inspired by the state-of-art simplifications on the third-order HGKS, we extend the HGKS to the case of arbitrary high-order accuracy and eliminate its unnecessary high-order dissipation terms. Secondly, instead of computing the derivatives of particle distribution function and their complex moments, we introduce a Lax-Wendroff procedure to compute the high-order derivatives of macroscopic quantities directly. The new scheme takes advantage of both HGKS and the Lax-Wendroff procedure, so that it can be easily extended to the case of arbitrary high-order accuracy with practical significance. Typical numerical tests are carried out by EHGKS, with the third, fifth and seventh-order accuracy. The presence of good resolution on the discontinuities and flow details, together with the optimal CFL numbers, validates the high accuracy and strong robustness of EHGKS. To compare the efficiency, we present the results computed by the EHGKS, the original HGKS and Runge-Kutta-WENO-GKS. This further demonstrates the advantages of EHGKS.
This paper presents a Graphics Processing Units (GPUs) acceleration method of an iterative scheme for gas-kinetic model equations. Unlike the previous GPU parallelization of explicit kinetic schemes, this work features a fast converging iterative scheme. The memory reduction techniques in this method enable full three-dimensional (3D) solution of kinetic model equations in contemporary GPUs usually with a limited memory capacity that otherwise would need terabytes of memory. The GPU algorithm is validated against the DSMC simulation of the 3D lid-driven cavity flow and the supersonic rarefied gas flow past a cube with grids size up to 0.7 trillion points in the phase space. The performance of the GPU algorithm is assessed by comparing with the corresponding parallel CPU program using Message Passing Interface (MPI). The profiling on several models of GPUs shows that the algorithm has a medium to high level of utilization of the GPUs computing and memory resources. A $190times$ speedup can be achieved on the Tesla K40 GPUs against a single core of Intel Xeon-E5-2680v3 CPU for the 3D lid-driven cavity flow.