Abstract
Based on physicsinformed deep learning method, the deep learning model is proposed for thermal fluid fields reconstruction. This method applied fullyconnected layers to establish the mapping function from design variables and space coordinates to physical fields of interest, and then the performance characteristics Nusselt number Nu and Fanning friction factor f can be calculated from the reconstructed fields. Compared with reconstruction model based on convolutional neural network, the improved model shows no constrains on mesh generation and it improves the physical interpretability by introducing conservation laws in loss functions. To validate this method, the forced convection of the waterAl_{2}O_{3} nanofluids is utilized to construct training dataset. As shown in this paper, this deep neural network can reconstruct the physical fields and consequently the performance characteristics accurately. In the comparisons with other classical machine learning methods, our reconstruction model is superior for predicting performance characteristics. In addition to the effect of training size on prediction power, the extrapolation performance (an important but rarely investigated issue) for important design parameters are also explored on unseen testing datasets.
Similar content being viewed by others
Introduction
In recent decades, miniaturization technology facilitates the development of higher efficient equipment, such as electronic components, refrigeration, transportation and so on. However, increased performance has created an urgent need for removing the higher heat loads to protect device from high temperature^{1}. To resolve this challenge, microchannel heat exchanger (MCHE), known as micro electromechanical systems (MEMS), is widely applied in microscale devices due to its large surface to volume ratios, high convective heat transfer coefficient and smaller size^{2}. An obvious fact is that a heat transfer process is determined by the heat transfer device and the utilized working fluid. Thus, the conventional working fluids, such as water and other refrigerant showing scant improvement in thermal properties become one of the limiting factors for higher heat coefficient. The nanofluids combining the base fluid and solid nanoparticles with better thermal conductivity show a promising application in MCHE.
The concept of nanofluids is first proposed by Choi^{3} in 1995, their following works^{4,5} demonstrated that the nanofluids with metallic nanoparticles suspended in conventional heat transfer fluid enhanced heat transfer and reduced pressure drop dramatically. This excellent performance has led to a great upsurge in the study of nanofluids so far. Many stateofart reviews have been reported on the investigations and applications of nanofluids. Gupta^{6} reviewed the experimental investigations of forced convective heat transfer with different nanofluids. In the review of Ghadimi^{7}, the characteristics, numerical model and measurement of thermal conductivity and viscosity were introduced. The thermal and hydraulic performance of nanofluids flowing in mini and micro channels were reported by Sarkar^{8}.
As reported above, nanofluids have higher thermal conductivity and corresponding heat transfer than base fluid. However, the heat transfer coefficient shows complicated relationship with not only thermal conductivity but several factors, such as heat capacity, viscosity, flow pattern and so on. Based on numerical and experimental results, many conventional correlations are derived for different operation and nanofluids. In the conventional fitting methods, they suppose that the mapping function from design parameters to target variables satisfies some specific type of function. And then the rest of the fitting work is how to determine the coefficients. This method is simple and convenient, and the precision is acceptable if a proper function is settled. However, the functions between the design variables and objective functions are unknown in most cases, and they are usually selected by experience and numerous attempts. The selection process of functions becomes laborious if massive design parameters are considered or the relationship shows strong nonlinear.
Recently, more and more researchers adopt machine learning methods, such as adaptive neurofuzzy inference system^{9}, least squares support vector machine^{10}, Gaussian process regression^{11}, artificial neural network (ANN)^{12,13,14,15} and so on, to predict thermophysical properties and thermodynamic performance for nanofluids. The advantages of those machine learning methods include: no specific type of function should be supposed in advance; more design parameters can be included; higher calculation efficiency for prediction; and more importantly, the fitting ability is stronger for nonlinear functions. However, these approaches of conventional fitting methods or machine methods overlooked a fact that most target variables, such as the thermodynamic performance, can be regarded as a kind of refinement from physical fields. Those methods only focus on the mapping functions between design variables and some target variables, which cause the lack of physics interpretability and limit their scope of application. Different from these traditional machine learning methods, we propose a reconstruction framework utilizing deep neural network (DNN) to reconstruct physical fields instead of thermodynamic performance in this study. Once the physical fields are reconstructed, any interested thermodynamic performance can be extracted from fields directly.
As the most wellknown methods in supervised learning, the neural network was demonstrated to approximate any function with sufficiently large and deep network by the universal approximation theorem in 1989^{16}. Recently, according to the widespread growth of data and the rapid advances of supercomputer, the power and flexibility of DNNs have led to a series of breakthroughs for computer vision, natural language processing and many other directions of artificial intelligence. In thermal and fluid mechanics, many complex tasks, such as superresolution reconstruction^{17,18,19}, turbulence model improvement^{20,21,22}, field reconstruction^{23,24,25,26,27}, flow control^{28}, design and optimization^{29,30} and so on, can be accomplished impressively with deep learning architectures.
Traditionally, the physical fields are obtained by means of experiments measuring physical variables in limited space or numerical simulations solving conservation equations. Despite significant advances in experiments and numerical simulations in recent decades, the progress of experiments and simulations is still timeconsuming, laborious and need prior experience. The advantages of DNN enables physical field prediction quickly without manual intervention. Mostly, the physical fields are treated as images (spatial data) or videos (spatial–temporal data), and then many prediction or reconstruction models inspired by computational vision^{31,32,33} are reported. Guo^{23} proposed an approximation model with encoder and decoder to predict flow field from geometry and boundary representations based on DNN. Hennigh^{24} proposed LatNet to compress numerical simulations obtained by Lattice Boltzmann Method using convolutional neural network (CNN). LatNet is composed of three parts, an encoder to compress the state of simulations, a compression mapping to learn the dynamics on compressed state and a decoder realizing the decompress process. Lee^{34} utilized a multiscaled generative adversarial network (GAN) to predict time series of laminar vortex shedding over a cylinder based on previous fields. Kim^{35} synthesized discrete velocity fields in space and time from a set of reduced parameters. In our previous study^{36,37}, a reconstruction model with GAN and fields gradient loss is firstly proposed to predict the physical fields of nanofluids microchannel based on design variables, limited measurement and the effect of training size, measuring uncertainty and measuring layouts are discussed in detail. The imageinspired reconstruction models can obtain the overall physical fields in one prediction at the millisecond level and capture the spatial correlations among grid points efficiently. Despite great potential, practical implementation applying CNN models to computational or experimental fluid dynamics remains limited. Firstly, the field reconstruction task realized by CNN models is still a black box without physics interpretability, and the successes of CNN models mainly rely on the powerful feature extraction and the ability to resolve nonlinear problems. In essence, the thermal and fluids examples are driven by prior knowledge of conservation laws. Secondly, due to the special convolution operations on feature, the input or output fed to CNN model should be preprocessed to structure grid data. In addition to some examples with simple geometry or structure whose grids can be easily transformed to required structured data^{38}, the physical fields of most thermal and fluids examples with complicated geometry should be interpolated to designed structured grid points, which may cause the lack of information around large gradience. Besides, the definition of default values in areas without fields information is a pending problem.
For thermal and fluid mechanics, there exists an obvious prior physical knowledge, which is conservation principles, and all available data respect the physical laws given by conservation principles. Many trials have been made to incorporate and enforce known flow physics in applications of DNN. Zhao^{39} combined domain knowledge and ANN to predict the critical heat flux. In the multiscaled GAN constructed by Lee^{34}, the conservation principles are formulated using the triangle inequality to approximate to the original forms and the loss function of conservation principles is minimized to compare the difference between predicted fields and ground truth. However, the introduction of physical informed loss function requires the predesigned nontrainable convolutional filters^{38} or loss equations around the whole fields^{34,35}, and it heavily depends on the mesh information. Recently, physicsinformed neural networks (PINNs) introduced by Karniadakis and Raissi^{40,41} that are trained to solve supervised tasks respecting physical laws are introduced. According to their explorations of physics informed neural network on surface breaking crack identification^{42}, biological reactions^{43}, compartmental disease transmission models^{44} vortexinduced vibrations^{45}, heat transfer problems^{46} and other flow mechanism^{47}, the physics conservation principles can be utilized to be loss function and regularization term with the automatic differentiation and the backpropagation mechanism, which can be essentially regarded as prior knowledge to constrain the space of admissible solutions. Another advantage is physicsinformed neural networks can provide a meshfree solver as the discrete interpolators in both space and time over the computational domain, which can efficiently handle the unstructured mesh of any numerical methods. Thus, the possibility of using PINNs to approximate flow in idealized stenosis^{48}, arterial blood pressure^{49} and highspeed flow^{50} are investigated. Moreover, Karniadakis^{51} extended PINNs (XPINNs) to space–time domain decomposition for nonlinear partial differential equations in arbitrary complexgeometry domains. Based on XPINNs and another extension (namely the conservative PINNs^{52}), a distributed framework for PINNs^{53} is proposed with several advantages, such as parallelization capacity, large representation ability, efficient hyperparameter tuning and is particularly effective for multiscale and multiphysics problems.
To the best of authors’ knowledge, this is the first attempt of applying PINNs on the field reconstruction for nanofluids convection problem. In this study, we applied PINNs to reconstruct physical fields (pressure, temperature and velocity) of nanofluids convection from varying design variables, including Reynold number, nanofluids properties, geometric parameters and boundary conditions. The primary contributions of this work are listed as followed:

1.
The physicsinformed and meshfree prediction model for nanofluids convection in microchannels is proposed to reconstruct all interested physical fields and then extract the heat transfer characteristics (Nu and f for instance) directly.

2.
This method enforces the conservation laws by introducing mass, momentum and energy continuity equations to guide the training of deep learning neural network.

3.
To evaluate the accuracy of our model in the sense of theory and engineering, the performance of reconstructed fields, conservational residual and the heat transfer and flow characteristics interested in engineering are discussed in detail.

4.
Except for the effect of training size, we also focus on an important but lack of attention issue, that is the extrapolation ability of the neural network on the unknown domains.
The main context of this paper is organized as followed: In section B, the overall architecture of applied physics informed neural network is presented and the data set of nanofluids convection applied to train and test reconstruction network is described. Next, the prediction performance is analyzed for physical field visualizations. Detailed physical fields distributions, evaluation criteria and performance are conducted in section C. And then the comparisons of our method and other surrogate models on the prediction performance of performance characteristics are presented. Moreover, the effect of training size and the extrapolation performance are investigated in the latter part of section C. Finally, the conclusions are summarized in section D.
Methods
Overall architecture
In physicsinformed neural network, the mathematical physics is enforced as partial differential equations (PDE) or ordinary differential equations. The steady governing equations for heat transfer and flow of nanofluids can be expressed as Eq. (1), including the mass, momentum and energy conservation principles.
where \({{\mathcal{N}}}(\bullet)\) are the nonlinear PDE operators representing the conservation principles, x is the space coordinate, θ is a state parameters set to describe the physical system such as fluid properties, boundary conditions, and geometry of the domain, which can be expressed as a D_{θ}dimensional vector; ρ and ν represent density and viscosity of the fluid, respectively; b_{f} is the body force; and Ω denotes the fluid computational domain. The velocities u, pressure p and temperature t can be regarded as functions of the space coordinate x, and variable parameters θ. Specifically, the physical fields can be uniquely determined when suitable boundary conditions are prescribed,
where \({\mathcal{B}}\) is the general differential operators that define the boundary conditions and ∂Ω denotes the boundary regions. Given a set of specific parameters θ, the mapping functions of physical fields, i.e. u(x), p(x) and t(x), can be obtained by discretizing corresponding governing equations in the Eq. (1) using numerical methods, such as Finite Difference /Finite Volume /Finite Element methods. However, the numerical process requires timeconsuming mesh generation and iteratively solving large linear/nonlinear systems. Due to the tedious regeneration of computational mesh, the traditional numerical methods become more challenging with diverse geometries involved.
Different from the traditional numerical methods, the PINNs simply focuses on the reconstruction of physical scalars for single temporal and spatial point rather than the whole physical fields in one prediction. This strategy enables the deep neural network solving the conservation principle quickly without the trouble of mesh generation. To describe the process of numerical method and reconstruction model, some mathematic expressions are listed as follows.
As shown in Eq. (3), the training data is obtained by the traditional numerical or experimental methods based on design parameters θ and coordinates x, and the acquisition method of raw physical fields \({{\varvec{\uppsi}}} = \{ p{,}{\mathbf{u}}^{T} {,}t\}^{T}\) is considered as a mapping \({\mathbb{F}}{(}{\mathbf{x}}{,}{{\varvec{\uptheta}}}{)}\). Correspondingly, the welltrained deep neural network model constructs an approximate mapping \(\widetilde{\mathbb{F}}({\varvec{x}},{\varvec{\theta}};\Theta )\) with learnable parameters \(\Theta\) substituting for the traditional method \({\mathbb{F}}{(}{\mathbf{x}}{,}{{\varvec{\uptheta}}}{)}\) to reconstruct physical fields \({\hat{{\varvec{\uppsi}}}} = \{ \hat{p}{,}{\hat{\mathbf{u}}}^{T} {,}\hat{t}\}^{T}\). According to the description of the heat transfer and flow problem in Eq. (1–2), the key point of reconstructing physical fields is trying to conform with the nonlinear PDEs \({{\mathcal{N}}}({\mathbf{x}},{{\varvec{\uptheta}}}) = {\mathbf{0}}\) and differential operator \({\mathcal{B}}{(}{\mathbf{x}},{{\varvec{\uptheta}}}{)} = {\mathbf{0}}\). Thus, the cost function of the model is considered as the combination shown in Eq. (4). Remarkably, as described in Eq. (5), the training process of the model can be regarded as the optimization process of the proper learning parameters \(\Theta^{*}\) with the cost function \({\mathcal{L}}\).
Nanofluids heat convection problem
In this study, a twodimensional convection problem for the waterAl_{2}O_{3} nanofluids in microchannels are validated regarding to this physics informed neural network model. Though the application of nanofluids become engineering gradually and the application scenario becoming more and more complex, the fundamental researches on the heat and flow behavior in simplified microchannel is still important and necessary. Thus, a simple microchannel model as depicted in Fig. 1 is studied. The microchannel is composed by top wall and bottom wall, where the bottom wall is a smooth plate with infinite length while the top wall is a roughed infinite plated with two grooves/protrusions. The total length of the microchannel is a and the height is H. The inlet and outlet are extended to ensure fully developed and prevent the nanofluids from flowing backward. The lengths of extensions are both a_{1}. The locations of two grooves/protrusions are determined by the interval of a_{3} and the length between the inlet and first groove/protrusion a_{2}. And the geometries of two grooves/protrusions are defined by two parameters: radius R_{1} and R_{2}, and relative depth δ_{1} = d_{1}/R_{2} and δ_{2} = d_{2}/R_{2}. It should be noticed that the depth of the groove is positive d > 0 while the depth of protrusion is negative d < 0. Five geometric parameters related to the location and shapes of grooves/protrusions, including interval of a_{3}, radius R_{1} and R_{2}, and relative depth δ_{1} and δ_{2}, are set as design variables and the other geometric constants are listed in Table 1.
The waterAl_{2}O_{3} nanofluids is composed of base fluid water and nanoparticles Al_{2}O_{3} with a diameter of 30 nm at a certain proportion. In practice, the heat transfer and flow problem for nanofluids is a kind of multiphase flow. To simplify the simulation, the following assumptions are made: the nanoparticles are all spheres with a diameter of 30 nm and distributed uniformly in the base fluid, so the twophase nanofluids can be equivalent to a singlephase fluid. The thermophysical properties of basefluid water and nanoparticle Al_{2}O_{3} are shown in Table 2, and thermophysical properties nanofluids are calculated as followed.
The density of applied nanofluids ρ_{n} is calculated as:
where φ denotes the volume fraction of nanofluids, ρ_{b} and ρ_{p} represent the density of base fluid water and nanoparticles Al_{2}O_{3}, respectively. The specific heat capacity^{54} of applied nanofluids Cp_{n} is calculated as:
where \(Cp_{{\text{b}}}\) and \(Cp_{{\text{p}}}\) represent the specific heat capacity of base fluid water and nanoparticles Al_{2}O_{3}, respectively. The thermal conductivity^{55} of nanofluids λ_{n} is defined as:
where λ_{b} and λ_{p} represent the specific heat capacity of base fluid water and nanoparticles Al_{2}O_{3}, and \(\Delta\) is shown as below:
The dynamic viscosity coefficient^{1} of nanofluids \(\mu_{{\text{n}}}\) is formulated as
where \(\mu_{{\text{b}}}\) dynamic viscosity coefficient of water.
Considering the infinite microchannel studied in this work, the calculation models are simplified to twodimensional models. Thus, mathematic expression of conservation equations for incompressible nanofluids in Eq. (1) can be written as:
where u and v are the velocity along x and y coordinate; p and t represent temperature and pressure quantities; \(\alpha_{{\text{n}}} { = }\lambda_{{\text{n}}} /\rho_{{\text{n}}} Cp_{{\text{n}}}\) is defined as the thermal diffusion coefficient.
The numerical simulations are all obtained by solving conservation equations with the commercial numerical software FLUENT. To confirm the simulation accuracy, the SIMPLE algorithm is adopted to couple pressure and velocity item and the calculation domains are discretized by the secondorder upwind scheme. As listed in Eq. (14–15), the inlet condition is set as velocity input and the inlet temperature of nanofluids is defined at 293 K; the nonslip wall condition is imposed on both up and down walls and the same constant heat flux is set for top and bottom walls.; the outlet pressure of atmosphere is utilized.
Reynold number Re is defined as:
where the hydraulic diameter D_{h} is 2H, and \(u_{\infty }\) is the inlet velocity. The mesh generations for all cases are performed by software ICEMCFD. As shown in Fig. 2, the structured grids are divided into the whole computational domain and the grids around grooves/protrusions are refined. According to the physical geometry, the mesh model is composed of three parts: inlet extension, outlet extension and reconstruction region which is utilized as training and testing datasets.
Dataset description
To embrace the physics information as much as possible, about 6000 cases with eight design variables for the waterAl_{2}O_{3} nanofluids in the microchannel including properties of nanofluids, geometric parameters and boundary conditions are sampled by Latin Hypercubic Sampling method. The raw dataset is divided into two parts: the training dataset (50%, 3000 cases) and the testing dataset (50%, 3000 cases).
The configurations and layouts of two protrusions and grooves are important for heat transfer and flow behavior. Thus, the geometric parameters including radium R_{1}/R_{2}, relative depth δ_{1}/δ_{2} of grooves/protrusions and the interval length a_{3} between two grooves/protrusions are taken as design variables. Besides, the volume fraction φ of nanofluids is set as one of the design variables because it becomes a determining factor for the thermophysical properties once the nanoparticle is determined. Finally, the boundary conditions including Reynold number Re indicating the inlet velocity and the heat flux q imposed on top and bottom walls are also taken into consideration. All the design variables and their ranges are listed in Table 3.
Apparently, each case for microchannel with nanofluids is determined by design variables \({\varvec{\theta}}=\{{\theta }_{1},{\theta }_{2},\cdots ,{\theta }_{{D}_{{\varvec{\theta}}}}{\}}^{T}\), which is \({\varvec{\theta}}=\{Re, \varphi , {a}_{3},{R}_{1},{R}_{2},{\delta }_{1},{\delta }_{2},q{\}}^{T}\) in this study. For each case, the computational domain is divided by 800 × 40 grid points and only 550 × 40 grid points containing main heat transfer and flow characteristics in the reconstruction region are used. Every grid point is considered as a sample point located by corresponding coordinates \({\varvec{x}}=\{x, y{\}}^{T}\) for training and testing dataset. Therefore, the input for a sample point is the combination of design variables and coordinates \({\varvec{\xi}}=\{{{\varvec{x}}}^{T}, {{\varvec{\theta}}}^{T}{\}}^{T}\) and the output is the interested physical quantities \({\varvec{\psi}}=\{p,u,v,t{\}}^{T}\), that is pressure p, temperature t and velocity attributes along two coordinates u and v.
With 550 × 40 sample points in each case, there are 132,000,000 sample points can be collected in all 6000 cases. The input of the training dataset is represented as \({\mathcal{D}}_{{\varvec{\xi}}}^{N}=\{{{\varvec{\xi}}}_{1},{{\varvec{\xi}}}_{2},\dots ,{{\varvec{\xi}}}_{N}\}\), while the output physical fields can be defined as \({\mathcal{D}}_{{\varvec{\psi}}}^{N}=\{{{\varvec{\psi}}}_{1},{{\varvec{\psi}}}_{2},\dots ,{{\varvec{\psi}}}_{N}\}\). The number of sample points in training dataset is \({N}_{s}=\mathrm{66,000,000}\) for 3000 cases. It should be emphasized that 3000 cases for training are selected randomly and then all the sample points are disrupted together. In the testing process, all the sample points in one specific case with same design variables should be taken as input to predict the whole physical fields.
Implementation of deep learning model
Focusing on the nanofluids convection, we establish the deep learning model as shown in Fig. 2 by deep neural network. To ensure the network convergence, the nondimensionalization and normalization methods are involved to scale the input and output tensor. Thus, the input coordinates should be nondimensionalized and normalized while the input parameters of governing equations just need to be normalized. After subsequent operations of deep neural network, interested physical quantities experienced nondimensionalization and normalization operations are predicted. Generally, the measurement loss of the difference between true and predicted normalized physical scalars (the output of DNN after inverse normalization) along with proper optimization algorithm can be utilized to drive the training process of DNN. However, for incorporating the first principle, the conservation loss constructed based on normalized governing equations are required by leveraging automatic difference mechanism. If the output fields are need, then an inverse nondimensionalization should be conducted after the inverse normalization. In the following, we introduce the nondimensionalization and normalization methods applied, the detailed network structure and loss function, and the learning strategy employed in this deep learning framework.
Nondimensionalization and normalization
The order of magnitude of the different physical quantities, pressure, velocities, temperature have a significant relative difference, e.g., p ∼ 10^{4} Pa, u ∼ 1 m/s, v ∼ 10^{–1} m/s and t ∼ 10^{2} K, which casts great difficulty on the training of the neural network. The significant difference in magnitude of the parameters creates a systematic problem for the training of the physicsinformed neural network, as this difference in scales will have a severe impact on the magnitude of the backpropagated gradients that adjust the neural network parameters during training. To overcome this problem, we employ a nondimensionalization and normalization technique with the purpose of scaling the input and the output of the neural networks in a proper scale (e.g., (p^{*}, t^{*}, u^{*}, v^{*}) ∼ O(1)) and normalizing the spatial and temporal coordinates to have zero mean and unit variance for training the neural networks more efficiently. Although there could be a way to weight the components of the loss function to mitigate the bias casted into the loss function due to this discrepancy across scales, this process would require a lot of guesswork and tuning. On the other hand, the proposed nondimensionalization strategy achieves the goal of normalizing the variables in a physically justified and intuitive manner that adheres to the requirements of standard neural net initialization strategies (e.g., Xavier initialization) and yields a robust workflow that is free from adhoc hacks and guesswork. For the purpose of nondimensionalization we introduce some characteristic variables, which are commonly used in multiscale physics modeling^{50} in order to simplify the equations. For this problem the characteristic length H and the characteristic velocity U_{∞}. are applied. Thus, u^{*} and v^{*} are the nondimensional velocity along x and y coordinate; p^{*} and t^{*} represent nondimensional pressure and temperature quantities. The nondimensional quantities are defined as following:
Substituting the Eqs. (14) into the Eq. (11), then the conservation equations can be simplified as Eq. (15).
The Prandtl Number Pr applied in governing equations are shown as below:
After the dimension, eight design variables and two coordinates are considered as input. However, the distributions for different variables show a large deviation. Thus, the Zscore normalization method^{56} is utilized to scale input. The formulation of the normalization method for each sample point is:
where \({\mathbf{x}}^{\prime}\) means the normalized coordinate and \({\mathbf{\theta^{\prime}}}\) means the normalized design variables; \({{\varvec{\upxi}}}^{\prime}\) indicates the normalized input; \({\mathbb{E}}[\bullet]\) and \({\mathbb{V}}[\bullet]\) indicate the expectation and variance operation along the column vectors.
Network structure and loss functions
As described, the input information is a 1D vector and the output, interested physical quantities, is also a 1D vector with different size. Thus, the fullyconnected (FC) operations are utilized to transform the input feature space to target feature space. For a further description, we use a 1D vector \({{\varvec{\eta}}}_{i}^{{D}_{i}}\in {\mathbb{R}}^{{D}_{i}}\) with a length of \(D_{i}\) to represent the output tensor of ith FC layer. Specially, the input information can be indicated by i = 0. Then the mathematic expressions of FC layers can be defined as follows:
where \({\mathbf{W}}_{i}^{{D}_{i1},{D}_{i}}\in {\mathbb{R}}^{{D}_{i1}\times {D}_{i}}\) is the weights of ith FC layer with the size of D_{i1} × D_{i} , \({\mathbf{b}}^{{D}_{i}} \in\)\({\mathbb{R}}^{{D}_{i}}\) is the bias of ith FC layer with size of 1 × D_{i} and N_{l} means the number of FC layers. As shown in Eq. (16), a FC layer consists of three operations in sequence: multiply of input vector and weights, the addition of bias and the activation function operation \(\sigma\) which is used to perform nonlinear transformation. To prevent the problem of vanishing gradient, the activation function ReLU is utilized and the expression can be written as:
It should be noted that the selection of activation function has an important effect on the training and convergence of the deep learning methods. There have been some works on providing more efficient and appropriate activation functions for PINN model. For example, Jagtap et al.^{57} proposed an adaptive activation function with the additional scalable parameter introduced in the network. By leveraging the scalable hyperparameter, the increased convergence and better performance can be achieved. In the following, they^{58} further extended this scalable parameter from global adjustment to the hidden layer even the neurons, and the results demonstrate the improved training speed and accuracy. Based on the idea of adaptive activation functions, the deep Kronecker neural networks^{59} is proposed by Jagtap. In this paper, we mainly focused on the physical field prediction for the nanofluid in microchannel. Thus, the standard activation function of ReLU is utilized.
In this heat transfer and flow of nanofluids, we employ 10 hidden layers and 32 neurons per hidden layer per output variable (i.e. 4 × 32 = 128 neurons per hidden layer). Since the active function is settled, the learnable parameters \(\Theta\) in this network will be the weight matrix and bias vector in each FC layer.
After series transformation of all the FC layers, the direct output is the normalized physical quantities and a reverse normalization operation as shown below is required to get proper predicted physical quantities.
In this reverse normalization method, \({\hat{{\varvec{\uppsi}}}}\) is the predicted physical quantities and \({{\varvec{\uppsi}}}^{\prime}\) is the normalized physical quantities after inverse nondimensionalization obtained at the last layer of the model.
The field reconstruction completed by the deep learning model is a kind of regression method to predict physical quantities from input information. Thus, we utilize a loss function \({\mathcal{L}}_{\psi }\) evaluating the distance between predicted physical quantities and the real ones, which is a loss function of the physical field measurements:
For a heat transfer and flow problem, the final simulation results in each control volume should approximately satisfy the conservation laws, including mass, momentum, and energy conservations. To make the most of physical principles, the partial derivatives for physical quantities \({\hat{\varvec{\uppsi}}}{ = }\{ \hat{p},\hat{t},\hat{u},\hat{v}\}^{T}\) are solved by the automatic differential mechanism of PyTorch to acquire the residuals of conservations. The applied residuals of mass, momentum along x and y coordinate and energy conservations are formulated as follows:
where \(\user2{\mathcal{L}}_{c}\), \(\user2{\mathcal{L}}_{mx}\), \(\user2{\mathcal{L}}_{my}\) and \(\user2{\mathcal{L}}_{e}\) means the loss function of mass, momentum along x coordinate, momentum along y coordinate, and energy conservations, respectively, and \(\widehat{p}\), \(\widehat{t}\), \(\widehat{u}\) and \(\widehat{v}\) denotes the predicted physical quantities. The more accurate the predicted physical fields \(\widehat{{\varvec{\psi}}}\), the closer to zero the loss function of conservation laws \(\user2{\mathcal{L}}_{g} = \{ \user2{\mathcal{L}}_{c} ,\user2{\mathcal{L}}_{mx} ,\user2{\mathcal{L}}_{my} ,\user2{\mathcal{L}}_{e} \}\) is, which means that the control volumes approximately respect the conservation laws. The total loss function is composed by the field loss function and a weighted sum of conservation loss functions:
The weight of conservations \(\lambda_{\user2{\mathcal{L}}}\) aims at balancing the scale of field loss and conservation loss. The training is a search process for a pair of optimal weight matrix \({\mathbf{W}}^{*}\) and bias \({\mathbf{b}}^{*}\) to minimize the total loss function. As shown in Eq. (28), the optimal weight matrix \({\mathbf{W}}^{*}\) and bias \({\mathbf{b}}^{*}\) are the hyperparameters when total loss function \(\user2{\mathcal{L}}_{total}\) reaches minimized value. It should be emphasized that this work focuses on the improvement of the incorporation with conservation laws, thus the boundary condition and initial conditions are removed from loss functions. The investigation of boundary and initial conditions can be found in the application of PINN on cardiovascular flows^{49}.
Performance criteria
Since the deep learning model can predict physical quantities for one sample point in one set, the whole physical fields of a case can be reconstructed by all corresponding sample points fed to model. For better testing and verifying the model, we evaluate the reconstruction power with four criteria in whole fields.
Two field error criteria named relative L_{1} and relative L_{2} for each field are used as a metric for evaluation^{45}. The calculation for them are presented as follows:
where \({\Vert \cdot \Vert }_{1}\) means the norm1 operation, \({\Vert \cdot \Vert }_{2}\) means the norm2 operation, and N is the number of sample points of the field in one case. Except for those two normal criteria indicating the deviations of predicted quantities, we also adopt the residuals of governing equations (mass conservation R_{c}, x momentum conservation R_{mx}, y momentum conservation R_{my} and energy conservation R_{e}) referencing from the numerical software FLUENT (Eq. (28–5) in chapter 28.15 of Fluent help document) to evaluate the accuracy of physical fields.
In addition, two common performance characteristics, Nusselt number Nu and fanning friction factor f to test the accuracy of predicted fields. The Nu representing thermal performance is calculated as:
where h is heat transfer coefficient; \(q\) is heat flux density which is a constant for each case; \(\overline{t}_{w}\) and \(\overline{t}_{f}\) denote the averaged wall temperature and averaged bulk temperature. The f indicating the hydraulic performance is calculated as:
where \(t_{w}\) is the mean temperature of top and bottom walls; \(t_{f}\) is mean fluid temperature; \(p_{{{\text{in}}}}\) is the average pressure at inlet; and \(p_{{{\text{out}}}}\) is the average pressure at outlet. They are shown in Eq. (34).
The relative error (RE) for performance characteristics Nu or f can be calculated as:
where \(y\) is the parameter calculated from original fields and \(\hat{y}\) is the predicted value extracted from the reconstructed fields. On some level, Nu and f criteria can be considered as another integral weighted form of the field error.
Results
Performance analysis of reconstruction
In this section, we demonstrate that the improved model can reliably reconstruct physical fields for nanofluids flowing in microchannels. The training dataset consists of 3000 unique cases (3000 × 550 × 40 sample points) randomly selected from 6000 cases in total. And the remaining 3000 cases are used as test dataset to illustrate the reconstruction performance from four aspects: physical field visualizations, physical field error, residuals of conservation equations and relative error of performance characteristics.
A collection of examples with varying design parameters is presented to visualize the true, predicted physical fields and absolute error distributions in Fig. 3. Based on the geometry, the microchannels can be divided into four types: protrusiongroove, grooveprotrusion, protrusionprotrusion and groovegroove microchannels. Considering the different flow and heat transfer behaviors for four types of microchannels, two examples for each type of microchannels are displayed. Besides, the design variables (\([Re,q, \varphi ,R_{1} ,R{}_{2},\delta_{1} ,\delta_{2} , a_{3} ]\)) are listed at the top of each image to describe the corresponding case.
It can be seen from Fig. 3 that true and predicted pressure both drop down along microchannel and reduces to zero at the outlet as nanofluids developing. Compared with smooth pressure fields of ground truth, the model can reconstruct the pressure field identically in a coarse structure with few discontinuous lines in dramatically changing parts. Due to the heat flux conditions on walls, the nanofluids near walls is heated along microchannel while the temperature of nanofluids in middle keeps around the inlet temperature of 293 K. It should be noted that the two obvious hightemperature zones are found in the leading edge of groove or the trailing edge of protrusion, respectively. Even with some dramatic changes in temperature distributions, our approach can generate plausible temperature fields that close to ground truth. And the prediction error mainly concentrated around dramatical changes areas, especially for the leading and trailing edge of groove/protrusion. As for velocity u, different phenomena are observed around grooves and protrusions. In protrusion, a highvelocity zone forms when the microchannel narrows and then flows towards the top wall after the groove. No significant change is found in groove except for lower u in the expansion of microchannel. From the plots in Fig. 3, the reconstructed u distributions are consistent with the true distributions. The distribution of velocity v is much more complex, which a lowvelocity zone and a highvelocity zone are formed around grooves and protrusions. Though it can be found that the absolute error is much large compared with the velocity field, the large errors only exist at local points, making little influence on global velocity distributions. It can be inferred that predicted velocity v fields resemble the ground truth flow fields. As discussed above, our model can predict plausible physical fields, including pressure, temperature, and velocities, for all kinds of microchannels.
In Fig. 4, three curves A, B and C are plotted in microchannels to describe the prediction performance near bottom wall, the center of microchannel and near top wall, respectively. Along three curves, the comparisons between predicted pressure, temperature and velocity (u, v) attributes and true fields are presented in Fig. 5, where the shaded areas indicate the groove and protrusion. As shown in plots, the model can favorably predict four field attributes in detail, while velocity v is found close to ground truth with marginal deviations in curve A. This may be partly due to the low values of velocity v near wall. This brilliant prediction ability enables the deep learning model to gather arbitrary points of physical attributes with favorable accuracy and analyze the heat transfer and flow behavior.
To quantitatively investigate the reconstruction performance, the two evaluation criteria relative L_{1} and L_{2} error as well as the residuals of four conservation equations, mass, momentum, and energy, are discussed in Fig. 6. For four physical attributes, all averaged L_{1} and L_{2} error is lower than 0.02 and 0.1. The low field criteria distributions indicate that our approach can reconstruct fields for different design variables with high fidelity. From Fig. 6b, the residuals of predicted conservation equations are close to the true residuals and the maximum deviation may be the residual of predicted mass equation which is tenfold true residual. This result illustrates that the reconstructed physical fields almost satisfy the conservation equations. In some way, it can be conjectured that the physicalinformed model no longer simply reconstructed physical fields, but approximate the inherent physical laws based on prior information of conservation equations to improve the reconstruction performance.
In flow and heat transfer problem, the most important performance characteristics are Nusselt number Nu representing thermal performance and Fanning friction factor f representing hydraulic performance. In recent decades, numerous studies focused on the surrogate models predicting Nu and f based on measurements or design variables. Herein, the surrogate model is replaced by the reconstruction model, and the characteristics can be extracted from generated fields by integral operations shown in Eq. (32). In Fig. 7, we present the characteristics prediction performance. It is observed that the relative errors of two performance characteristics distributed around 0 with negligible bias. Besides, the relative error of Nu is high to 13% while it is less than 5% for most examples. Likewise, though the maximum relative error of f is high to 23%, most errors are less than 10%. It can be inferred that the accuracies of performance characteristics are high enough to complete diverse engineering applications in the range of studied cases. Overall, the proposed physicsinformed reconstruction framework generalizes well for the tasks of field reconstruction and performance prediction.
Compared with classical surrogate models
In the traditional design process, numerical simulations are conducted over and over upon the previous unqualified designs until a satisfying result is obtained. Although the computational ability is powerful nowadays, the design periodic is still dragged by massive numerical simulations which is timeconsuming for iterative solution of highdimensional equations. To accelerate the design process, numerical simulations are replaced by surrogate models to approximate the objective functions, such as Nu and f. However, single surrogate model can only construct one mapping function, and the surrogate model should be renewed once the objective functions changed. Nearly all the objective function is a kind of abstract for physical fields which means that they can be extracted from fields by mathematical methods. Based on this fact, our approach obtains the objective functions directly from physical fields predicted by PINNs, which makes it possible to get any multiple mapping functions together.
In flow and heat transfer problems, performance characteristics Nu and f are widely used as objective functions. Thus, the prediction performance of different traditional surrogate models including Linear Regression (LR), Polynomial Regression (PR), Supported Vector Regression (SVR), Artificial Neural Network (ANN), Gauss Process Regression (GPR), Random Forest (RF), Extreme Gradient Boosting (XGB), our previous reconstruction model constructed by the deep convolutional neural network^{36} (RDCNN) and our approach in this paper for Nu and f are plotted in Fig. 8. It should be noted that RDCNN and PINNs are both reconstruction models and they adopt same extraction method of Nu and f from physical fields. However, the RDCNN is different from PINNs by its imagetreated physical fields and deep convolutional neural network for reconstruction. To distinguish the prediction performance, three global criteria Rsquare (R^{2}), mean squared error (MSE) and mean absolute error (MAE) and relative error (RE) of performance characteristics are utilized. The mathematical expressions of three global criteria are described below:
where y means the original performance characteristics, \(y^{\prime}\) means the performance characteristics predicted by surrogate models, \(\overline{y }\) is the mean value of original performance characteristics and N is the total number of testing dataset, which is 3000 for this comparison (3000 samples are applied as training dataset).
It can be observed from Fig. 8a,b that reconstruction model RDCNN shows the best prediction performance of Nu ,and our model is slightly lower than RDCNN with 0.992 R^{2} score, 0.081 MSE, 0.152 MAE and 88.6% for ± 10% Relative Error (RE), respectively. Among other classical surrogate models, XGB provides best prediction accuracy which is still lower than our reconstruction models. For the prediction performance of f, the improved model outperforms other models with 0.994 R^{2} score, 3.89 × 10^{–5} MSE, 2.58 × 10^{–4} MAE and 92.6% for ± 10% Relative Error (RE). As for other classical surrogate models, the prediction power of XGB is far more than others, but it is still inferior to reconstruction models considering comprehensive performance. The results confirm that the reconstruction models are beneficial to predict performance characteristics based on predicted physical fields compared with regular surrogate models.
Effect of training size
An inherent problem of datadriven machine learning approaches is that model performance strongly depends on the quality of training dataset. Thus, it is significant to investigate the effect of training size on the reconstruction performance of our reconstruction model. In this section, four training sizes are utilized: 3000 groups of sample points (50%), 2000 groups (33%), 1000 groups (17%) and 500 groups (8%) for all 6000 groups in raw data. Besides, the remain 3000 groups of sample points are taken as testing dataset for all the models with different training size. In the following, the effect of training size is studied from three factors: relative errors of physical fields, residuals of conservation equations and relative errors of performance characteristics.
Figure 9 shows the comparisons of relative field errors L_{1} and L_{2} for four different training size. It is apparent that physical relative errors decrease as training size enlarges from 500 to 3000 groups of sample points and the decrement drops down gradually. Averaged physical errors L_{1} and L_{2} for all physical fields are under 0.05 with training size greater than 2000 (33%). Once the training size is up to 2000, increasing training size shows a very limited improvement of reconstruction accuracy.
In Fig. 10, it is observed that the residuals of predicted conservation equations with four training size are all below 2 × 10^{–3}, while the residuals of true conservation equations are lower than 10^{–4}. Similar to the physical field relative errors, the residuals of conservation equations keep decreasing till training size is higher than 2000. The varying trend indicates that the reconstruction performance is difficult to be improved through reducing residuals of conservation equations with adequate training samples (2000 groups of sample points in this study).
In addition, the relative errors of attentionattracting performance characteristics Nu and f are presented in Fig. 11. From the box plots for relative error of Nu, it can be found that the relative error of thermal performance is distributed around 0 and the maximum relative error decreases from 12 to 5% with training size increasing from 500 to 3000 groups. As for hydraulic performance f, the averaged relative error fluctuates around 0 and the maximum relative error also decreases from 15 to 8% with larger training size. The higher relative error for f may due to the much smaller values of f and the higher relative errors of pressure attribute. Similar to physical fields relative errors and residuals of conservation equations, the relative errors of performance characteristics show little reduction when training size is larger than 2000.
Overall, the increase of training size has an obvious positive effect on reconstruction performance, including reducing physical fields relative errors, regulating residuals of conservation equations and decreasing relative errors of performance characteristics. However, this improvement of reconstruction power is constrained if the training samples are adequate. In the light of this study, training data of 2000 groups is enough to attain excellent PINN model with acceptable accuracy.
Extrapolation performance
As discussed above, the applied physicsinformed model shows excellent capability for multiple evaluation criteria of interested fields and characteristics. However, it should be noted that the reconstruction of fields above can be regarded as interpolation due to randomly divided datasets whose design variables of training and testing set are under the same probability distribution. To demonstrate the generality and scalability of our method, it is an intuitive idea to investigate extrapolation power of the approach—one of the most important problem but perhaps rarely mentioned in recent research. In this section, we evaluate the extrapolation capabilities from two aspects: physical fields relative errors L_{2} and relative errors of performance characteristics, Nu and f. Moreover, we repartition different training data set based on the extrapolation of four important parameters Re, q, φ and a_{3}, covering the design information related to geometry, nanofluids and boundary conditions.
From Figs. 12, 13, 14, 15, the relative L_{2} error of physical fields and relative error of characteristics are plotted with varying design variables for training and testing datasets. In addition, the pressure fields in the testing dataset are displayed for visual observation. An obvious result can be found that relative L_{2} error and relative error of performance rise gradually as design variables extend outward the training ranges. This indicates a noticeable conclusion that the farther the design variables are away from the training interval, the worse the reconstruction performance. The design variable a_{3}, which provides the least impact on the physical fields, has the best extrapolation performance. It is observed that the reconstructed fields in testing dataset show a relatively good match with the ground truth, the relative error L_{2} is less than 0.1 except some special cases and the relative error of characteristics concentrated between 10% and 10%. Then the extrapolation performance of φ ranks second, with reasonable physical fields, slightly higher relative error L_{2} and larger relative error of performance.
As for the boundary conditions of Re and q, the extrapolation performance along increasing variables is pretty good while a clear accelerating downward trend of relative L_{2} error and an increasing bias of increasing relative error of performance can be observed with decreasing variable. For the extrapolation of Re, the relative L_{2} errors of pressure and velocity u up to 0.5 and the relative error of f ups to 50% with lower Re while the relative error of Nu is much smaller. Besides, the reconstructed pressure fields can only obtain coarse flow patterns in low Re while plausible predictions of pressure fields are found with high Re. The reason for this result is the significant influence of varying Re on pressure. Likewise, due to the close relationship between q and temperature field, the relative L_{2} errors of temperature field (maximum of 0.5) and the relative error of Nu (maximum of 25%) is much higher with decreasing q. Besides, the reconstructed physical fields agree well with ground truth with lower q. The results show that our model can predict multiple physical fields and performance characteristics with favorable accuracy if Re and q increase, and acceptable accuracy if Re and q slightly decrease.
In summary, the prediction performance become worse with design variables deviated from training ranges and this is determined by the regression essence of neural network. Besides, the larger the influence of design variable on fields, the less satisfactory the extrapolation performance. Even though the accuracy decreases with the design variable extending, our approach can reconstruct plausible physical fields and predict the thermal and hydraulic performance accurately. Especially for φ and a3, the relative error of performance characteristics ranging from 10% to 10% for almost all cases and the physical fields visualizations are in good agreement with ground truths. It indicates that our model enables relatively accurate prediction out of the training range to some extent.
Conclusion
In this study, a physics informed deep neural network incorporating the first principle, which is conservation laws in thermal and fluids mechanism, is proposed to reconstruct physical fields for nanofluids convection with design variables as input, including nanofluids volume fraction, geometric parameters and boundary conditions parameters. The main results are concluded as follows:

1.
The prediction power of our model is validated from four factors: the physical fields visualizations resemble the ground truth with reasonable details; the relative L_{1} and L_{2} for physical fields are quite lower than 0.02 and 0.1; the residuals of conservations close to the results of numerical simulations and the relative error for Nu and f are less than 10% for most cases.

2.
Compared with classical surrogate models, reconstruction models show superior prediction performance of either Nu or f due to reconstructed fields (RDCNN ranks 1 for Nu and our model ranks 1 for f).

3.
As indicated in the results with different training sizes, the more sample points involved in training, the more powerful the physicsinformed model is for reconstructing physical fields. In nanofluids convection, 2000 groups of sample points enable the physicsinformed model to achieve best prediction performance approximately.

4.
The evaluations of reconstruction performance with extended design variables demonstrate that the proposed model shows certain parametric extrapolation ability for the heat transfer and flow of nanofluids.
Abbreviations
 \({\varvec{\psi}}\) :

True physical quantities
 \(\widehat{{\varvec{\psi}}}\) :

Predicted physical quantities
 R:

Conservation residual
 \({\mathbf{x}}\) :

Coordinates vector
 \({\mathbf{x}}^{\prime}\) :

Coordinate after normalization and nondimensionalization
 x :

X coordinate (μm)
 y :

Y coordinate (μm)
 x ^{*} :

Nondimensional x coordinate
 y ^{*} :

Nondimensional y coordinate
 b _{f} :

Body force
 u :

Velocity vector
 p :

Pressure (Pa)
 t :

Temperature (K)
 u :

Velocity along x coordinate (m/s)
 v :

Velocity along y coordinate (m/s)
 p ^{*} :

Nondimensional pressure
 t ^{*} :

Nondimensional temperature
 u ^{*} :

Nondimensional x velocity
 v ^{*} :

Nondimensional y velocity
 \(\widehat{p}\) :

Predicted pressure (Pa)
 \(\widehat{t}\) :

Predicted temperature (K)
 \(\widehat{u}\) :

Predicted x velocity (m/s)
 \(\hat{v}\) :

Predicted y velocity (m/s)
 H :

Microchannel height (μm)
 d :

Groove depth (μm)
 a :

Length of microchannel
 a _{1} :

Length of extension (μm)
 a _{2} :

Length between inlet and first groove/protrusion (μm)
 a _{3} :

Groove interval (μm)
 \(u_{\infty }\) :

Inlet velocity (m·s^{−}^{1})
 \(C_{P}\) :

Specific heat (J·kg^{−}^{1}·K^{−}^{1})
 \(q\) :

Heat flux of boundary (W·m^{−}^{2})
 R :

Groove radius (μm)
 D _{h} :

Hydraulic length (μm)
 D _{θ} :

The dimensions of variable parameters
 P r :

Prandtl number
 Nu :

Nusselt number
 f :

Fanning friction factor
 Re :

Reynolds number
 t _{f} :

Mean temperature of fluid (K)
 t _{w} :

Mean temperature of walls (K)
 Δt :

Temperature difference (Pa)
 Δp :

Pressure drop along channel (Pa)
 N _{s} :

Number of sample points in training dataset
 \({\mathbb{F}}\left(\bullet \right)\) :

Traditional numerical or experimental methods
 \({\tilde{\mathbb{F}}}\left(\bullet \right)\) :

Approximate mapping for physical fields
 b ^{*} :

The optimal Bias of fullyconnected layer
 \({\mathbb{E}}[\bullet]\) :

Expectation operation
 \({\mathbb{V}}[\bullet]\) :

Variance operation
 \({\mathbf{W}}\) :

Weights of fullyconnected layer
 b :

Bias of fullyconnected layer
 \({\mathcal{B}}\) :

General differential operators that define boundary conditions
 \({\mathbb{R}}\) :

Real vector space
 s :

The first moment estimations
 r :

The second moment estimations
 R^{2} :

Determination coefficient for performance characteristics
 \(L_{1}\) :

Relative field norm1 error
 \(L_{2}\) :

Relative field norm2 error
 \({\mathbf{W}}\) ^{*} :

The optimal weight of fullyconnected layer
 \({\mathbf{\mathcal{D}}}\) :

Sampling space of dataset
 \({\mathbf{\mathcal{N}}}(\bullet)\) :

Nonlinear partial differential equations operators
 \(\Theta\) :

Learnable parameter set
 \(\Theta^{*}\) :

Optimal learnable parameters
 α:

Thermal diffusion coefficient
 \({{\varvec{\upxi}}}\) :

Input vector
 \({\mathcal{L}}(\bullet)\) :

Loss function
 \({{\varvec{\Omega}}}\) :

Computational domain
 ρ :

Density (kg·m^{−}^{3})
 \(\lambda\) :

Thermal conductivity (W∙m^{−}^{1}∙K^{−}^{1})
 \({{\varvec{\upxi}}}^{\prime}\) :

Normalized input
 μ :

Dynamic viscosity (Pa·s)
 \(\delta\) :

A small constant value for numerical stability
 \(\varphi\) :

Volume fraction of nanofluids (%)
 δ :

Groove relative depth
 \({\mathbf{\theta^{\prime}}}\) :

Normalized design variables
 θ :

Variable parameters
 σ (·) :

Active function operator
 \(\beta\) :

Attenuation coefficients
 \({{\varvec{\upeta}}}\) :

The output tensor of fullyconnected layer
 n:

Nanofluids
 p:

Nanoparticle
 b:

Base fluid
 c:

Mass conservations
 bottom:

Bottom wall
 in:

Microchannel inlet
 out:

Microchannel outlet
 e:

Energy conservations
 g:

Governing equation
 mx:

Momentum conservations along x coordinate
 my:

Momentum conservations y coordinate
 top:

Top wall
References
Mohammed, H. A., Bhaskaran, G., Shuaib, N. H. & Saidur, R. Heat transfer and fluid flow characteristics in microchannels heat exchanger using nanofluids: A review. Renew. Sustain. Energy Rev. 15(3), 1502–1512. https://doi.org/10.1016/j.rser.2010.11.031 (2011).
Whitesides, G. M. The origins and the future of microfluidics. Nature 442(7101), 368–373. https://doi.org/10.1038/nature05058 (2006).
S. U. S. Choi, “Enhancing thermal conductivity of fluids with nanoparticles,” in American Society of Mechanical Engineers, Fluids Engineering Division, 1995, vol. 231, pp. 99–105.
Eastman, J. A., Choi, S. U. S., Li, S., Yu, W. & Thompson, L. J. Anomalously increased effective thermal conductivities of ethylene glycolbased nanofluids containing copper nanoparticles. Appl. Phys. Lett. 78(6), 718–720. https://doi.org/10.1063/1.1341218 (2001).
Choi, S. U. S., Li, S. & Eastman, J. A. Measuring thermal conductivity of fluids containing oxide nanoparticles. J. Heat Transfer 121(2), 280–289. https://doi.org/10.1115/1.2825978 (1999).
Gupta, M., Arora, N., Kumar, R., Kumar, S. & Dilbaghi, N. A comprehensive review of experimental investigations of forced convective heat transfer characteristics for various nanofluids. Int. J. Mech. Mater. Eng. 9(1), 1–21. https://doi.org/10.1186/s407120140011x (2014).
Ghadimi, A., Saidur, R. & Metselaar, H. S. C. A review of nanofluid stability properties and characterization in stationary conditions. Int. J. Heat Mass Transf. 54(17–18), 4051–4068. https://doi.org/10.1016/j.ijheatmasstransfer.2011.04.014 (2011).
Sarkar, J. A critical review on convective heat transfer correlations of nanofluids. Renew. Sustain. Energy Rev. 15(6), 3271–3277. https://doi.org/10.1016/j.rser.2011.04.025 (2011).
Baghban, A., Kahani, M., Nazari, M. A., Ahmadi, M. H. & Yan, W. M. Sensitivity analysis and application of machine learning methods to predict the heat transfer performance of CNT/water nanofluid flows through coils. Int. J. Heat Mass Transf. 128, 825–835. https://doi.org/10.1016/j.ijheatmasstransfer.2018.09.041 (2019).
Ahmadi, M. H., Ahmadi, M. A., Nazari, M. A., Mahian, O. & Ghasempour, R. A proposed model to predict thermal conductivity ratio of Al 2 O 3 /EG nanofluid by applying least squares support vector machine (LSSVM) and genetic algorithm as a connectionist approach. J. Therm. Anal. Calorim. 135(1), 271–281. https://doi.org/10.1007/s109730187035z (2019).
C. K. I. Williams and C. E. Rasmussen, “Gaussian processes for regression,” 1996. doi: https://doi.org/10.1016/01654896(94)900086.
HemmatEsfe, M. & Afrand, M. Predicting thermophysical properties and flow characteristics of nanofluids using intelligent methods: focusing on ANN methods. J. Thermal Anal. Calorimetry 140(2), 501–525. https://doi.org/10.1007/s10973019087892 (2020).
Bagherzadeh, S. A. et al. Minimize pressure drop and maximize heat transfer coefficient by the new proposed multiobjective optimization/statistical model composed of ‘ANN + Genetic Algorithm’ based on empirical data of CuO/paraffin nanofluid in a pipe. Phys. A 527, 121056. https://doi.org/10.1016/j.physa.2019.121056 (2019).
Maleki, A., Haghighi, A., IrandoostShahrestani, M. & Abdelmalek, Z. Applying different types of artificial neural network for modeling thermal conductivity of nanofluids containing silica particles. J. Thermal Anal. Calorimetry https://doi.org/10.1007/s1097302009541x (2020).
Wu, H. et al. Present a new multi objective optimization statistical Pareto frontier method composed of artificial neural network and multi objective genetic algorithm to improve the pipe flow hydrodynamic and thermal properties such as pressure drop and heat transfer. Phys. A 535, 122409. https://doi.org/10.1016/j.physa.2019.122409 (2019).
Hornik, K., Stinchcombe, M. & White, H. Multilayer feedforward networks are universal approximators. Neural Netw. 2(5), 359–366. https://doi.org/10.1016/08936080(89)900208 (1989).
Fukami, K., Fukagata, K. & Taira, K. Superresolution reconstruction of turbulent flows with machine learning. J. Fluid Mech. 870, 106–120. https://doi.org/10.1017/jfm.2019.238 (2019).
Lee, Y., Yang, H. & Yin, Z. PIVDCNN: Cascaded deep convolutional neural networks for particle image velocimetry. Exp. Fluids 58(12), 1–10. https://doi.org/10.1007/s0034801724561 (2017).
Liu, B., Tang, J., Huang, H. & Lu, X. Y. Deep learning methods for superresolution reconstruction of turbulent flows. Phys. Fluids 32, 2. https://doi.org/10.1063/1.5140772 (2020).
B. Tracey, K. Duraisamy, and J. J. Alonso, “A machine learning strategy to assist turbulence model development,” 53rd AIAA Aerospace Sciences Meeting, no. January, pp. 1–22, 2015, doi: https://doi.org/10.2514/6.20151287.
Ling, J., Kurzawski, A. & Templeton, J. Reynolds averaged turbulence modelling using deep neural networks with embedded invariance. J. Fluid Mech. 807, 155–166. https://doi.org/10.1017/jfm.2016.615 (2016).
B. Tracey, K. Duraisamy, and J. J. Alonso, “Application of supervised learning to quantify uncertainties in turbulence and combustion modeling,” 51st AIAA Aerospace Sciences Meeting including the New Horizons Forum and Aerospace Exposition 2013, no. January, pp. 1–18, 2013, doi: https://doi.org/10.2514/6.2013259.
X. Guo, W. Li, and F. Iorio, “Convolutional neural networks for steady flow approximation,” Proceedings of the ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, vol. 13–17Augu, pp. 481–490, 2016, doi: https://doi.org/10.1145/2939672.2939738.
Hennigh, O. LatNet: Compressing lattice boltzmann flow simulations using deep neural networks. (2017)
Li, Y., Wang, H., Mo, K., Zeng, T.: Reconstruction of simulationbased physical field by reconstruction neural network method. 2018.
Bhatnagar, S., Afshar, Y., Pan, S., Duraisamy, K. & Kaushik, S. Prediction of aerodynamic flow fields using convolutional neural networks. Comput. Mech. 64(2), 525–545. https://doi.org/10.1007/s00466019017400 (2019).
Sekar, V., Jiang, Q., Shu, C. & Khoo, B. C. Fast flow field prediction over airfoils using deep learning approach. Phys. Fluids 31, 5. https://doi.org/10.1063/1.5094943 (2019).
Ren, F., Bao Hu, H. & Tang, H. Active flow control using machine learning: A brief review. J. Hydrodyn. 32(2), 247–253. https://doi.org/10.1007/s4224102000260 (2020).
J. Viquerat, J. Rabault, A. Kuhnle, H. Ghraieb, A. Larcher, and E. Hachem, “Direct shape optimization through deep reinforcement learning,” arXiv, 2019.
Wang, Y., Liu, T., Zhang, D. & Xie, Y. Dualconvolutional neural network based aerodynamic prediction and multiobjective optimization of a compact turbine rotor. Aerosp. Sci. Technol. 116, 106869. https://doi.org/10.1016/j.ast.2021.106869 (2021).
I. J. Goodfellow et al., “Generative Adversarial Nets,” 2014.
K. He, X. Zhang, S. Ren, and J. Sun, “Identity mappings in deep residual networks,” Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics), vol. 9908 LNCS, pp. 630–645, 2016, doi: https://doi.org/10.1007/9783319464930_38.
M. Sundermeyer, R. Schlüter, and H. Ney, “LSTM neural networks for language modeling,” 13th Annual Conference of the International Speech Communication Association 2012, INTERSPEECH 2012, vol. 1, pp. 194–197, 2012.
Lee, S. & You, D. Datadriven prediction of unsteady flow over a circular cylinder using deep learning. J. Fluid Mech. 879, 217–254. https://doi.org/10.1017/jfm.2019.700 (2019).
Kim, B. et al. Deep fluids: A generative network for parameterized fluid simulations. Comput. Graph. Forum 38(2), 59–70. https://doi.org/10.1111/cgf.13619 (2019).
Liu, T., Li, Y., Xie, Y. & Zhang, D. Deep learning for nanofluid field reconstruction in experimental analysis. IEEE Access 8, 64692–64706. https://doi.org/10.1109/ACCESS.2020.2979794 (2020).
Liu, T., Li, Y., Jing, Q., Xie, Y. & Zhang, D. Supervised learning method for the physical field reconstruction in a nanofluid heat transfer problem. Int. J. Heat Mass Transfer 165, 120684. https://doi.org/10.1016/j.ijheatmasstransfer.2020.120684 (2021).
Gao, H., Sun, L. & Wang, J. X. PhyGeoNet: Physicsinformed geometryadaptive convolutional neural networks for solving parameterized steadystate PDEs on irregular domain. J. Comput. Phys. 428, 110079. https://doi.org/10.1016/j.jcp.2020.110079 (2021).
Zhao, X., Shirvan, K., Salko, R. K. & Guo, F. On the prediction of critical heat flux using a physicsinformed machine learningaided framework. Appl. Thermal Eng. https://doi.org/10.1016/j.applthermaleng.2019.114540 (2020).
Raissi, M., Perdikaris, P. & Karniadakis, G. E. Physics informed deep learning (Part II): Datadriven discovery of nonlinear partial differential equations. Part II, 1–19 (2017).
Raissi, M., Perdikaris, P. & Karniadakis, G. E. Physics informed deep learning (Part I): Datadriven solutions of nonlinear partial differential equations. Part I, 1–22 (2017).
Shukla, K., Di Leoni, P.C., Blackshire, J., Sparkman, D. & Karniadakis, G.E. Physicsinformed neural network for ultrasound nondestructive quantification of surface breaking cracks. arXiv:2005.03596 [cs, stat], May 2020, Accessed: Nov. 22, 2021. [Online]. Available: http://arxiv.org/abs/2005.03596
Yazdani, A., Raissi, M. & Karniadakis, G. E. Systems biology informed deep learning for inferring parameters and hidden dynamics. Syst. Biol. https://doi.org/10.1101/865063 (2019).
Raissi, M., Ramezani, N. & Seshaiyer, P. On parameter estimation approaches for predicting disease transmission through optimization, deep learning and statistical inference methods. Lett. Biomath. https://doi.org/10.1080/23737867.2019.1676172 (2019).
Raissi, M., Wang, Z., Triantafyllou, M. S. & Karniadakis, G. E. Deep learning of vortexinduced vibrations. J. Fluid Mech. 861, 119–137. https://doi.org/10.1017/jfm.2018.872 (2019).
Zobeiry, N. & Humfeld, K. D. A physicsinformed machine learning approach for solving heat transfer equation in advanced manufacturing and engineering applications. Eng. Appl. Artif. Intell. 101, 104232. https://doi.org/10.1016/j.engappai.2021.104232 (2021).
Raissi, M., Yazdani, A. & Karniadakis, G. E. Hidden fluid mechanics: Learning velocity and pressure fields from flow visualizations. Science 367(6481), 1026–1030. https://doi.org/10.1126/science.aaw4741 (2020).
Sun, L., Gao, H., Pan, S. & Wang, J. X. Surrogate modeling for fluid flows based on physicsconstrained deep learning without simulation data. Comput. Methods Appl. Mech. Eng. 361, 112732. https://doi.org/10.1016/j.cma.2019.112732 (2020).
Kissas, G. et al. Machine learning in cardiovascular flows modeling: Predicting arterial blood pressure from noninvasive 4D flow MRI data using physicsinformed neural networks. Comput. Methods Appl. Mech. Eng. 358, 112623. https://doi.org/10.1016/j.cma.2019.112623 (2020).
Mao, Z., Jagtap, A. D. & Karniadakis, G. E. Physicsinformed neural networks for highspeed flows. Comput. Methods Appl. Mech. Eng. 360, 112789. https://doi.org/10.1016/j.cma.2019.112789 (2020).
Aeee, J. & Karniadakis, G. E. Extended physicsinformed neural networks (XPINNs): A generalized spacetime domain decomposition based deep learning framework for nonlinear partial differential equations. CiCP 28(5), 2002–2041. https://doi.org/10.4208/cicp.OA20200164 (2020).
Jagtap, A. D., Kharazmi, E. & Karniadakis, G. E. Conservative physicsinformed neural networks on discrete domains for conservation laws: Applications to forward and inverse problems. Comput. Methods Appl. Mech. Eng. 365, 113028. https://doi.org/10.1016/j.cma.2020.113028 (2020).
Shukla, K., Jagtap, A. D. & Karniadakis, G. E. Parallel physicsinformed neural networks via domain decomposition. J. Comput. Phys. 447, 110683. https://doi.org/10.1016/j.jcp.2021.110683 (2021).
Ma, J., Nie, B. & Xu, F. Transient flows on an evenly heated wall with a fin. Int. J. Heat Mass Transf. 118, 235–246. https://doi.org/10.1016/j.ijheatmasstransfer.2017.10.117 (2018).
Mahian, O. et al. A review of entropy generation in nanofluid flow. Int. J. Heat Mass Transf. 65, 514–532. https://doi.org/10.1016/j.ijheatmasstransfer.2013.06.010 (2013).
Lundberg, J. Lifting the crowncitation zscore. J. Informet. 1(2), 145–154. https://doi.org/10.1016/j.joi.2006.09.007 (2007).
Jagtap, A. D., Kawaguchi, K. & Karniadakis, G. E. Adaptive activation functions accelerate convergence in deep and physicsinformed neural networks. J. Comput. Phys. 404, 109136. https://doi.org/10.1016/j.jcp.2019.109136 (2020).
Jagtap, A. D., Kawaguchi, K. & EmKarniadakis, G. Locally adaptive activation functions with slope recovery for deep and physicsinformed neural networks. Proc. R. Soc. A. 476, 2239. https://doi.org/10.1098/rspa.2020.0334 (2020).
Jagtap, A. D., Shin, Y., Kawaguchi, K. & Karniadakis, G. E. Deep Kronecker neural networks: A general framework for neural networks with adaptive activation functions. Neurocomputing 468, 165–180. https://doi.org/10.1016/j.neucom.2021.10.036 (2022).
Author information
Authors and Affiliations
Contributions
Y.L. and T.L. wrote the main manuscript text, T.L. provided the methods and Y.L. prepared all figures. All authors reviewed the manuscript.
Corresponding author
Ethics declarations
Competing interests
The authors declare no competing interests.
Additional information
Publisher's note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Li, Y., Liu, T. & Xie, Y. Thermal fluid fields reconstruction for nanofluids convection based on physicsinformed deep learning. Sci Rep 12, 12567 (2022). https://doi.org/10.1038/s41598022164631
Received:
Accepted:
Published:
DOI: https://doi.org/10.1038/s41598022164631
This article is cited by

Physicsinformed neural network with transfer learning (TLPINN) based on domain similarity measure for prediction of nuclear reactor transients
Scientific Reports (2023)

A guide to the preparation techniques of six classes of metal, metal oxide, and carbonbased nanofluids and the implications for their stability
Journal of Thermal Analysis and Calorimetry (2023)
Comments
By submitting a comment you agree to abide by our Terms and Community Guidelines. If you find something abusive or that does not comply with our terms or guidelines please flag it as inappropriate.