An accurate estimation of three-dimensional (3D) temperature fields in channel flows is challenging but critical for many important applications such as heat exchangers, radiation energy collectors, and enhanced geothermal systems. In this paper, we demonstrate the possibility of inferring temperature fields from concentration fields for laminar convection flows in a 3D channel using a machine learning (ML) approach. The study involves generation of data using 3D numerical simulations, application of deep learning methodology using conditional generative adversarial networks (cGANs), and analysis of how dataset selection affects model performance. The model is also tested for applicability in different convection scenarios. Results show that cGANs can successfully infer temperature fields from concentration fields, and the reconstruction accuracy is sensitive to the training dataset selected. In this study, we demonstrate how ML can be used to overcome the limitations of traditional heat and mass analogy functions widely used in heat transfer research.

Mixed convection flows occur in a wide range of industrial processes and natural phenomena.1–3 Free and forced convection effects are comparable in mixed convection, leading to buoyancy-driven secondary flows as well as unstable and inhomogeneous flow fields. The flow characteristics of mixed convection are known to enhance heat transfer up to 4–5 folds as compared to pure forced convection.4–8 Thus, a better understanding of heat and mass transfer in mixed convection flows is necessary to optimize the design of cooling systems for electronic devices, compact heat exchangers, cooling cores of nuclear reactors, and enhanced geothermal systems. For this purpose, extensive investigations have been performed to study heat transfer under different channel geometries and operating conditions.9–12 

In mixed convection, both free convection and forced convection contribute to the heat transfer process, where the interplay between bulk fluid flow caused by forced convection and buoyancy-induced flow caused by temperature gradients can lead to complex flow patterns. Applications involving heat transfer in channel flows often involve not only heat transport but also mass (solute) transport. Thus, buoyancy-induced flow, caused by temperature differences, leads to a complex coupling among fluid flow, mass and heat transport, and mixing in single 3D channels. Obtaining accurate 3D temperature fields is essential for understanding this complex coupled system and determining how buoyancy-induced flow affects heat transfer in channel flows.

A number of previous studies have attempted to measure and visualize the temperature fields in 3D channels. For example, Patil and Vijay Babu13 experimentally studied mixed convection through a plain square duct in a laminar flow. However, the thermocouples installed at axial locations along the wall of the duct could only measure the inner wall temperature and did not capture the temperature fields at cross sections. Sandhu and Siddiqui14 successfully measured temperature fields at cross sections of a flat-plate solar collector tube, albeit with limited resolution. Even in controlled laboratory settings, measuring and visualizing the temperature fields in 3D channels pose significant challenges, especially when mixed convection flows are involved.

The transformation between temperature and concentration fields is feasible when the dimensionless governing equations for heat and mass transfer processes are analogous. Thus, one may envision inferring temperature fields from concentration fields because measuring the 3D solute concentration fields is typically easier than measuring the 3D temperature fields. In heat and mass transfer research, analogy functions have been widely used. When heat and mass transfer processes follow dimensionless governing equations of the same form, the temperature and concentration fields are correlated and can be interconverted, a phenomenon referred to as the heat and mass transfer analogy.15 For example, the temperature profile of the boundary layer can be derived from the concentration profile using the dimensionless governing equation or, in some cases, empirical analogy functions.16 The heat and mass transfer analogy has found extensive applications in determining heat transfer characteristics, such as heat transfer coefficients, in a variety of convection systems.17–21 However, it is difficult to mathematically derive governing equations or analogy functions when multiple nonlinear effects should be taken into consideration in a heat transfer problem, such as mixed convection flows.

Recently, a variety of fields, including medicine, health care, robotics, and basic research, have benefited from machine learning (ML).22–26 Due to the adoption of nonlinear activation functions, ML algorithms have the advantage of expressing highly nonlinear functions without being restricted by mathematical functional forms. In the area of fluid mechanics and heat transfer, several ML algorithms have been effectively implemented to predict and analyze mass and heat transfer characteristics in variable systems.27–29 For example, convolutional neural networks (CNNs) were trained to predict the 2D distribution of local heat flux in a fully developed turbulent channel flow when the model inputs were the 2D images of stress and pressure fields.30 CNNs leverage convolutional operations to extract features and expressions from high-dimensional data, making them well-suited for regression tasks using images.31,32 To handle sequential data or time series data, recurrent neural networks (RNNs) with a recurrent mechanism have been widely used.33,34 Recently, physics-informed neural networks (PINNs) have been introduced to infer the flow and pressure fields from limited snapshots of solute concentration fields.35 PINNs fit their solutions to the governing equations (e.g., conservation of mass, momentum, and energy), thereby reducing the dependence of neural networks on the size of the dataset.36–40 Generative adversarial networks (GANs) is another promising approach. With their unique network architectures, GANs can generate realistic and diverse data samples similar to those in the training, even without access to the original samples.41–44 For example, a previous study successfully employed GANs to generate the temperature fields in a 2D domain induced by heat conduction.45 Most importantly, GANs can effectively reduce the output space using conditional input, while CNNs can constrain their output only by adjusting their network parameters based on sophisticatedly designed loss functions. Therefore, GANs are often easier to train than CNNs and other deep neural networks.

In this study, we investigate the possibility of reconstructing the temperature fields directly from the concentration fields using a ML model. Considering the attractive characteristics of GANs, we employ the cGAN strategy. We trained the cGAN model using a series of concentration fields and corresponding temperature fields (training dataset). The cGAN model is trained to recognize the analogy function between two correlated scalar fields, concentration fields, and temperature fields. After training, the cGAN model is shown to successfully reconstruct the temperature fields from the input concentration fields (testing dataset).

Here, we consider heat and mass transfer in laminar channel flows. We consider a 3D channel with a width (W) of 60 mm, height (H) of 15 mm, and length (L) of 130 mm, as shown in Fig. 1. At an inlet, water at 35 °C ( T 0) enters with a uniform velocity distribution of 3 mm / s. The bottom wall temperature ( T b ) is maintained constantly at 43 °C, while the other walls are thermally insulated. Although the geometry and flow conditions are simplified and idealized, they serve as a good proxy for identifying critical flow and fluid-related factors affecting the mixed convection flow in a horizontal channel or fracture. A passive solute tracer is injected along with water, only at the bottom half of the inlet, to provide solute concentration in the channel. The Reynolds number ( R e = u l ν ) of the reference case in this study is 72, which is obtained using the characteristic length of the channel ( l = 2 W H W + H ), the injection velocity of water ( u ), and the kinematic viscosity of water ( ν ). The Rayleigh number ( R a = ρ β Δ T l 3 g η α ) of the reference case in this study is around 2.82 × 10 6, where ρ is the density of water, β is the thermal expansion coefficient of water, Δ T is the temperature difference between the bottom wall and the inlet, η is the dynamic viscosity of water, and α is the thermal diffusivity of water.

FIG. 1.

Geometry of the flow channel with concentration and temperature fields. At the inlet, water enters with a constant temperature of T 0 and a uniform velocity. A passive solute tracer is injected along with water at the bottom half of the inlet. The bottom wall is heated constantly with temperature T b, and the other walls are thermally insulated. (a) Passive tracer concentration fields at selected cross sections in a rectangular channel. (b) Temperature fields at selected cross sections in a rectangular channel. There is a close correlation between the concentration and temperature fields.

FIG. 1.

Geometry of the flow channel with concentration and temperature fields. At the inlet, water enters with a constant temperature of T 0 and a uniform velocity. A passive solute tracer is injected along with water at the bottom half of the inlet. The bottom wall is heated constantly with temperature T b, and the other walls are thermally insulated. (a) Passive tracer concentration fields at selected cross sections in a rectangular channel. (b) Temperature fields at selected cross sections in a rectangular channel. There is a close correlation between the concentration and temperature fields.

Close modal
The transient flow and temperature fields are simulated using an open-source computational fluid dynamics (CFD) software OpenFOAM46 developed by OpenCFD Ltd. buoyantSimpleFoam is a flow solver that solves buoyancy-driven flow of compressible fluids. The fluid flow in a channel can be described by Navier–Stokes (N–S) equations that consider the mass and momentum conservations,
(1)
(2)
where u is the velocity field, p is the pressure field, g is the gravitational acceleration, and t is the time. To simulate the heat transfer in the channel, energy conservation is also taken into account as
(3)
where K = | u | 2 / 2 is the kinetic energy per unit mass, h is the enthalpy per unit mass, and α e f f is the effective thermal diffusivity. We coupled the buoyantSimpleFoam with a solver for an advection-diffusion equation (scalarTransportFoam) following a similar approach as shown in Cao et al.47 Solute transport in the channel is described by the advection-diffusion equation (ADE),
(4)
where C is the passive tracer concentration and D is the diffusion coefficient. In summary, fluid flow and heat transfer are solved by buoyantSimpleFoam, and solute transport and mixing are solved by scalarTransportFoam.

To reduce computation costs, only the half-section of the channel is modeled by using the yz-plane at x = 0 as a symmetric plane. Also, we used a structured mesh considering the domain geometry and computational efficiency. The time step size of the simulation model is 0.05 s, and the number of time steps is 7000. Using supercomputing resources at Minnesota Supercomputing Institute (MSI), it took about 4 days to simulate one case using 128 cores and 400 GB RAM. The simulation outputs are the solute concentration fields and temperature fields in the simulation domain obtained at every 20 time steps (1 s). Although the simulation is 3D, training and testing datasets for the ML model are created using 2D cross-sectional images of the concentration fields and the corresponding temperature fields across xy-planes to reduce the model training time. These 2D cross-sectional images are obtained every 10 mm along the z direction. Therefore, the total simulation results are 4550 pairs of cross-sectional concentration and temperature images with a resolution of 180 × 360. Note that not all of the simulation results need to be employed to create training and testing datasets for the ML model. There are three cases that do not require many data samples. First, at the early stage of simulation, when the dye has not passed through the channel (first 10 s), most parts of the channel have no concentration variation. Second, at the late stage of simulation, the fluid flow reaches a steady state (after 190 s), leading to minimal changes in concentration and temperature fields. Last, changes in the concentration and temperature fields at the cross sections near the inlet (z < 20 mm) and near the outlet (z > 120 mm) are not significant. For these cases, a small number of data samples are needed for training and testing. With this consideration, the final datasets consist of 1980 pairs of cross-sectional concentration images and temperature images with a resolution of 180 × 360 obtained across the xy-planes. This dataset is denoted as the entire dataset in the following sections. The dataset obtained at cross sections where z = 90 and 110 mm are used for testing, while the other images are used for training.

Conditional generative adversarial networks (cGANs) have been used to predict transient temperature field images from numerical array inputs describing the channel geometry, thermal and hydraulic boundary conditions, location, and time.48 cGAN is a type of GAN that involves the conditional generation of images by a generator model.41 cGAN relies on a generator that learns to generate new images, and a discriminator that learns to distinguish synthetic images from real images. In cGANs, the generator network learns to create an image close to the ground truth image to fool the discriminator network. The discriminator network, on the other hand, learns to distinguish ground truth images from fake images that the generator network creates. Various types of generator and discriminator networks have been developed for cGANs. Here, we use cGANs to estimate transient temperature fields from graphical inputs describing concentration fields in a three-dimensional (3D) channel. The details of the generator and discriminator architectures we used can be found in the supplementary material.

In the cGAN model, the generator produces a temperature field (Y) from a provided input vector (c) that looks similar to the CFD simulation result (X) and tries to deceive the discriminator. The discriminator tries to distinguish X from Y. During training, the discriminator learns to maximize the probability of correctly classifying the images, while the generator tries to minimize this probability by generating realistic samples. Thus, the objective of a cGAN is
where D(c, X or Y) is the probability that the discriminator classifies X or Y as the ground truth for a given c. E c , X o r Y is the expected value over the entire group of X or Y. The generator attempts to minimize L c G A N, while the discriminator tries to maximize L c G A N.
To improve the ability of the generator, L1 loss is often added as a generator objective. The L1 loss function measures the absolute distance between two images: the generated image X and the ground truth image Y. The objective of the L1 loss function is to minimize the absolute difference between the generated and ground truth images. Thus, if the generator is trained to minimize the L1 loss, the produced images become clearer than the results without the L1 objective. The L1 loss is formulated as
Thus, the final objective of a cGAN is described as
where the hyperparameter λ is a weight for L L 1 ( G ).

We optimized hyperparameters such as training epochs and λ to balance the training quality and computation duration. It took about 8 h to train the cGAN model using the GPU of a high-performance workstation (Dell Precision 7920 Tower). Training the cGAN model takes a much shorter time than CFD simulations. Therefore, having such a ML model can significantly lower computational costs than conventional numerical simulation techniques. Figure 2 shows the training losses of the optimized (reference) model as a function of epochs. From Fig. 2, we can see that L D, L L 1, and L T o t a l gradually decrease with the increase in epoch. L c G A N slightly increases over the epoch but remains relatively stable in general. The hyperparameter λ balances the two objectives for generator training. If λ increases, the generator focuses more on minimizing the mean absolute difference between X and Y, thereby improving the low-frequency accuracy. Considering the training duration, we conducted all the subsequent model training with 1000 epochs and λ = 104.

FIG. 2.

Training losses of the optimized model as a function of epochs.

FIG. 2.

Training losses of the optimized model as a function of epochs.

Close modal

The image contrast, determined by concentration and temperature gradients, can affect model accuracy. A previous work reported that the cGAN model tended to generate blurred features in the temperature fields when the Michelson image contrast was below 0.011.48 Additionally, concentration and temperature fields significantly change with location. For example, concentration and temperature gradients are smaller at the cross sections near the outlet because of flow mixing. Consequently, how the field images are sampled from the entire dataset will determine the model’s accuracy and generalizability. To investigate the sensitivity of data sampling on reconstruction accuracy, we trained two additional models and compared their results with those of the model trained by the entire dataset, denoted as the reference model. In the first additional model (denoted as model 1), pairs of concentration and temperature images sampled at z = 20–50 mm are used for training, while the remaining data samples are used for testing. In the second additional model (denoted as model 2), pairs of concentration and temperature images sampled at z = 90–120 mm are used for training, and the remaining images are used for testing. This means that in model 1, the training dataset is sampled closer to the inlet, while in model 2, the training dataset is sampled closer to the outlet.

In Fig. 3, the results of the different models are qualitatively compared at an early stage, t = 50 s, for selected z locations. The first row shows the concentration fields at different cross sections from simulation, which are used as input data in the ML model. The second row shows the temperature fields at different cross sections from simulation, serving as ground truth for comparison with the reconstructed results from the ML model. As expected, the concentration and temperature gradients are larger near the inlet (z = 40 mm) compared to those near the outlet (z = 100 mm). The third to fifth rows show the reconstructed results from different models.

FIG. 3.

Qualitative comparison of the results of different models at t = 50 s, for selected z locations.

FIG. 3.

Qualitative comparison of the results of different models at t = 50 s, for selected z locations.

Close modal

Results show that the reference model performed the best because it was trained using the figures from both near the inlet and near the outlet. The reference model can reconstruct the temperature fields across the channel, regardless of the gradient. Model 1 was trained with the images near the inlet, where the concentration and temperature gradients are larger. Therefore, at the cross section near the inlet (z = 40 mm), the reconstructed result from model 1 is better than that from model 2. However, model 1 fails to reconstruct the temperature fields at the cross section near the outlet (z = 100 mm) because it did not have enough experience inferring the temperature fields from the concentration fields, when the gradient is small. Overall, model 2 performed better than model 1.

Error metrics are commonly used to evaluate the performance of ML models by comparing predicted values with actual values. Typical error metrics include the mean squared error (MSE), root mean squared error (RMSE), mean absolute error (MAE), and R-squared (R2). Each error metric serves a distinct purpose and provides insights into different aspects of the model performance. Here, we focus on RMSE because it is a standardized error metric that offers a measure of the average magnitude of errors in the same units as the target variable. This facilitates the interpretation and comparison of different models.

Figure 4 shows how RMSE changes with locations at three different times for different models. Black lines indicate the reference model, blue lines indicate model 1, and red lines indicate model 2. Results generally show that model 2 exhibits smaller RMSE than model 1, indicating the better performance of model 2, consistent with our qualitative comparison with Fig. 3. For model 1, the reconstruction zone (or unseen zone) is when z = 50–120 mm (not including z = 50 mm), while for model 2, the reconstruction zone is when z = 20–90 mm (not including z = 90 mm). In the reconstruction zone, the maximum RMSE ranges from 0.19 to 0.24 for model 1 at different times, whereas for model 2, the maximum RMSE ranges from 0.1 to 0.19 at different times. Based on the quantitative and qualitative comparisons, model 2 is considered to have superior performance to model 1. This result suggests that using the figures with smaller concentration and temperature gradients but with more complex spatial patterns as the training dataset can help the ML model in better inferring the temperature fields from the concentration fields. For channel flows, one may benefit from better measurement of the concentration fields near the outlet in order to estimate the temperature fields across the channel.

FIG. 4.

RMSE changes as a function of location at three different times for different models.

FIG. 4.

RMSE changes as a function of location at three different times for different models.

Close modal

From Fig. 4, we observed that the RMSE tends to increase with time for all three models. At later times, because of mixing, concentration and temperature gradients decrease. Therefore, models trained with data at early times, where the concentration and temperature gradients are large without exhibiting much unique features in fields, may not perform well when reconstructing the temperature fields at later times. However, the capability of reconstructing the temperature fields at later times is important for most applications. To investigate the possibility of using early-time data to reconstruct late-time temperature fields, an additional model (denoted as model 3) was trained using simulation results from the first 40 s as the training dataset, while results at 50 s and 100 s were used as the testing dataset. The reconstructed results are shown in Fig. 5. Based on qualitative analysis, the accuracy of model 3 when reconstructing the temperature fields at later times is relatively poor, especially near the outlet. At later times and near the outlet, due to increased mixing, the concentration and temperature gradients are smaller with more unique field features than at the early times. Therefore, a model trained with data at early times has a hard time reconstructing the temperature fields at later times.

FIG. 5.

Qualitative comparison of the reconstructed results of model 3 at different times at selected z locations.

FIG. 5.

Qualitative comparison of the reconstructed results of model 3 at different times at selected z locations.

Close modal

Figure 6 shows how the RMSE changes with time for the reference model and model 3. At the cross section near the inlet, the RMSE of the reference model gradually increases with time. At the cross sections in the middle of the channel and near the outlet, the RMSE of the reference model stays constant at early times but increases at later times. For model 3, the RMSE increases dramatically at around 50 s. After 50 s, the change in RMSE with time is not very significant. Although the qualitative analysis in Fig. 5 shows that the accuracy of model 3 when reconstructing the temperature fields at later times is relatively not good, the RMSE at late times from Fig. 6 is about 0.14, which is acceptable. We can infer that the cGAN has the capability of reconstructing the temperature fields at later times even when we only have data at early times, although the result may not be highly accurate. The cGAN will exhibit better extrapolation capability in the time domain when there is a greater similarity in field features between early times and later times.

FIG. 6.

RMSE changes as a function of time for the reference model and model 3 at three different locations.

FIG. 6.

RMSE changes as a function of time for the reference model and model 3 at three different locations.

Close modal

The results so far suggest that the ML approach can be a powerful tool for overcoming the limitations of traditional heat and mass analogy approaches. In particular, we confirmed the possibility of using an ML model to infer the temperature fields from the concentration fields for the laminar mixed convection in a 3D horizontal channel. However, ensuring the generalization of ML models to untrained conditions, such as different flow regimes or geometries, remains a major challenge. Developing models that are more adaptable to varying convection scenarios is an important topic. In this section, we test the applicability of our model in different convection scenarios.

In thermal convection problems, the governing nondimensional parameter for describing convection scenarios is the Richardson number (Ri). The Richardson number (Ri), which is the ratio of the Grashof number (Gr) to the Reynolds number (Re), represents the importance of natural convection relative to forced convection. Typically, natural convection is negligible when Ri < 0.1, forced convection is negligible when Ri > 10, and both are not negligible when 0.1 < Ri < 10. The Richardson number is defined as
where g is the gravitational acceleration, β is the thermal expansion coefficient, T h o t is the hot wall temperature, T r e f is the reference temperature, l is the characteristic length, and v is the characteristic velocity. For the rectangular channel we used, with a width (W) of 60 mm and a height (H) of 15 mm, the characteristic length is l = 4 W H 2 ( W + H ) = 24 mm = 0.024 m. The inlet flow velocity is taken as the characteristic velocity and the thermal expansion coefficient ( β ) is 1.6 × 10 5 + 9.6 × 10 6 × T a v g.

To test the applicability of our model to different convection scenarios, we simulated various scenarios by changing the temperature differences T h o t T r e f and characteristic velocities ( v ). The goal is to test the applicability of the reference model, trained at specific Ra and Re, to scenarios with different Ra and Re. Table I shows the Ri, Ra, and Re for different cases. The Ri for the case we used to train our model (reference case) is marked in red. Subsequently, we employed the trained reference model to reconstruct the temperature fields of various cases and calculated the averaged RMSE. Table II shows the averaged RMSE for different cases, with the average RMSE for the reference case marked in red. To better visualize how the RMSE changes across different convection scenarios, Fig. 7 is plotted.

FIG. 7.

Map of the averaged RMSE at different convection scenarios and a few representative cases (red circles) showing the corresponding concentration and temperature fields. A reference model (red star) trained under mixed convection conditions can be applicable for reconstructing the cases under different flow regimes. Model accuracy is more sensitive to Ra than to Re.

FIG. 7.

Map of the averaged RMSE at different convection scenarios and a few representative cases (red circles) showing the corresponding concentration and temperature fields. A reference model (red star) trained under mixed convection conditions can be applicable for reconstructing the cases under different flow regimes. Model accuracy is more sensitive to Ra than to Re.

Close modal
TABLE I.

Ri, Ra, and Re for different cases. The case in the red cell is the reference case.

Ri v ( mm / s )3.1256.2512.52550100
Δ T ( ° C ) Ra    Re 75 150 300 600 1200 2400 
0.654 × 106 17.44 4.36 1.09 0.27 0.07 0.02 
1.342 × 106 35.80 8.95 2.24 0.56 0.14 0.03 
2.066 × 106 55.08 13.77 3.44 0.86 0.22 0.05 
2.824 × 106 75.30 18.82 4.71 1.18 0.29 0.07 
10 3.616 × 106 96.44 24.11 6.03 1.51 0.38 0.09 
Ri v ( mm / s )3.1256.2512.52550100
Δ T ( ° C ) Ra    Re 75 150 300 600 1200 2400 
0.654 × 106 17.44 4.36 1.09 0.27 0.07 0.02 
1.342 × 106 35.80 8.95 2.24 0.56 0.14 0.03 
2.066 × 106 55.08 13.77 3.44 0.86 0.22 0.05 
2.824 × 106 75.30 18.82 4.71 1.18 0.29 0.07 
10 3.616 × 106 96.44 24.11 6.03 1.51 0.38 0.09 
TABLE II.

The averaged RMSE for different cases. The case in the red cell is the reference case.

RMSE v ( mm / s )3.1256.2512.52550100
Δ T ( ° C ) Ra   Re 75 150 300 600 1200 2400 
0.654 × 106 0.0696 0.0713 0.0715 0.0716 0.0714 0.0715 
1.342 × 106 0.0577 0.0560 0.0598 0.0525 0.0581 0.059 
2.066 × 106 0.0675 0.066 0.0647 0.0439 0.0581 0.0534 
2.824 × 106 0.0866 0.0813 0.0796 0.0723 0.0777 0.0823 
10 3.616 × 106 0.0958 0.094 0.089 0.0827 0.0877 0.0913 
RMSE v ( mm / s )3.1256.2512.52550100
Δ T ( ° C ) Ra   Re 75 150 300 600 1200 2400 
0.654 × 106 0.0696 0.0713 0.0715 0.0716 0.0714 0.0715 
1.342 × 106 0.0577 0.0560 0.0598 0.0525 0.0581 0.059 
2.066 × 106 0.0675 0.066 0.0647 0.0439 0.0581 0.0534 
2.824 × 106 0.0866 0.0813 0.0796 0.0723 0.0777 0.0823 
10 3.616 × 106 0.0958 0.094 0.089 0.0827 0.0877 0.0913 

Figure 7 shows how the RMSE changes with Ra and Re. In Fig. 7, colors represent the value of RMSE, where yellow indicates a large RMSE while blue means a small RMSE. The reference case is marked by a red star. Four representative cases with different Re and Ra are shown in Fig. 7 as red circles and the corresponding images are presented. From Fig. 7, it can be observed that the model trained under mixed convection conditions can still be applicable for reconstructing the cases under natural or forced convection conditions. The maximum RMSE in this figure is about 0.1, which is acceptable. With an increase in Ra, the RMSE increases significantly, while changes in Re have less effect on the RMSE. The impact of changes in Re on the RMSE is more pronounced in regions with higher Ra compared to regions with smaller Ra. This indicates that model accuracy is more sensitive to changes in Ra than changes in Re in the corresponding regimes of Ra and Re. Thus, when using the trained model on a case with different Δ T or velocities, the difference in Δ T between the trained case and the test case will have a greater impact on model performance than the difference in velocity. In general, the results show that the ML approach can be used to infer temperature fields in unseen flow scenarios, especially when Ra does not vary significantly.

To understand the complex convective transport processes, both mass concentration and temperature fields are critical information. However, experimentally measuring the temperature field is more challenging and requires sophisticated experimental setups. Our results in this section show that the trained model can be applied to different flow conditions. This implies that if we have the concentration and temperature datasets of one experiment, we could use the dataset to train the model and reconstruct the temperature fields of different experiments under different flow conditions. The concentration fields of different experiments need only be measured, which is relatively easier than measuring the temperature fields.

Through the use of machine learning, we showed the possibility of inferring 3D temperature fields from 3D concentration fields for laminar mixed convection in 3D horizontal channel flows. Due to the scarcity of available experimental data, we conducted 3D numerical simulations to generate sufficient data on concentration fields and temperature fields for training and testing a cGAN model. Our study demonstrated that 3D temperature fields can be accurately estimated from 3D concentration fields. We further investigated the effects of a training dataset on model performance and found that the reconstruction accuracy of the ML approach is sensitive to the sampling of the training dataset. ML models trained with data exhibiting smaller concentration and temperature gradients, but spatially more complex patterns, are shown to perform better. Furthermore, we tested the applicability of our model to different convection scenarios by employing the trained model to reconstruct the temperature fields of various cases under different convection scenarios. The concentration and temperature fields at different convection scenarios are obtained by simulating the cases with varying degrees of temperature difference and injection velocity. When applying the trained model to different cases under different convection scenarios, the maximum RMSE obtained was 0.1, which is acceptable. Additionally, model accuracy was found to be more sensitive to changes in Ra than changes in Re. This study shows the promise of using machine learning for inferring 3D temperature fields, which is important for a wide range of applications.

The supplementary material describes how the reference model parameters were optimized and selected. In particular, we address the generator and discriminator architectures of conditional generative adversarial networks (cGANs).

The authors acknowledge the support of the National Science Foundation (NSF) under Grant No. CBET-2053413.

The author has no conflicts to disclose.

Hongfan Cao: Data curation (lead); Formal analysis (equal); Investigation (equal); Methodology (equal); Software (equal); Visualization (lead); Writing – original draft (lead). Beomjin Kwon: Conceptualization (equal); Formal analysis (supporting); Funding acquisition (supporting); Investigation (equal); Project administration (supporting); Software (supporting); Supervision (equal); Visualization (equal); Writing – review & editing (equal). Peter K. Kang: Conceptualization (lead); Formal analysis (supporting); Funding acquisition (lead); Investigation (equal); Methodology (equal); Project administration (equal); Resources (equal); Supervision (lead); Visualization (equal); Writing – review & editing (equal).

The data that support the findings of this study are available from the corresponding author upon reasonable request.

1.
A.
Bejan
,
Convection Heat Transfer
, 2nd ed. (
Wiley
,
New York
,
1995
).
2.
W. M.
Kays
and
M. E.
Crawford
,
Convective Heat and Mass Transfer
, 4th ed. (
McGraw Hill
,
New York
,
2005
).
3.
T. L.
Bergman
,
A. S.
Lavine
,
F. P.
Incropera
, and
D. P.
Dewitt
,
Fundamentals of Heat and Mass Transfer
, 7th ed. (
Wiley
,
New York
,
2011
).
4.
D. G.
Osborne
and
F. P.
Incropera
, “
Laminar, mixed convection heat transfer for flow between horizontal parallel plates with asymmetric heating
,”
Int. J. Heat Mass Transfer
28
,
207
217
(
1985
).
5.
S.
Ostrach
and
Y.
Kamotani
, “
Heat transfer augmentation in laminar fully developed channel flow by means of heating from below
,”
J. Heat Transfer
97
,
220
225
(
1975
).
6.
M.
Yasuo
and
U.
Yutaka
, “
Forced convective heat transfer between horizontal flat plates
,”
Int. J. Heat Mass Transfer.
9
,
803
817
(
1966
).
7.
A. E.
Bergles
and
R. R.
Simonds
, “
Combined forced and free convection for laminar flow in horizontal tubes with uniform heat flux
,”
Int. J. Heat Mass Transfer
14
,
1989
2000
(
1971
).
8.
S. T.
McComas
and
E. R. G.
Eckert
, “
Combined free and forced convection in a horizontal circular tube
,”
J. Heat Transfer
88
,
147
152
(
1966
).
9.
X.
Chen
,
C.
Wang
,
Y.
Wu
,
C.
Chen
, and
C.
Ma
, “
Numerical simulation of mixed convection heat transfer of molten salt in horizontal square tube with single surface heating
,”
Appl. Therm. Eng.
104
,
282
293
(
2016
).
10.
F.
Koffi
,
C.
Abid
,
M.
Medale
, and
F.
Papini
, “
Mixed convection flow in a horizontal rectangular channel subjected to a horizontal thermal gradient
,”
Heat Mass Transfer
47
,
1251
1260
(
2011
).
11.
C.
Gau
,
Y. C.
Jeng
, and
C. G.
Liu
, “
An experimental study on mixed convection in a horizontal rectangular channel heated from a side
,”
J. Heat Transfer
122
,
701
707
(
2000
).
12.
T.
Maré
,
N.
Galanis
,
I.
Voicu
, and
J.
Miriel
, “
Experimental analysis of mixed convection in inclined tubes
,”
Appl. Therm. Eng.
26
,
1677
1683
(
2006
).
13.
S. V.
Patil
and
P. V.
Vijay Babu
, “
Experimental studies on mixed convection heat transfer in laminar flow through a plain square duct
,”
Heat Mass Transfer
48
,
2013
2021
(
2012
).
14.
G.
Sandhu
and
K.
Siddiqui
, “
Investigation of the fluid temperature field inside a flat-plate solar collector
,”
Heat Mass Transfer
50
,
1499
1514
(
2014
).
15.
T. L.
Bergman
,
A. S.
Lavine
,
F. P.
Incropera
, and
D. P.
Dewitt
,
Fundamentals of Heat and Mass Transfer
(
John Wiley & Sons, Inc.
,
Hoboken, NJ
,
2011
).
16.
S. W.
Churchill
and
H.
Ozoe
, “
Correlations for laminar forced convection in flow over an isothermal flat plate and in developing and fully developed flow in an isothermal tube
,”
J. Heat Transfer
95
,
416
419
(
1973
).
17.
S.
Aravinth
, “
Prediction of heat and mass transfer for fully developed turbulent fluid flow through tubes
,”
Int. J. Heat Mass Transfer
43
,
1399
1408
(
2000
).
18.
J.
Devera
,
T.
Hyhlík
, and
R.
Kalinay
, “
Similarity of temperature and concentration field in Poiseuille–Benard channel flow
,”
EPJ Web Conference
(EDP Sciences,
2019
), Vol.
213
, p.
02012
.
19.
R.
Kumar
and
S. P.
Mahulikar
, “
Numerical re-examination of Chilton-Colburn analogy for variable thermophysical fluid properties
,”
J. Heat Transfer
139
,
071701
(
2017
).
20.
S.
Han
and
R. J.
Goldstein
, “
The heat/mass transfer analogy for a simulated turbine blade
,”
Int. J. Heat Mass Transfer
51
(
21–22
),
5209
5225
(
2008
).
21.
S.
Han
and
R. J.
Goldstein
, “
The heat/mass transfer analogy for a simulated turbine endwall
,”
Int. J. Heat Mass Transfer
51
(
11–12
),
3227
3244
(
2008
).
22.
B. J.
Erickson
,
P.
Korfiatis
,
Z.
Akkus
, and
T. L.
Kline
, “
Machine learning for medical imaging
,”
Radiographics
37
,
505
515
(
2017
).
23.
J.
Bell
,
Machine learning : hands-on for developers and technical professionals
(John Wiley & Sons,
2014
).
24.
W. D.
Smart
and
L. P.
Kaelbling
, “
Effective reinforcement learning for mobile robots
,” in
Proceedings 2002 IEEE International Conference on Robotics & Automation
(IEEE,
2002
), Vol.
4
, pp.
3404
3410
.
25.
A. L.
Beam
and
I. S.
Kohane
, “
Big data and machine learning in health care
,”
J. Am. Med. Assoc.
319
,
1317
1318
(
2018
).
26.
S. L.
Brunton
,
B. R.
Noack
, and
P.
Koumoutsakos
, “
Machine learning for fluid mechanics
,”
Annu. Rev. Fluid Mech.
52
,
477
508
(
2020
).
27.
C.
Wang
,
C.
Xu
,
C.
Wang
, and
D.
Tao
, “
Perceptual adversarial networks for image-to-image transformation
,”
IEEE Trans. Image Process.
27
(
8
),
4066
(
2018
).
28.
B.
Kwon
,
F.
Ejaz
, and
L. K.
Hwang
, “
Machine learning for heat transfer correlations
,”
Int. Commun. Heat Mass Transfer
116
,
104694
(
2020
).
29.
M.
Kang
and
B.
Kwon
, “
Deep learning of forced convection heat transfer
,”
ASME J. Heat Transfer
144
(
2
)
021801
(
2022
).
30.
J.
Kim
and
C.
Lee
, “
Prediction of turbulent heat transfer using convolutional neural networks
,”
J. Fluid Mech.
882
A18
(
2020
).
31.
R.
Sharma
,
A. B.
Farimani
,
J.
Gomes
,
P.
Eastman
, and
V.
Pande
, “Weakly-supervised deep learning of heat transport via physics informed loss,” arXiv:1807.11374 (2018).
32.
J.
Dai
,
Y.
Lu
, and
Y. N.
Wu
, “Generative modeling of convolutional neural networks,” arXiv:1412.6296 (2014).
33.
N.
Srivastava
,
E.
Mansimov
, and
R.
Salakhutdinov
, “
Unsupervised learning of video representations using LSTMs
,” in
International Conference on Machine Learning
(PMLR,
2015
), pp.
843
852
.
34.
X.
Shi
,
Z.
Chen
,
H.
Wang
,
D. Y.
Yeung
,
W. K.
Wong
, and
W. C.
Woo
, “
Convolutional LSTM network: A machine learning approach for precipitation nowcasting
,”
Adv. Neural Inf. Process. Syst.
1
,
802
810
(
2015
).
35.
M.
Raissi
,
A.
Yazdani
, and
G. E.
Karniadakis
, “
Hidden fluid mechanics: Learning velocity and pressure fields from flow visualizations
,”
Science
367
(
6481
),
1026
1030
(
2020
).
36.
G.
Han
,
L.
Sun
, and
J.
Wang
, “
Phygeonet: Physics-informed geometry-adaptive convolutional neural networks for solving parameterized steady-state PDEs on irregular domain
,”
J. Comput. Phys.
428
,
110079
(
2021
).
37.
W.
Nils
,
M.
Weinmann
,
M.
Neidlin
, and
R.
Klein
, “
Spline-PINN: Approaching PDEs without data using fast, physics-informed hermite-spline CNNs
,” in
Proceedings of the AAAI Conference on Artificial Intelligence
(AAAI,
2022
), Vol.
36
, no.
8
, pp.
8529
8538
.
38.
G.
Viktor
,
A.
Heinlein
, and
A.
Klawonn
, “Learning the solution operator of two-dimensional incompressible Navier-Stokes equations using physics-aware convolutional neural networks,” arXiv:2308.02137 (2023).
39.
S.
Cai
,
Z.
Wang
,
S.
Wang
,
P.
Perdikaris
, and
G. E. M.
Karniadakis
, “
Physics-informed neural networks for heat transfer problems
,”
J. Heat Transfer
143
(
6
),
060801
(
2021
).
40.
N.
Zobeiry
and
K. D.
Humfeld
, “
A physics-informed machine learning approach for solving heat transfer equation in advanced manufacturing and engineering applications
,”
Eng. Appli. Artif. Intell.
101
,
104232
(
2021
).
41.
M.
Mirza
and
S.
Osindero
, “Conditional generative adversarial nets,” arXiv:1411.1784 (2014).
42.
J.
Goodfellow
,
J.
Pouget-Abadie
,
M.
Mirza
,
B.
Xu
,
D.
Warde-Farley
,
S.
Ozair
,
A.
Courville
, and
Y.
Bengio
, “
Generative adversarial nets
,”
Adv. Neural Inf. Process. Syst.
27
, 2672–2680 (
2014
).
43.
P.
Isola
,
J.-Y.
Zhu
,
T.
Zhou
, and
A. A.
Efros
, “
Image-to-image translation with conditional adversarial networks
,” in
Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition
(IEEE,
2017
), pp.
1125
1134
.
44.
Z.
Yi
,
H.
Zhang
,
P.
Tan
, and
M.
Gong
, “
DualGAN: Unsupervised dual learning for image-to-image translation,
” in
Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition
(IEEE,
2017
), pp.
2868
2876
.
45.
A. B.
Farimani
,
J.
Gomes
, and
V. S.
Pande
, “Deep learning the physics of transport phenomena,” arXiv:1709.02432 (2017) (accessed July 10, 2020).
46.
H. G.
Weller
,
G.
Tabor
,
H.
Jasak
, and
C.
Fureby
, “
A tensorial approach to computational continuum mechanics using object-oriented techniques
,”
Comput. Phys.
12
(
6
),
620
631
(
1998
).
47.
H.
Cao
,
S.
Yoon
,
Z.
Xu
,
L. J.
Pyrak-Nolte
,
E.
Bresciani
, and
P. K.
Kang
, “
Emergence of unstable focused flow induced by variable-density flows in vertical fractures
,”
Water Resour. Res.
59
(
12
),
e2023WR034729
(
2023
).
48.
M.
Kang
,
N. P.
Nguyen
, and
B.
Kwon
, “
Deep learning model for rapid temperature map prediction in transient convection process using conditional generative adversarial networks
,”
Therm. Sci. Eng. Prog.
49
,
102477
(
2024
).