Uploaded by Facaprom Tech

FULLTEXT02

advertisement
Mälardalen University Doctoral Dissertation 399
Jerol Soibam MACHINE LEARNING TECHNIQUES FOR ENHANCED HEAT TRANSFER MODELLING
2024
ISBN 978-91-7485-625-5
ISSN 1651-4238
Address:
P.O. Box 883, SE-721 23 Västerås. Sweden
P.O. Box 325, SE-631 05 Eskilstuna. Sweden
E-mail: info@mdu.se Web: www.mdu.se
Machine Learning Techniques
for Enhanced Heat Transfer
Modelling
Jerol Soibam
!" #$ %&'(')
*+,-+),./'..
*-/.-)',
% 01&2 Mälardalen University Press Dissertations
No. 399
MACHINE LEARNING TECHNIQUES FOR ENHANCED HEAT TRANSFER MODELLING
Jerol Soibam
Akademisk avhandling
som för avläggande av teknologie doktorsexamen i energi- och miljöteknik vid
Akademin för ekonomi, samhälle och teknik kommer att offentligen försvaras
tisdagen den 13 februari 2024, 09.00 i Delta, Mälardalens universitet, Västerås.
Fakultetsopponent: Andrea Ianiro, University Charles III of Madrid
Akademin för ekonomi, samhälle och teknik
Abstract
With the continuous growth of global energy demand, processes from power generation to electronics
cooling become vitally important. The role of heat transfer in these processes is crucial, facilitating
effective monitoring, control, and optimisation. Therefore, advancements and understanding of heat
transfer directly correlate to system performance, lifespan, safety, and cost-effectiveness, and they serve
as key components in addressing the world's increasing energy needs.
The field of heat transfer faces the challenge of needing intensive studies while retaining fast
computational speeds to allow for system optimisation. While advancements in computational power
are significant, current numerical models lack in handling complex physical problems such as illposed. The domain of heat transfer is rapidly evolving, driven by a wealth of data from experimental
measurements and numerical simulations. This data influx presents an opportunity for machine learning
techniques, which can be used to harness meaningful insights about the underlying physics.
Therefore, the current thesis aims to the explore machine learning methods concerning heat transfer
problems. More precisely, the study looks into advanced algorithms such as deep, convolutional,
and physics-informed neural networks to tackle two types of heat transfer: subcooled boiling and
convective heat transfer. The thesis further addresses the effective use of data through transfer learning
and optimal sensor placement when available data is sparse, to learn the system behaviour. This
technique reduces the need for extensive datasets and allows models to be trained more efficiently. An
additional aspect of this thesis revolves around developing robust machine learning models. Therefore,
significant efforts have been directed towards accounting for the uncertainty present in the model,
which can further illuminate the model’s behaviour. This thesis shows the machine learning model's
ability for accurate prediction. It offers insights into various parameters and handles uncertainties and
ill-posed problems. The study emphasises machine learning's role in optimising heat transfer processes.
The findings highlight the potential of synergistic application between traditional methodologies and
machine learning models. These synergies can significantly enhance the design of systems, leading to
greater efficiency.
ISBN 978-91-7485-625-5
ISSN 1651-4238
To my family,
Acknowledgements
I would like to express my deep gratitude to my principal supervisor, Prof.
Rebei Bel Fdhila. His guidance and unwavering support have been invaluable throughout my research journey. I am particularly grateful for his
confidence in my ideas and the freedom he has given me to explore them.
I am thankful to my co-supervisor, Prof. Konstantinos Kyprianidis, for his
guidance and encouragement. His innovative ideas and thought-provoking
discussions have greatly enriched my research. His insightful guidance
has shaped my academic journey, offering constant inspiration and learning opportunities.
Profound thanks to my co-supervisor, Assoc Prof. Ioanna Aslanidou,
whose support has been crucial both academically and personally during
my PhD journey. Her encouragement in times of challenge was a beacon
of hope, keeping me focused and resilient. Working with her has been an
exceptionally positive and insightful experience.
I would also like to express my gratitude to my manager, Dr. Anders
Avelin, for his extensive support and guidance throughout my PhD journey. His encouragement to pursue new opportunities has been invaluable.
A heartfelt thanks to my colleagues in the department for the meaningful
conversations during lunch and fika breaks, which have significantly enhanced my experience throughout my time at the university.
I seize this moment to express my gratitude to Amare, Valentin, and Achref
for engaging conversations and enjoyable collaboration we have had.
I also take this opportunity to thank Dimitra-Eirini Diamantidou, who has
stood by my side through thick and thin. Her unwavering support and love
have been a constant source of motivation, especially in times when I lost
hope in myself. Thank you for the enriching conversations on life as well
as research, and for the wonderful trips we have had over the years. I look
forward to more of them.
Mälardalen University Press Dissertations
i
I extend my immense gratitude to my siblings and parents, who have been
my greatest support in life. Thank you for constantly believing in me, for
your endless patience, and for the warmth that our family shares. Your
love, sacrifices, and encouragement have given me the strength to keep
moving forward and to make the best out of every situation. I am eternally
grateful for the faith you have consistently shown in me.
Jerol Soibam
Västerås, Sweden, September 2023
ii
Jerol Soibam
Summary
The growing global energy demands and environmental impacts necessitate efficient industrial systems. The systems’ effectiveness heavily depends on the control of fluid flows and heat transfer. Methods such as
detailed numerical simulations have been utilised over the years, despite
their time-consuming nature and the need for intricate modelling. Such
simulations have generated significant volumes of data. The accumulated
data sets the stage for the application of machine learning techniques,
which provide a promising solution for real-time monitoring and management of heat transfer systems.
This thesis comprises several studies which focuses on the potential
use of machine learning for heat transfer problems. More specifically,
various machine learning architectures have been investigated to study
subcooled boiling and convective heat transfer. In the segment dealing
with boiling heat transfer, the research explores various methodologies
for predicting crucial parameters such as wall temperature and void fraction in subcooled flow boiling heat transfer. A deep neural network serves
as the core computational tool in this exploration. The model’s input incorporates the near local flow parameters of the heated minichannel under
subcooled boiling. The predictions obtained from the model demonstrate
notable agreement with numerical simulation data, reinforcing the efficacy of the employed machine learning techniques in modelling complex
thermal dynamics. Moreover, the study focuses on predicting the uncertainty present in the deep learning model while estimating the quantities
of interest.
The thesis also unveils a novel convolutional neural network model
for tracking bubble dynamics in a mini-channel, heated from one side.
This model utilises high-speed camera images as its primary input, despite
the inherent challenges in the form of shadows, reflections, and noise.
Through transfer learning, the model successfully detects and classifies
the bubble boundaries with impressive accuracy. Moreover, it can estimate local parameters delivering a detailed understanding of the spatialtemporal behaviour of the bubbles, an achievement that could be instrumental for various industrial applications.
The critical heat flux, a crucial parameter in the design of any industrial boiling system, is another focal point of this thesis. A parametric
Gaussian process regression approach is introduced to predict critical heat
Mälardalen University Press Dissertations
iii
flux with significantly reduced predictive error compared to traditional
empirical methods and look-up tables. This model provides robust insights and aligns well with the underlying physics, emphasising the potential of machine learning in overcoming the challenges of empirical correlation.
In the context of convective heat transfer, the thesis presents a strategy
to handle ill-posed boundary conditions when there is a lack of accurate
thermal boundary conditions. To tackle this challenge, the study employs
a physics-informed neural network. As the thermal boundary is unknown
near the source, the network has been optimised using only a few sensors points to discover the thermal boundary. It simultaneously represents
velocity and temperature fields while enforcing the Navier-Stokes and energy equations at random points within the domain, thereby functioning
as an inverse problem. The aim is to reproduce the global flow field and
temperature profile with sparse data. Furthermore, the research employs
transfer learning to account for different parameters, such as the position
and size of the source term within the enclosure domain.
To enhance the model’s robustness, the QR pivoting technique is employed to determine the optimal sensor location within the domain. Most
importantly, the sensors derived from QR pivoting ensure the effective
capture of the dynamics, leading to enhanced model accuracy when integrated with the physics-informed neural network. To further enhance the
model’s robustness and generalisability, an ensemble physics-informed
neural network is implemented. This method serves two essential purposes, it mitigates the risks associated with overfitting and underfitting,
which are prevalent in machine learning models. Furthermore, it accounts
for inherent uncertainty in model predictions, offering a measure of confidence in the model’s outputs. This acknowledgement and quantification
of uncertainty not only increases the robustness of the model but also enhances transparency, which refers to the model’s ability to indicate the reliability of its predictions. Consequently, the model can pinpoint regions
of reliable prediction and potential inaccuracies, broadening its applicability in addressing complex heat transfer problems with unknown boundary
conditions. The methods proposed in this study have demonstrated good
agreement with the underlying physics represented by numerical results.
In conclusion, this research paves the way for innovative solutions in
heat transfer studies, harnessing the capabilities of machine learning techniques. The results demonstrate that these methodologies can handle the
complexities and uncertainties inherent in predicting heat transfer phenomena, suggesting promising opportunities for further exploration and
potential industrial applications.
iv
Jerol Soibam
Sammanfattning
Den ökande globala energiefterfrågan och miljöpåverkan kräver effektiva
industriella system. Dessa systemens effektivitet beror i stor utsträckning på kontrollen av vätskeflöden och värmeöverföring. Under åren har
metoder som detaljerade numeriska simuleringar använts, trots att de är
tidskrävande och kräver komplicerad modellering. Dessa simuleringar har
genererat enorma datamängder. Den ackumulerade data lägger grunden
för användningen av maskininlärningstekniker, som erbjuder en lovande
lösning för realtidsövervakning och hantering av värmeöverföringssystem.
Denna avhandling omfattar flera studier som riktar in sig på den potentiella användningen av maskininlärning för värmeöverföringsproblem.
I avhandlingen har olika arkitekturer av maskininlärning undersökts för
att studera underkyld kokning och konvektiv värmeöverföring. I det avsnitt som handlar om kokningsvärmeöverföring utforskas olika metoder för
att förutsäga avgörande parametrar, såsom väggtemperatur och tomfraktion vid underkyld flödeskokning. Ett djupt neuralt nätverk är det centrala verktyget i denna utforskning. Modellen tar in data från närliggande
flödesparametrar i den uppvärmda minikanalen under underkyld kokning.
Förutsägelserna från modellen stämmer väl överens med numerisk simulering, vilket bekräftar maskininlärningsteknikernas effektivitet i att modellera komplexa termiska dynamiker. Vidare syftar studien till att förutsäga osäkerheten i den djupa inlärningsmodellen, vilket förstärker insikten
om modellens pålitlighet.
Avhandlingen introducerar även en innovativ modell baserad på konvolutionella neurala nätverk för att spåra bubblors dynamik i en minikanal,
uppvärmd från en sida. Modellen använder bilder från höghastighetskameror, trots utmaningar som skuggor, reflektioner och brus. Med hjälp av
överföringsinlärning kan modellen framgångsrikt detektera och klassificera bubblor med imponerande noggrannhet. Modellen ger en detaljerad
förståelse av bubblors rumsliga och temporala beteende, vilket kan vara
avgörande i industriella sammanhang.
En annan central del av denna avhandling är den kritiska värmeöverföringsflödet, en avgörande parameter i designen av industriella kokningssystem. En parametrisk Gaussian process regression-metod introduceras, som visat sig minska prediktiva fel jämfört med traditionella
metoder. Denna metod inte bara förstärker modellens robusthet utan ökar
Mälardalen University Press Dissertations
v
också transparensen, vilket understryker maskininlärningens potential att
bemöta empiriska korrelationers utmaningar.
Inom konvektiv värmeöverföring presenterar avhandlingen en strategi
för att hantera osäkra gränsförhållanden när det saknas exakta termiska
gränsbetingelser. Här används ett fysikinformerat neuralt nätverk som
kan representera både hastighets- och temperaturfält. Målet är att med
sparsam data återskapa både flödesfält och temperaturprofiler. För att
öka modellens tillförlitlighet och generaliserbarhet föreslås även en ensemblemetod för det fysikinformerade neurala nätverket. Denna metod
adresserar riskerna med överanpassning och ger en mätning av förtroendet för modellens utdata.
Sammanfattningsvis ger denna forskning nya möjligheter för innovativa lösningar inom värmeöverföringsstudier genom att använda maskininlärning. Resultaten understryker dessa metodologiers förmåga att hantera de komplexa utmaningarna och osäkerheterna i att förutsäga värmeöverföringsfenomen, vilket pekar på lovande framtida utforskningar och
industriella tillämpningar.
vi
Jerol Soibam
List of papers
Publications included in the thesis
This thesis is based on the following papers, which are referred to in the
text by their roman numerals:
I. Soibam, J., Aslanidou, I., Kyprianidis, K, and Bel Fdhila. R. (2020).
A Data-Driven Approach for the Prediction of Subcooled Boiling
Heat Transfer. In Proc. of the 61st International Conference of
Scandinavian Simulation Society, SIMS-36, September 22-24, 2020,
Oulu, Finland.
II. Soibam, J., Rabhi, A., Aslanidou, I., Kyprianidis, K, and Bel Fdhila. R.(2020). Derivation and Uncertainty Quantification of a
Data-Driven Subcooled Boiling Model. Multidisciplinary Digital
Publishing Institute (MDPI), Energies 13 (22), 5987, 2020.
III. Soibam, J., Rabhi, A., Aslanidou, I., Kyprianidis, K, and Bel Fdhila. R. (2021). Prediction of the Critical Heat Flux using Parametric Gaussian Process Regression. In proc. of the 15th International
Conference on Heat Transfer, Fluid Mechanics, and Thermodynamics, HEFAT, 26-28 July 2021, Amsterdam, Netherlands.
IV. Soibam, J., Aslanidou, I., Kyprianidis, K, and Bel Fdhila. R. (2023).
Inverse flow prediction using PINNs in an enclosure containing heat
sources. 8th Thermal and Fluids Engineering Conference , ASTFE,
27-30 March 2023, Maryland, USA.
V. Soibam, J., Scheiff, V., Aslanidou, I., Kyprianidis, K, and Bel Fdhila. R. (2023). Application of deep learning for segmentation of
bubble dynamics in subcooled boiling. International Journal of
Multiphase Flow, 2023.
VI. Soibam, J., Aslanidou, I., Kyprianidis, K, and Bel Fdhila. R. (2023).
Inverse flow prediction using ensemble PINNs and uncertainty quantification. International Journal of Heat and Mass Transfer (Under
Review), 2023.
Mälardalen University Press Dissertations
vii
The author’s contribution to the included publications
In all the appended papers, the author conceptualised, developed and performed the simulations, analysed the numerical results and wrote the drafts
and the final versions of the papers. The co-authors actively contributed
by enhancing the conceptual foundations and offered valuable insights to
improve the quality of the research.
Publications not included in the thesis
• E. Helmryd Grosfilley, G. Robertson, J. Soibam, and J.-M. Le Corre.
Investigation of Machine Learning Regression Techniques to Predict Critical Heat Flux over a Large Parameter Space. 20th International Topical Meeting on Nuclear Reactor Thermal Hydraulics
(NURETH-20), Washington, D.C., USA, Aug. 20 – 25, 2023.
• Aslanidou, I., Soibam, J. Comparison of machine learning approaches
for spectroscopy applications. In Proc. of the 63st International
Conference of Scandinavian Simulation Society, SIMS-63, Trondheim, Norway, September 20-21, 2022.
viii
Jerol Soibam
Contents
Acknowledgements . . . . . . . . . . . . . . . . . . . . . . . . .
i
Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . iii
Swedish summary . . . . . . . . . . . . . . . . . . . . . . . . . . v
List of papers . . . . . . . . . . . . . . . . . . . . . . . . . . . . . vii
List of figures . . . . . . . . . . . . . . . . . . . . . . . . . . . . xi
List of tables . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xiii
Nomenclature . . . . . . . . . . . . . . . . . . . . . . . . . . . . xiii
Preface . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xvii
1 I NTRODUCTION . . . . . . . . . . . . . . . . . . . . . . . . . . 1
1.1 Background . . . . . . . . . . . . . . . . . . . . . . . . . . . 1
1.2 Challenges and motivation . . . . . . . . . . . . . . . . . . . 4
1.3 Research approach . . . . . . . . . . . . . . . . . . . . . . . 6
1.4 Research framework . . . . . . . . . . . . . . . . . . . . . . 7
1.5 Summary of appended papers . . . . . . . . . . . . . . . . . 9
1.6 Contribution to knowledge . . . . . . . . . . . . . . . . . . . 12
1.7 Thesis outline . . . . . . . . . . . . . . . . . . . . . . . . . . 13
2 T HEORETICAL BACKGROUND . . . . . . . . . . . . . . . . . .
2.1 Overview of machine learning . . . . . . . . . . . . . . . . .
2.2 Application of ML in heat transfer and fluid dynamics . . . .
2.2.1 Data mining and processing . . . . . . . . . . . . . . . .
2.2.2 Control and optimisation . . . . . . . . . . . . . . . . . .
2.2.3 Flow modelling . . . . . . . . . . . . . . . . . . . . . . .
15
15
16
17
18
19
3 M ETHODOLOGY . . . . . . . . . . . . . . . . . . . . . . . . .
3.1 Framework of the thesis . . . . . . . . . . . . . . . . . . . .
3.2 Data availability . . . . . . . . . . . . . . . . . . . . . . . .
3.3 Architectures . . . . . . . . . . . . . . . . . . . . . . . . . .
3.3.1 Deep neural network . . . . . . . . . . . . . . . . . . . .
3.3.2 Physics informed neural network . . . . . . . . . . . . . .
3.3.3 Convolutional neural network . . . . . . . . . . . . . . .
3.3.4 Gaussian process regression . . . . . . . . . . . . . . . .
3.4 Uncertainty quantification for robustness . . . . . . . . . . .
23
23
24
26
26
28
29
31
32
Mälardalen University Press Dissertations
ix
3.4.1
3.4.2
Monte Carlo dropout . . . . . . . . . . . . . . . . . . . . 32
Deep ensemble . . . . . . . . . . . . . . . . . . . . . . . 33
4 R ESULTS AND DISCUSSION . . . . . . . . . . . . . . . . . . .
4.1 Case study on boiling heat transfer . . . . . . . . . . . . . . .
4.1.1 DNN for subcooled boiling . . . . . . . . . . . . . . . . .
4.1.2 Computer vision for subcooled boiling . . . . . . . . . .
4.1.3 Critical heat flux . . . . . . . . . . . . . . . . . . . . . .
4.2 Case study on convective heat transfer . . . . . . . . . . . . .
4.2.1 Forced convection . . . . . . . . . . . . . . . . . . . . .
4.2.2 Mixed convection . . . . . . . . . . . . . . . . . . . . . .
4.3 Discussion of the research questions . . . . . . . . . . . . . .
5
35
35
35
38
41
43
43
46
50
C ONCLUSIONS AND FUTURE WORK . . . . . . . . . . . . . . . 57
B IBLIOGRAPHY . . . . . . . . . . . . . . . . . . . . . . . . . . . . 61
PAPERS . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 71
x
Jerol Soibam
List of figures
1.1
1.2
1.3
1.4
2.1
2.2
3.1
3.2
3.3
3.4
3.5
3.6
3.7
3.8
3.9
4.1
4.2
4.3
4.4
Brief overview of uses of machine learning for heat transfer
and fluid dynamics. . . . . . . . . . . . . . . . . . . . . . . . 3
Research approach adopted in this thesis. . . . . . . . . . . . 7
Holistic overview of the research framework employed in the
thesis. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8
Schematic diagram of the relationship between the research
questions and the appended papers. . . . . . . . . . . . . . . . 12
Research trend of using machine learning for heat transfer and
fluid dynamics (Source: scopus & (Ardabili et al., 2023)) . . . 15
Potential applications of ML in the domain of heat transfer
and fluid mechanics. . . . . . . . . . . . . . . . . . . . . . . 16
Detailed framework of the learning system. . . . . . . . . . .
Numerical computational domain and experimental setup. . .
Numerical computational domain for forced and mixed convection. . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
Deep neural network architecture. . . . . . . . . . . . . . . .
PINN framework for forced and mixed convection. . . . . . .
Convolutional neural network for bubble segmentation and
tracking. . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
Framework for parametric Gaussian process regression. . . . .
Monte Carlo dropout architecture with a dropout ratio of 20%
for training and testing. . . . . . . . . . . . . . . . . . . . . .
Deep ensemble architecture for M = 2. . . . . . . . . . . . .
Temperature field prediction and the relative error between
the CFD and DE model of the minichannel. . . . . . . . . . .
Void Fraction field prediction and the relative error between
the CFD and DE model of the minichannel. . . . . . . . . . .
Wall void fraction profile by the DNN, MC dropout and DE
models along the arc of the minichannel . . . . . . . . . . . .
Comparison of mask between ground truth (human eyes), classical image processing and CNN model. a) Ground truth mask
b) Classical mask c) CNN mask d) Normal image with a surface void fraction of 6.45% e) IoU for classical method and f)
IoU for CNN model . . . . . . . . . . . . . . . . . . . . . .
Mälardalen University Press Dissertations
23
25
26
27
29
30
31
33
34
36
37
37
38
xi
4.5
Masks given by the model with different noise conditions: a)
Original image with the mask, b) Contrast enhancement, c)
Sharpening, d) Top-hat filter with disk shape radius = 20, e)
Gaussian Blur σ = 0.8, f) Gaussian noise σ = 10, g) Gaussian
noise σ = 25, h) Black dead pixel 0.02% and i) IoU for noisy
images . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
4.6 Results extracted from the binary mask given by CNN model
to study local bubbles behaviour: a) Trajectory of the bubbles
with velocity magnitude and growing bubble images (Note:
axes are not in scale), b) Bubble diameter with time, c) Bubble Reynolds number with time. . . . . . . . . . . . . . . . .
4.7 Validation result of the CHF for pGPR and LUT predicted
values . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
4.8 Weighting factor and prediction obtained from pGPR model .
4.9 Case studies for different sensor placements to infer temperature distribution in the enclosure domain . . . . . . . . . . . .
4.10 Case studies for different sensor placement to infer temperature distribution in the enclosure domain . . . . . . . . . . . .
4.11 Temperature distribution: Transfer learning case study for two
rectangular sources . . . . . . . . . . . . . . . . . . . . . . .
4.12 Comparision of L2 relative error between QR and randomly
selected sensors over time for the component u∗ and θ ∗ . . . .
4.13 Sensitivity of QR sensors and uncertainty quantification of
PINN . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
4.14 Sensitivity of QR sensors and uncertainty quantification of
PINN . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
4.15 Comparison of flow field profiles for constant heat flux at t =
80, as predicted by PINN and DNS, with corresponding L2
error for u∗ , v∗ , p∗ , θ ∗ . . . . . . . . . . . . . . . . . . . . . .
4.16 Predictive performance of the PINN at random point which
was not a part of QR sensor nor training . . . . . . . . . . . .
xii
39
40
42
42
44
45
45
46
47
48
49
50
Jerol Soibam
List of tables
1.1
Overview of appended papers presented in this thesis. Sub.
boiling: Subcooled boiling, CHF: Critical Heat Flux & Conv.:
Convection . . . . . . . . . . . . . . . . . . . . . . . . . . . 11
3.1
Data split for training, validation, and testing at different heat
flux and flow conditions. . . . . . . . . . . . . . . . . . . . . 24
4.1
Performance of the DNN, MC dropout, and DE models on
validation dataset of the computational domain, VF: Void Fraction, Temp: Temperature. . . . . . . . . . . . . . . . . . . . . 35
Mälardalen University Press Dissertations
xiii
Nomenclature
Abbreviations
Adam
Adaptive Moment Estimation
ANN
Artificial Neural Network
CFD
Computational Fluid Dynamics
CNN
Convolutional Neural Network
DE
Deep Ensemble
DL
Deep Learning
DMD
Dynamic Mode Decomposition
DNN
Deep Neural Network
DNS
Direct Numerical Simulation
DRL
Deep Reinforcement Learning
GA
Genetic Algorithms
GPR
Gaussian Process Regression
GPU
Graphical Processing Unit
HPC
High Performance Commutating
LES
Large Eddy Simulation
MC
Monte Carlo
MCMC
Markov chain Monte Carlo
ML
Machine Learning
MSE
Mean Square Error
NLL
Negative Log-Likelihood
NN
Neural Network
PCA
Principal Component Analysis
PDE
Partial Differential Equation
pGPR
parametric Gaussian Process Regression
xiv
Jerol Soibam
PINN
Physics Informed Neural Network
PIV
Particle Image Velocimetry
POD
Proper Orthogonal Decomposition
QoIs
Quantities of Interest
RANS
Reynolds Average Navier-Stokes
ReLU
Rectified Linear Units
RMSEP
Root Mean Square Error Prediction
ROI
Region of Interest
ROM
Reduced Order Model
RQ
Research Question
SA
Spallart–Allmaras
UQ
Uncertainty Quantification
VF
Void fraction
YOLO
You Only Look Once
Machine learning symbols
ŷ
Predicted value
λ
Hyperparameter
λw
Weight decay
E(y∗ )
Predictive mean
L
Loss function
τ
Gaussian process precision
b
Biases
g(x)
Activation function
l
Length scale
Lr
Learning rate
M
Number of networks
N
Size of the dataset
p
Probability of neurons not dropped
R2
Coefficient of determination
Mälardalen University Press Dissertations
xv
T
Repetition of the prediction step
Var(y∗ )
Predictive variance
W
Weights
x
Input signals
X∗
Input features
x∗
New dataset
y
Target value
L1
Lasso regression
L2
Ridge regression
Physics symbols
α
Void fraction
μ
Dynamic viscosity
θ∗
Non-dimensional temperature
l
Non dimensional arc length
p
Pressure
p∗
Non-dimensional pressure
Pec
Péclet number
q
Heat flux
Re
Reynolds number
Ri
Richardson number
T
Fluid temperature
t∗
Non-dimensional time
Twall
Wall temperature
u
Velocity in x direction
u∗
Non-dimensional velocity in x direction
v
Velocity in y direction
v∗
Non-dimensional velocity in p direction
x
Minichannel length in x axis
y
Minichannel length in y axis
xvi
Jerol Soibam
Preface
This doctoral thesis was completed at the Future Energy Center (FEC)
at the School of Business, Society and Engineering (EST) at Mälardalen
University in partial fulfilment of the requirements for acquiring the degree of Doctor of Philosophy in Energy and Environmental Engineering.
The study was carried out under the guidance of Prof. Rebei Bel Fdhila, Prof. Konstantinos Kyprianidis and Assoc Prof. Ioanna Aslanidou.
The thesis summarises the study conducted by the author during the PhD
project, which started on 18 June 2019. It also provides a concise overview
of the six enclosed publications, while the comprehensive research findings are available in the appendix.
Mälardalen University Press Dissertations
xvii
1
Introduction
This chapter introduces the research background, challenges, and motivations. It formulates research questions from literature gaps. Thereafter, it
presents the research methodology and framework, detailing the links between research topics and appended papers. The chapter concludes with
the thesis outline.
1.1
Background
Heat transfer is a fundamental process found in numerous sectors, ranging
from the cooling of electronic devices, supercomputers, and the functioning of nuclear reactors, to the design of industrial processes and equipment. It is characterised by a multitude of complex phenomena that span
a broad range of scales, including molecular, microscale, and macroscale.
The very essence of heat transfer plays a crucial role in the broader context of energy production, management, and consumption. Every device,
mechanism, or process that involves the production or consumption of
energy invariably incorporates elements of heat transfer. This includes
traditional forms of energy production such as combustion engines and
thermal power plants, as well as modern, renewable energy technologies
such as solar thermal and geothermal energy systems.
The ability to monitor and control the heat transfer process in a system
is critical for optimal performance, longevity, safety, and cost-effectiveness
in a wide range of applications. Moreover, as the world becomes increasingly energy-conscious, efficient management of heat has profound implications for energy conservation and sustainability. Improved heat transfer
systems can lead to significant energy savings, reduce environmental impact, and help meet global energy demand in a sustainable manner. For
instance, enhancing heat transfer efficiency in power generation and industrial processes could greatly reduce energy consumption and carbon
footprint.
Over the years, significant advancements have been made in understanding heat transfer principles, largely due to concerted efforts in experimental research and numerical modelling. These areas of research
have offered unique contributions that have synergistically expanded the
knowledge base. The experimental and modelling methods have generated a large volume of data over the years, enriching the scientific comMälardalen University Press Dissertations
1
Machine Learning Techniques for Enhanced Heat Transfer Modelling
munity with invaluable resources for future research and validation. However, despite these substantial advancements, modelling detailed physical problems for industrial applications remains a challenging endeavour
(Karniadakis et al., 2021). Accurately depicting heat transfer problems
often necessitates substantial computational time, a fine-tuned mesh grid
and well-defined boundary conditions (Alifanov, 2012). Moreover, setting up an accurate numerical simulation is a complex, time-consuming
process that depends on multiple factors such as defining geometry, generating mesh, selecting solvers, post-processing, and interpreting results
in a meaningful way (Molland and Turnock, 2007). Furthermore, when
these methods are subject to ill-posed problem conditions or high levels
of uncertainty, the validity of the model becomes questionable (Colaco
et al., 2006). The sensitivity of the solution can vary drastically with even
a slight change in the input parameters or the initial conditions, potentially
rendering the results less reliable or unrepeatable.
Despite these hurdles, a distinct opportunity arises from the immense
volume of data accrued from years of experimental research and numerical
modelling. This data accumulation provides an ideal basis for the application of emerging analytical tools. Specifically, the introduction of machine learning (ML) into the field of heat transfer and fluid mechanics is a
notable development (Brunton et al., 2020), reminiscent of the paradigm
shift brought about by the advent of high-performance computing (HPC)
decades ago. Just as HPC expanded the boundaries of research capability,
enabling larger and more intricate problem-solving, ML has the potential
to make a comparable, if not greater, impact. The application of ML in this
context can be used to process and learn from vast datasets, and presents
a potent approach for addressing complex and often ill-posed problems
in heat transfer and fluid mechanics (Taira et al., 2017, 2020). Moreover,
in scenarios where data is limited, machine learning holds another distinct advantage – the ability to integrate physics-based principles into the
models. This unique fusion of data-driven learning with physics-informed
insights allows for the creation of robust and reliable models even when
data is scarce (Parish and Duraisamy, 2016). By embedding the foundational principles of physics into the learning process, these models not
only adapt to data but also remain consistent with the underlying physical
laws. This approach enhances the robustness of the model, thereby further
extending the utility and impact of ML in the study of heat transfer and
fluid mechanics (Raissi et al., 2019b).
Within the domain of heat transfer and fluid mechanics, recent years
have witnessed an escalating interest in the potential of ML methods to
augment engineering solutions as shown in Figure 1.1. Utilising ML or
data-driven techniques within fluid mechanics isn’t a recent development.
2
Jerol Soibam
Chapter 1. Introduction
Figure 1.1: Brief overview of uses of machine learning for heat transfer and
fluid dynamics.
Indeed, these methods were explored in previous decades (Pollard et al.,
2016; Brunton et al., 2020). However, compared to the modern ML techniques, earlier studies were not as impactful. This transformation has been
driven by the current intersection of big data, high-performance computing, sophisticated algorithms, and significant industrial investment, leading to open-source tools and accessible benchmark datasets. An influential development in the contemporary ML age is the emergence of deep
learning (DL) (Goodfellow et al., 2016), achieving human-level performance in several demanding tasks, such as image recognition (Krizhevsky
et al., 2012) and control systems (Mnih et al., 2015; Silver et al., 2018).
The success of these data-driven techniques has primarily relied on extensive datasets (Deng et al., 2009), open-source algorithms promoting reproducible research, and powerful computational resources like graphical
processing unit (GPU) hardware.
In recent decades, substantial progress in experimental methods and
large-scale numerical simulations have made big data in fluid mechanics and heat transfer a reality (Pollard et al., 2016). Despite this, the
field of fluid mechanics remains primarily focused on the physical mechanisms underlying fluid phenomena. ML frameworks have proven crucial
in instances where these mechanisms are challenging to investigate. A
prominent example is the use of Proper Orthogonal Decomposition (POD)
(Rowley and Dawson, 2017) to examine near-wall flow structures at varying Reynolds numbers in turbulent channel flow (Podvin et al., 2010). The
Mälardalen University Press Dissertations
3
Machine Learning Techniques for Enhanced Heat Transfer Modelling
past decade has seen impressive advances in ML, spurred by commercial
successes in natural language processing and advertising sectors (Otter
et al., 2021). With the confluence of these technologies, the techniques
become relevant for heat transfer and engineering challenges, enabling
the engineering field to capitalise on these potent methods. A promising
aspect of the area of heat transfer and fluid dynamics is that these burgeoning techniques are robust enough to handle large datasets and depict
intricate nonlinear dynamics frequently encountered in these fields (Brunton et al., 2020).
1.2
Challenges and motivation
The rapid advancement of ML or DL techniques has the potential to significantly improve the solutions to engineering problems, particularly in
the areas of heat transfer and fluid mechanics. However, these domains
require extensive research before the full potential of ML/DL can be exploited. Precise quantification of underlying physical mechanisms in heat
transfer and fluid dynamics is crucial for their proper analysis. These
fields exhibit complex, multiscale phenomena, presenting particular challenges. For instance, turbulence flow fields introduce significant complexities due to their nonlinearities and multiple spatiotemporal scales, which
may not be readily addressed by currently available ML algorithms. In
addition, another notable challenge is to ensure the alignment of ML models with the conservation laws of physics. Traditional ML models might
produce physically inconsistent results, leading to inaccurate and untrustworthy predictions. To solve this problem, researchers are exploring techniques that incorporate physical laws and principles into the training process of ML models.
ML models fundamentally rely on data-driven mechanisms to formulate informed predictions or decisions. The efficacy of these ML models
is inherently contingent upon the quality and volume of the available data
(Iten et al., 2020). In situations where data is either limited or fails to
represent a wide spectrum of instances, the performance of ML models
may be critically compromised. The scarcity or lack of heterogeneous
data could lead to an inadequate learning experience for the models, potentially jeopardising their predictive accuracy and decision-making capabilities (Brunton et al., 2016). Moreover, it’s pivotal to acknowledge
situations in which data collection may not be practical or even possible (Schmidt and Lipson, 2009). Such situations introduce an extra layer
of complexity to the efficient training of ML models. Restrictions could
originate from a variety of sources, including logistical difficulties, safety
4
Jerol Soibam
Chapter 1. Introduction
protocols, privacy regulations, or the lack of requisite technological methods or tools for data collection.
A significant obstacle presented by these machine ML models is the
inherent lack of transparency, or interpretability (Lipton, 2018). Unlike
traditional numerical simulations, which often follow a clear and understandable set of steps, ML algorithms frequently function as a sort of
’black box’. In other words, while these models are capable of generating
predictions, the underlying logic or reasoning that drives these predictions
is often elusive or even hidden. This characteristic of ML models creates a
significant barrier, particularly in disciplines such as heat transfer and fluid
dynamics. In these fields, the comprehension of the underlying physical
phenomena is not just desirable but often essential. The ’black box’ nature of ML models makes it challenging to understand these phenomena
fully and therefore can limit their applicability in these areas. Therefore,
it becomes paramount to test the validity and robustness of these models
before they can be integrated into engineering systems. It is not enough
for a model to just make accurate predictions; it must also be reliable
and resistant to potential changes in the input data or underlying conditions (Zhang et al., 2019). Moreover, even a model that exhibits strong
predictive performance may not always meet the requirements of a given
situation. In such cases, quantifying the uncertainty associated with the
model’s predictions becomes necessary. Uncertainty quantification allows
us to understand the potential range and likelihood of different outcomes,
providing a fuller picture of the model’s capabilities and limitations (Yang
et al., 2020; Zhu et al., 2019). This, in turn, can inform decision-making
and risk management, adding a layer of depth and robustness to the use of
ML models in engineering and other fields.
Despite the challenges posed by the ’black box’ nature of ML models, these difficulties should not be perceived as insurmountable barriers.
Rather, they can serve as powerful catalysts, spurring the development
and refinement of more effective and efficient ML techniques. Indeed,
applying ML to fields with known physical principles, such as fluid dynamics and heat transfer, has the potential to solve complex problems in
these areas while also contributing to a deeper understanding of the underlying mechanics of ML algorithms (Brunton et al., 2020). The interplay
between ML and these traditionally physics-based fields can cultivate a
mutually beneficial relationship. For instance, the use of known physical
laws can be a valuable resource during the training of neural networks
(Karniadakis et al., 2021). These laws can provide crucial guidance to the
learning process of the model, ensuring that the networks’ predictions do
not defy the physical principles they are supposed to model. This integration of physical laws with ML may not only improve the model’s preMälardalen University Press Dissertations
5
Machine Learning Techniques for Enhanced Heat Transfer Modelling
dictive accuracy, but also its interpretability, since the predictions would
align with established physical understanding. Therefore, despite its challenges, ML represents a powerful tool with a unique capacity to handle
high dimensional, nonlinear problems common in fields like heat transfer.
The journey towards overcoming its current limitations can also open up
opportunities for enhancing the capabilities of ML models, thereby advancing the broader fields of artificial intelligence and data science.
Acknowledging the challenges and opportunities outlined, it is apparent that the utilisation of ML/DL techniques in the field of heat transfer
and fluid mechanics necessitates comprehensive investigation. The intricate complexities which characterise these fields demand a precise exploration into the capabilities and limitations of such techniques. A more
detail on the literature and gaps can be found in Chapter 2. The issues surrounding data availability and quality, alignment with physical laws, and
model interpretability play pivotal roles in shaping the efficacy of ML/DL
methods. The current scenario thus raises several research questions that
require attention to further push the boundaries of ML applications in heat
transfer. Consequently, this thesis focuses on the following research questions (RQ):
• RQ1: How effectively can deep learning techniques capture the
intricate nature of heat transfer based on data, and how successfully
can these techniques extrapolate to unseen scenarios?
• RQ2: How to incorporate known physics of heat transfer in deep
learning models to improve the generalisability in estimating unknown parameters and system behaviours?
• RQ3: How can machine learning methods secure reliable prediction outcomes in situations where heat transfer data are scarce?
• RQ4: How can uncertainty quantification contribute to enhancing
the robustness and reliability of machine learning models utilised to
simulate and predict heat transfer?
The formulated research questions are expressed in a general manner, however, the conclusions drawn are based on selected case studies
within the research framework presented in Section 1.4. The above research questions are further discussed in detail in Section 4.3.
1.3
Research approach
The research approach followed in this thesis is an iterative cycle as depicted in Figure 1.2. The process starts with the "Research Objectives,"
6
Jerol Soibam
Chapter 1. Introduction
Figure 1.2: Research approach adopted in this thesis.
which provide the foundation for the entire thesis. These objectives define
the goals and aim of the research, offering a clear direction for the investigation. Upon setting clear and achievable research objectives, the next
step is to formulate preliminary "Research Questions". These inquiries
are derived from the research objectives, focusing on and specifying the
areas of investigation. Once the initial research questions are established,
a comprehensive "Literature Review" is conducted. This step involves a
systematic analysis and interpretation of the existing body of knowledge
related to the research topic. The purpose of this review is to identify
gaps in current knowledge, understand the context of the research, and
further refine the research questions and objectives. Following the literature review, the "Research Framework" is developed. This serves as the
blueprint for the research, outlining the theoretical foundations, hypotheses or assumptions, and methodologies for data collection and analysis.
The framework provides a roadmap for how the research questions will
be answered. The last phase in this research cycle, called as "Data Collection," is the pivotal cornerstone that supports the entire research. Here,
data on heat transfer is gathered from a variety of sources, including experiments, numerical models, sensors, and existing literature. This data
then acts as the catalyst for refining the research framework. It guides
the selection of the most fitting ML algorithm and prompts the adjustment
of the other stages in the research process. The process is iterative, with
each stage revisited and refined based on the data until the final research
questions are precisely articulated. This approach ensures a thorough and
adaptable exploration of the research topic.
1.4
Research framework
The proposed research framework presents a comprehensive and adaptive strategy for ML model selection, development, and validation in the
Mälardalen University Press Dissertations
7
Machine Learning Techniques for Enhanced Heat Transfer Modelling
specific context of heat transfer data. Recognising the variability in data
availability across different tasks and domains, the framework provides a
robust plan for these variations, thereby optimising the performance and
reliability of ML models in heat transfer applications.
In scenarios where there is an abundance of heat transfer data, the
framework uses powerful architectures such as Deep Neural Networks
(DNNs) and Computer Vision models. These architectures, renowned for
their ability to model complex patterns and relationships, typically require
large datasets to achieve optimal performance. However, in many practical settings, large datasets may not be readily available. To navigate this
challenge, the framework introduces the concept of transfer learning. This
technique enables models, initially trained on extensive datasets, to be
fine-tuned for specific tasks using a smaller amount of heat transfer data,
thus capitalising on the generalised knowledge extracted from the primary
dataset.
On the contrary, when dealing with scarce heat transfer data, the research framework employs methods that can operate efficiently with smaller datasets. These include Physics-Informed Neural Networks (PINNs)
and Gaussian Processes. By incorporating domain knowledge related to
heat transfer or exploiting the inherent structure of the data, these methods
can yield substantial results even with limited data. This strategy becomes
particularly beneficial in scenarios where the collection of heat transfer
data is costly, time-consuming, or otherwise challenging.
To enhance the robustness and transparency of the developed models,
the research framework integrates the principle of Uncertainty Quantification (UQ). UQ provides a measure of the model’s confidence in its predictions, thus increasing the model’s validity and interpretability. This component is of significant importance in contemporary heat transfer appli-
Figure 1.3: Holistic overview of the research framework employed in the thesis.
8
Jerol Soibam
Chapter 1. Introduction
cations, where understanding the uncertainty associated with predictions
can critically aid in decision-making processes.
In conclusion, the proposed research framework provides an adaptable
and resilient methodology for ML model selection and validation, specifically tailored for the varying conditions of heat transfer data availability.
This framework serves as the backbone of the investigation, guiding the
development and conclusions of the associated publications in addressing
the research questions.
1.5
Summary of appended papers
Paper I: A Data-Driven Approach for the Prediction of Subcooled Boiling Heat Transfer
This study explores a data-driven technique using a supervised deep
neural network (DNN) to understand boiling heat transfer in subcooled
boiling flow within a 3 x 400 mm minichannel. The model predicts wall
temperature and void fraction, using data from numerical simulations split
into training, validation, and testing datasets. The DNN model, validated
and evaluated through these datasets, effectively predicts these quantities
with high accuracy for both interpolation and extrapolation datasets. The
results demonstrate the model’s ability to reproduce the subcooled boiling
pattern and its satisfactory generalisation property.
Paper II: Derivation and Uncertainty Quantification of a Data-Driven
Subcooled Boiling Model
The paper explores two deep learning techniques for subcooled boiling heat transfer. The first predicts deterministic quantities of interest
(QoIs), while the second predicts uncertainties in the model when estimating QoIs, utilising methods like Monte Carlo dropout and Deep Ensemble.
The results demonstrated that these uncertainty quantification (UQ) models outperformed the deterministic model. Furthermore, they exhibited
strong performance on unseen and extreme extrapolation datasets. They
also accurately reproduced physics even at a heat flux beyond training parameters. On average, all models had a root mean square error percentage
(RMSEP) under 5% for wall temperature and under 2% for a void fraction
with R2 scores of 0.998 and 0.999 respectively.
Paper III: Prediction of the Critical Heat Flux using Parametric Gaussian Process Regression
Understanding the critical heat flux is crucial for industrial boiling system design and safety, but most existing studies rely on empirical knowledge, leading to a ± 30% predictive error. This study addresses this gap by
Mälardalen University Press Dissertations
9
Machine Learning Techniques for Enhanced Heat Transfer Modelling
using machine learning to uncover hidden features in experimental data,
enhancing accuracy and efficiency in data usage. Experimental data from
low-pressure, low-flow boiling flows in tubes, a parametric Gaussian process regression model is utilised to predict the critical heat flux. This
model’s predictions are compared to experimental measurements and values from a critical heat flux look-up table, resulting in improved prediction accuracy and insight into input parameter relevance. Besides offering
prediction uncertainty, the model aligns with the underlying physics and
demonstrates robustness, which entails potential extension to other geometries, datasets, and operating conditions.
Paper IV: Inverse flow prediction using PINNs in an enclosure containing heat sources
This study focuses on addressing an ill-posed problem in simulating
heat transfer due to the lack of precise thermal boundary conditions. To
overcome this issue, a physics-informed neural network is used, with a
few known temperature measurements in the domain. The network helps
simulate the velocity and temperature fields while enforcing the NavierStokes and energy equations. The study serves as an inverse problem,
aiming to recreate the global flow field and temperature profile with limited data points. It also explores transfer learning for various parameters
like the position and size of the heat source within the enclosure domain,
aiding in the effective design of thermal systems. The results indicate
good agreement between the proposed method and the physics reflected
by the numerical outcomes.
Paper V: Application of deep learning for segmentation of bubble dynamics in subcooled boiling
In this research, a convolutional neural network model is implemented
to monitor bubble dynamics in a heated vertical rectangular mini-channel.
The model utilises images captured by a high-speed camera, overcoming
noise challenges such as shadows, reflections, and chaotic bubbles. The
model is trained using transfer learning, which reduces the need for a large
dataset and computational resources. The validated model demonstrates
98% accuracy in bubble detection and is robust under various conditions.
Moreover, the model accurately identifies bubble edges and predicts bubble masks with an 85% average intersection over union, providing a detailed understanding of individual bubbles, including their coalescence,
oscillation, and collisions. This facilitates the estimation of local parameters and a comprehensive grasp of their spatial-temporal behaviour.
10
Jerol Soibam
Chapter 1. Introduction
Paper VI: Inverse flow prediction using ensemble PINNs and uncertainty quantification
Lack of accurate thermal boundary conditions in numerical simulations often leads to challenges in heat transfer problems. This study employs a physics-informed neural network to handle these ill-posed problems with limited sensor data. The network, complying with the NavierStokes and energy equations, reconstructs the flow field around a square
cylinder, identifying unknown thermal boundaries. Optimal sensor placement is achieved through the QR pivoting technique, improving model
accuracy. An ensemble physics-informed neural network is used to enhance robustness, generalisability, and to provide a measure of model uncertainty, hence improving applicability in handling complex heat transfer
problems with unknown boundaries.
Overview
The presented research papers cover a variety of case studies within the
areas of heat transfer. In addition, they utilise different types of data and
architectures. Therefore, this information is presented in the Table 1.1, to
improve the clarity of the each research paper’s focus. Finally, the link
between the publications and the RQs is shown in Figure 1.4.
Table 1.1: Overview of appended papers presented in this thesis. Sub. boiling:
Subcooled boiling, CHF: Critical Heat Flux & Conv.: Convection
Paper
Case study
Data type
Architecture
UQ
I
Sub. boiling
Numerical
(Rabhi et al., 2021)
DNN
No
II
Sub. boiling
Numerical
(Rabhi et al., 2021)
MC Dropout
Deep Ensemble
Yes
III
CHF
Experimental
(Kim et al., 2000)
pGPR
Yes
IV
Forced conv.
Numerical
Inverse PINN
Transfer learning
No
V
Sub. boiling
Experimental (Image)
(Scheiff et al., 2023)
CNN
Transfer learning
No
VI
Mixed Conv.
Numerical
(Fraigneau, Y., 2019)
Ensemble PINN
QR-pivoting
Yes
Mälardalen University Press Dissertations
11
Machine Learning Techniques for Enhanced Heat Transfer Modelling
Figure 1.4: Schematic diagram of the relationship between the research
questions and the appended papers.
1.6
Contribution to knowledge
• Designed and implemented a deep neural network (DNN) architecture with the capability to address complex physical phenomena,
such as subcooled boiling. By successfully implementing a DNN
for heat transfer problems, the application of artificial neural networks to challenging scientific problems is expanded.
• Introduced an innovative computer vision technique for automated
bubble detection and estimation of statistical aspects of bubble dynamics. This technique provides a significant enabler to analyse and
understand fluid behaviour by automatically detecting and analysing
bubbles instead of the manual time-consuming image processing
and tracking.
• Proposed a physics-informed neural networks (PINN) while employing a sensor selection method (QR pivoting) to unveil unknown
parameters and reconstruct the flow field. This proof-of-concept
methodology enhances the efficiency and accuracy of the model,
especially in scenarios where sensor data plays a crucial role in
comprehending complex physical processes.
• Implemented a reliable model that accounts for uncertainty (Deep
Ensemble and Monte Carlo dropout) in both deep neural network
(DNN) and physics-informed neural network (PINN) architectures.
These models enhance the trustworthiness of predictions and simulations by recognizing and quantifying uncertainty.
12
Jerol Soibam
Chapter 1. Introduction
1.7
Thesis outline
The thesis is based on the appended papers and contains the following
chapters:
Chapter 1: Introduction
This chapter provides the research background, research challenges,
research questions formulated based on the knowledge gap, and research framework. It also shows the link to the appended papers and
the research questions in this thesis.
Chapter 2: Theoretical background
In this chapter a comprehensive literature review based on the use of
ML applications for heat transfer and fluid mechanics is presented.
The chapter also highlights the type of ML algorithms that can be
used for engineering problems.
Chapter 3: Methodology
This chapter presents a detailed ML framework used in this thesis.
Furthermore, it describes the data and architectures employed in this
thesis.
Chapter 4: Results and discussion
This chapter presents the key results obtained in this study, and provides a detailed discussion of the findings based on the research questions and limitations. The results are divided into boiling heat transfer and convective heat transfer.
Chapter 5: Conclusions and future work
This chapter presents the major conclusions of the thesis and key
contributions to the knowledge. The thesis ends by discussing potential future studies.
Mälardalen University Press Dissertations
13
2
Theoretical Background
In this chapter, a comprehensive literature review based on applications
of machine learning in the field of fluid mechanics and heat transfer is
provided.
2.1
Overview of machine learning
In recent years, ML, with a particular emphasis on deep learning, has
resulted in significant progress in tasks related to classification and regression. These methods are recognised for their inherent non-linearity
and adaptability across a broad range of applications. This versatility has
led to their use in a variety of fields, enabling the development of innovative solutions based on data in areas such as natural language processing,
computer vision, and robotics. More recently, these methodologies have
begun to leave a remarkable imprint on the natural sciences.
A trend that reflects this progress is the increasing application of ML
in the fields of heat transfer and fluid dynamics, as illustrated in Figure
2.1. The Figure reveals a sharp increase in research activity in this area
starting around 2015. This increasing surge of interest and research is
fundamentally driven by two intertwined factors. Firstly, the progressive
Figure 2.1: Research trend of using machine learning for heat transfer and fluid
dynamics (Source: scopus & (Ardabili et al., 2023))
Mälardalen University Press Dissertations
15
Machine Learning Techniques for Enhanced Heat Transfer Modelling
maturation and sophistication of ML algorithms have made them more accessible, effective, and applicable to a broader range of complex problems.
This is particularly noticeable in technical fields such as heat transfer and
fluid dynamics, where the intricate systems and phenomena involved can
greatly benefit from the predictive modelling and pattern recognition capabilities offered by ML techniques.
Secondly, the present era is characterised by unprecedented access to
extensive computational resources, thanks to advances in technology and
computing power. The increasing availability of these resources has significantly lowered the barriers to implementing ML techniques, thus encouraging their widespread adoption across multiple scientific fields. In
parallel with these developments, the explosion of big data across numerous sectors has provided a wealth of training data for these models. The
availability of these vast datasets accelerates the learning process of ML
algorithms and further propels the pace of advancements in these fields.
2.2
Application of ML in heat transfer and fluid dynamics
There has been a substantial historical intertwining of fluid mechanics
and machine learning. Numerous prevalent techniques utilised in today’s
ML landscape were initially explored and introduced within the context
of fluid dynamics many years ago (Brunton et al., 2020) (Pollard et al.,
2016). However, these early methods didn’t gain prominence until the recent convergence of vast data, high-power computing, intricate algorithms
(like deep learning), and significant industry investment. The use of ML
in heat transfer and fluid dynamics domains can be broadly classified as
shown in Figure 2.2.
Figure 2.2: Potential applications of ML in the domain of heat transfer and fluid
mechanics.
16
Jerol Soibam
Chapter 2. Theoretical Background
2.2.1
Data mining and processing
ML algorithms utilise optimisation methods to process data. Their objective is to discover a low-rank subspace for ideal data embedding, useful
for tasks like regression, clustering, and classification. In essence, ML is
a set of mathematical techniques for data mining. Principal Component
Analysis (PCA) is one of the prominent ML algorithms that elucidates
correlations in high-dimensional data. Lumley (1970) employed PCA to
model turbulent flows, suggesting that the instantaneous flow fields can be
depicted by a linear weighted sum of orthogonal basis vectors. A reduced
expression of the flow field can be derived from higher PCA modes. However, as PCA treats each instantaneous field independently, it’s difficult to
understand the temporal information of the flow field solely through PCA
mode. In recent developments, a convolutional neural network (CNN)
auto-encoder was used as a non-linear decomposition method for the flow
field around a cylinder (Murata et al., 2020), demonstrating that a nonlinear activation function reduces the reconstruction error compared to
PCA. Neural networks (NN) were used to investigate the classification of
wake topology behind a pitching airfoil for local vorticity measurements
(Colvert et al., 2018). The authors stated that their NN model was capable
of extracting the features from the wakes and mapping a time series of the
local vorticity.
CNNs can facilitate automated image analysis for object detection,
segmentation, and classification (Redmon et al., 2016; He et al., 2017;
Geng and Wang, 2020). CNNs effectively extract spatial information from
images, resembling human visual perception and proving crucial in scientific image analysis (Albawi et al., 2017). In studying boiling phenomena,
CNNs analyse bubble images to garner insights, owing to the rich statistical data (Fu and Liu, 2019). CNNs have detected the shift from nucleate to
film boiling in pool boiling experiments (Hobold and da Silva, 2019a) and
measured heat flux from boiling images (Hobold and da Silva, 2019b).
This correlation between bubble shapes and heat flux values facilitated effective tracking of the boiling process. CNNs also identified critical heat
flux (CHF) in pool boiling experiments by classifying based on spatial information (Rassoulinejad-Mousavi et al., 2021). Additionally, researchers
have utilised CNNs to predict boiling regimes, proving useful in averting
thermal accidents (Sinha et al., 2021). Finally, a CNN model has linked
bubble dynamics and boiling heat transfer, predicting boiling curves with
a mean error of 6% (Suh et al., 2021). YOLO (You Only Look Once), a
single-stage object detection CNN, has been employed to identify bubbles
during collision, merging, and rupture events (Wang et al., 2021). Twostage CNNs have been previously used to accurately identify and classify bubbles in a two-phase flow circuit for a nuclear power plant cooling
Mälardalen University Press Dissertations
17
Machine Learning Techniques for Enhanced Heat Transfer Modelling
system (Serra et al., 2020). Faster-Region-CNN, another CNN, detects
bubbles and estimates their shape by elliptical fitting, effective at reconstructing overlapping bubbles under a void fraction of 2% (Haas et al.,
2020). Mask R-CNN has also been used to segment and create masks
from an upward bubbly flow, using a mix of experimental and synthetic
datasets (Kim and Park, 2021; Fu and Liu, 2019).
2.2.2
Control and optimisation
Flow control can significantly enhance system performance, making it an
attractive prospect for technological advancements. Feedback flow control
modifies the behaviour of systems through sensor-informed actuation. It
stabilises unstable systems, reduces sensor noise, and compensates for external disturbances and uncertainty. However, challenges arise from fluid
non-linearity, high-dimensional data, and time delays. Many researchers
utilise ML techniques to overcome these issues in system identification,
sensor placements, and controls.
Artificial neural networks (ANNs) have garnered attention for system
identification and control, including aerodynamic applications (PHAN et al.,
1995). They have also been used for turbulence flow control to reduce
skin friction drag (Lee et al., 1997). Genetic algorithms (GA) are widely
used in active control for engineering applications (Dracopoulos, 2013;
Fleming and Purshouse, 2002) and in various flow control plants. GAs
have been used for control design in fluids for experimental mixing optimisation (Benard et al., 2016). Deep reinforcement learning (DRL) is
a promising candidate for flow control applications (Rabault et al., 2019;
Rabault and Kuhnle, 2019). Its use in exploring strong turbulent fluctuations during glider flight was investigated by Reddy et al. (2016).
Optimisation of fin thickness, width, and spacing in a heat sink with
plate fins was shown by Alayil and Balaji (2015). Their method used
ANN to maximise thermal performance, specifically reducing the time to
reach the set temperature. A predictive model for thermal performance
was developed by Bhamare et al. (2021), employing a range of ML and
deep learning methodologies. Their ANN model was able to predict the
heat flux accurately when compared with experimental values. Utilising
considerations such as node arrangement, power consumption, and cooling system configuration, Piatek
˛ et al. (2015) proposed a framework for
thermal modelling and control of HPC systems. DRL has been used to
create room temperature control policies (Di Natale et al., 2021). The
DRL agents use an augmented reward function, to balance comfort levels
and energy conservation, outperforming traditional rule-based controllers
both in simulation and real-world application.
18
Jerol Soibam
Chapter 2. Theoretical Background
2.2.3
Flow modelling
Turbulence, a phenomenon in fluid dynamics marked by unpredictable,
time-varying fluctuations in velocity and pressure, remains a persistent,
unresolved challenge (Jiménez and Moin, 1991; Drikakis, 2003). To accurately capture the transient flow physics with Computational Fluid Dynamics (CFD), an enormous number of grid points would be needed.
The most precise solution in fluid mechanics, directly solving the NavierStokes equations, known as Direct Numerical Simulation (DNS), is computationally impractical for high Reynolds numbers. In contrast, Reynolds
Average Navier Stokes (RANS) simulations, rely on approximated equations and are widely used in the industry due to their cost-effectiveness, at
the expense of accuracy. The RANS approach requires introducing new
terms to close the system of equations, with commonly used models being
k-epsilon, k-omega, and Spallart–Allmaras. However, no current models
are universally applicable across varied conditions (Brunton and Noack,
2015). Selecting the correct model, along with the appropriate boundary
and initial conditions (such as turbulence kinetic energy and dissipation
rate), is critical for accurately portraying the average flow behaviour.
Closure model
Turbulent and bubbly flows present significant spatial-temporal scale separation, making it computationally expensive to address all the scales in
simulations. Even with significant advances in computational power, full
resolution of all scales within a configuration remains decades away. As a
result, it’s typical to truncate the smaller scales and model their influence
on larger ones using a closure model, a strategy employed in RANS and
Large Eddy Simulation (LES). However, such models necessitate meticulous adjustment to correspond with fully resolved DNS simulations or experimental data. The application of ML algorithms for the development of
turbulence closures is an active field of research (Duraisamy et al., 2019).
ML methodologies have been utilised to discover and model inconsistencies in the Reynold stress tensor between RANS models and DNS simulations (Ling and Templeton, 2015). For LES closures, ANN has been used
by Maulik et al. (2019) to predict the turbulence source term from roughly
resolved quantities. In the case of heated cavity flow, a data-driven closure
model has been adopted to solve the Boussinesq approximations (San and
Maulik, 2018). A machine learning and field inversion framework was
devised by Parish and Duraisamy (2016) to construct corrective models
based on inverse modelling. This framework was subsequently used by
Singh et al. (2017) to create an NN-enhanced correction to the SpalartAllmaras RANS model, demonstrating excellent performance. A novel
Mälardalen University Press Dissertations
19
Machine Learning Techniques for Enhanced Heat Transfer Modelling
architecture that incorporated a multiplicative layer to embed Galilean invariance into tensor predictions was developed by Ling et al. (2016b). It
offers a unique and simple way to incorporate known physical symmetries and invariances into the learning architecture (Ling et al., 2016a),
which could prove essential in future endeavours that integrate learning
and physics.
Reduced order model
Over time, a significant amount of work has been devoted to the development of precise and efficient Reduced Order Models (ROMs) that can
encapsulate crucial flow and heat transfer behaviour with minimal computational expense. One such model reduction technique is the Galerkin
projection (Proctor et al., 2016) of Navier-Stokes equations onto an orthogonal basis of Proper Orthogonal Decomposition (POD) modes, which
benefits from a strong link to the governing equations. However, this approach is intrusive and poses challenges for non-linear systems. One of
the ML techniques employed in ROM is Dynamic Mode Decomposition
(DMD) (Rowley et al., 2009), an infinite-dimensional linear operator illustrating the temporal evolution of all measurement functions of the system. However, DMD is based on linear flow field measurements and the
derived models may not successfully capture non-linear transients. In contrast, ANNs can effectively handle large data volumes and non-linear challenges, such as those presented by near-wall turbulent flow (Milano and
Koumoutsakos, 2002). ANNs have been utilised by Sarghini et al. (2003)
to learn from LES channel flow data, enabling the identification and reproduction of the highly non-linear behaviour characteristic of turbulent
flows. The fusion of CNNs with ANNs has been deployed to create a
time-dependent turbulent inflow generator for a fully developed turbulent
channel flow (Fukami et al., 2019).
In the context of heat transfer, ANN models have been successfully
used to accurately predict convective heat transfer coefficients within a
tube (Jambunathan et al., 1996). Experimental data gathered from an impingement jet with variable nozzle diameter and Reynolds number, have
been employed to construct an ANN model. Later this model was used
to predict the Nusselt number with an error below 4.5% (Celik et al.,
2009). A random forest model was developed by Breiman (1996) for
predicting the heat transfer coefficient of complex geometries. When the
results from this model were compared with those from CFD simulations,
a strong correlation was noted, demonstrated by an R2 value of 0.966.
In their research, Jiang et al. (2013) explored the application of a Support Vector Machine (SVM) for predicting the Critical Heat Flux (CHF).
20
Jerol Soibam
Chapter 2. Theoretical Background
The results obtained from their model showed a strong agreement with
experimental data. Previously, Gaussian process regression (GPR) (Jiang
et al. (2020); Baraldi and Mangili (2015)) has been used to predict CHF
in minichannel. The advantage of the Gaussian process is that it can account for a priori knowledge and estimate model uncertainty. Traverso
et al. (2023) employed GPR to predict the heat transfer coefficient in a
microchannel. Their study results indicate that the GPR model exhibited consistent and reliable performance across diverse training datasets,
demonstrating its applicability in cost-effective engineering applications.
An in-depth review of the literature on the application of ML in heat transfer was conducted by Hughes et al. (2021). They evaluated the range of
ML’s applications, from creating efficient models for precise predictions
and robust optimisation to its usage in ROMs and optimisation of largescale systems.
Despite the impressive ability of ML models to accurately map highdimensional inputs to outputs, they typically demand large amounts of
data, the acquisition of which can be computationally costly and require
meticulous experiment design (Karniadakis et al., 2021). Consequently,
other researchers have pursued methods that combine physical principles
with ML techniques. The concept of physics-constrained learning was
introduced as early as the ’90s, aiming to solve classical differential equations (Lee and Kang, 1990; Lagaris et al., 1998). Recently, Rassi et al.
proposed a deep learning framework for addressing forward and inverse
problems known as the physics-informed neural network (PINN) (Raissi
et al., 2019b, 2020, 2019c). Sun et al. (2020) demonstrated the usage
of PINN in fluid flow applications without relying on any simulation data,
achieved by rigidly imposing initial and boundary conditions. Meanwhile,
Cai et al. (2021a) used PINN to derive velocity and pressure fields from
temperature data gathered from a background-oriented Schlieren experimental setup, designing a PINN framework capable of predicting both
fields without any information on initial and boundary conditions. Lucor
et al. (2022) employed PINN for surrogate modelling of turbulent natural
convection flows, primarily using direct numerical simulation (DNS) data
for network training. Cai et al. (2021b) used PINN for inverse heat transfer
problems and two-phase Stefan problems with a moving interface. They
further suggested that placing the thermal sensors based on the residual of
the energy equation gave the best performance.
From the literature survey, it is evident that extensive efforts have previously been exerted on probing the application of ML within the field
of fluid dynamics and heat transfer. The ML algorithms typically referenced in existing literature largely utilise deterministic methods that unfortunately fall short in supplying predictive uncertainty information. Yet,
Mälardalen University Press Dissertations
21
Machine Learning Techniques for Enhanced Heat Transfer Modelling
to understand the application of these ML models to engineering complexities is still in its infancy, with a wealth of information waiting to be
discovered. Therefore, this thesis aims to delve into the exploration of ML
techniques for heat transfer problems, placing a particular focus on experimental data processing and reduced order modelling. More specifically,
the present work examines the use of deep learning models for subcooled
boiling and the efficient and effective utilisation of data. Furthermore, this
thesis explores methods to navigate the complexities of unknown thermal
boundary (ill-posed) problems. The central theme of this thesis revolves
around developing robust ML models, with significant efforts directed towards accounting for the uncertainty present in the model, which can further illuminate the model’s behaviour.
22
Jerol Soibam
3
Methodology
In this chapter, a detailed framework of machine learning algorithms used
in this thesis is provided. Relevant theories and methods are presented in
detail.
3.1
Framework of the thesis
The thesis is based on a research framework that guides the exploration
of ML applications in heat transfer problems. Central to this framework
is the evaluation of the robustness of ML models, an aspect that directly
influences their performance and reliability. The framework delineates a
clear method for selecting and verifying ML models, taking into account
the nature and availability of data. It serves as the foundation for the
research presented in the associated publications, providing a structured
approach to addressing the research questions. In essence, the framework
not only organises the research process but also directs the exploration
and findings that form the crux of this thesis. The overall learning system adapted in this thesis, including the robustness assessment of the ML
models, is illustrated in Figure 3.1. Detailed explanations of each module
are provided in the subsequent sections.
Figure 3.1: Detailed framework of the learning system.
Mälardalen University Press Dissertations
23
Machine Learning Techniques for Enhanced Heat Transfer Modelling
3.2
Data availability
The data considered in this thesis mainly relies on heat transfer data namely
subcooled boiling and convective heat transfer data. The data used is either obtained from experiments or numerical studies which serve as the
backbone of the architecture to be trained. Furthermore, the quantity of
data plays a crucial role in determining what architecture is appropriate to
follow up the study.
Data for subcooled boiling heat transfer
The data used for training the Deep Neural Network (DNN) in Papers I
and II is of a numerical nature, specifically related to the phenomenon of
subcooled boiling heat transfer. It is extracted from numerical simulations
with varying degrees of heat flux and flow inlet conditions, as displayed in
Table 3.1. The computational domain used for CFD simulation is shown
in Figure 3.2a. For a more comprehensive understanding of the details
of data, readers are directed to Paper II and Rabhi et al. (2021). Note:
Numerical simulation is beyond the scope of this thesis.
Turning to Paper III, this utilises data to predict the critical heat flux,
sourced from pre-existing literature (Kim et al., 2000). The paper offers
a detailed breakdown of the data and includes a look-up table for the data
that was used in the validation process.
Up until this point, the focus has been on field data or point data. However, Paper V presents a shift in the nature of data used, as it incorporates
image data derived from an experimental study on subcooled boiling as
shown in Figure 3.2b. The image data, captured via a high-speed camera,
was directly used for labelling, thus eliminating the requirement for any
image processing techniques. Detailed descriptions and further specifics
regarding the data used in this case study are available in Paper V.
Table 3.1: Data split for training, validation, and testing at different heat flux
and flow conditions.
Cases
Training / Validation
Interpolation
Interpolation
Interpolation
Extrapolation
Extrapolation
Extreme extrapolation
24
Heat flux (q ) [W m−2 ]
Velocity (u) [ms−1 ]
[1000 - 29,000]
14,000
17,500
19,000
30,000
30,000
40,000
[0.05 - 0.2]
0.05
0.075
0.15
0.1
0.15
0.2
Jerol Soibam
Chapter 3. Methodology
Figure 3.2: Numerical computational domain and experimental setup.
Data for convective heat transfer
The data used in Paper IV and VI is an incompressible laminar flow scenario with convective heat transfer under the Boussinesq approximation.
In Paper IV, the case considered is a steady state forced convection with
constant heat flux boundary condition on the source as shown in Figure 3.3a. Whereas, in Paper VI the data used is from a transient mixed
convection setup for constant heat flux and constant temperature boundary on the source as shown in Figure 3.3b. The forced and mixed convection problems are solved using the governing equations - the NavierStokes equations and the energy equation, which are presented in their
non-dimensional form as follows:
∂ u∗ ∂ v∗
+
=0
∂ x ∗ ∂ y∗
∗
∗
∂ u∗
1 ∂ 2 u∗ ∂ 2 u∗
∗∂u
∗∂u
∗
+u
+v
= −∇px +
+
∂t ∗
∂ x∗
∂ y∗
Re ∂ x∗2 ∂ y∗2
∗
∗
1 ∂ 2 v∗ ∂ 2 v∗
∂ v∗
∗∂v
∗∂v
∗
+
u
+
v
=
−∇p
+
+
+ Riθ ∗
y
∂t ∗
∂ x∗
∂ y∗
Re ∂ x∗2 ∂ y∗2
2 ∗
∗
∗
∂ θ
∂θ∗
1
∂ 2θ ∗
∗∂θ
∗∂θ
+u
+v
=
+ ∗2
∂t ∗
∂ x∗
∂ y∗
Pec ∂ x∗2
∂y
(3.1)
where u∗ , v∗ , p∗ , and θ ∗ are the non-dimensional velocity-x, velocityy, pressure, and temperature fields respectively. The dimensionless quantities Re, Ri, and Pec correspond to the Reynolds number, Richardson
Mälardalen University Press Dissertations
25
Machine Learning Techniques for Enhanced Heat Transfer Modelling
Figure 3.3: Numerical computational domain for forced and mixed convection.
number, and Péclet number respectively. In the case of forced convection
Ri = 0. Moreover, data obtained from a direct numerical simulation (DNS)
(Fraigneau, Y., 2019) for the mixed convection at constant temperature on
the source was also used in this thesis. More details on the data can be
found in Papers IV and VI.
Data from experiments or numerical simulations often demonstrates
significant variations in magnitudes, uncertainties, units, and ranges. Such
variations can impact the effectiveness and efficiency of ML models, especially those that are sensitive to the scale of the features. To rectify
this, data processing is carried out, which often involves scaling. This
procedure ensures no specific feature dominates the model due to a larger
numeric range. Multiple scaling techniques exist and are selected based
on the nature of the data and the model being used. Furthermore, outliers or points that diverge significantly from other observations, can skew
model learning and hamper performance. The strategies range from eliminating outliers, and replacing them with statistical measures such as mean
or median, to employing robust models that are less sensitive to outliers.
3.3
Architectures
This thesis carefully investigates a spectrum of methodologies, aligning
each to the individual requisites and performance standards of each task.
It specifically concentrates on numerical data, experimental data, and sensor data. The ambition is to formulate robust and adaptable algorithms to
tackle both direct and inverse problems, with a focus on specific circumstances and requirements.
3.3.1
Deep neural network
An artificial neural network (ANN), with more than one hidden layer is
commonly known as a deep neural network (DNN). The DNN essentially
26
Jerol Soibam
Chapter 3. Methodology
involves a more complex structure of interconnected artificial neurons, or
nodes, extending the capabilities of traditional ANNs as shown in Figure
3.4. This complexity allows the DNN to capture higher-level features and
relationships within the data. It utilises a feed-forward phase where input
signals are passed sequentially through layers until they produce the outputs. The layers undergo a series of transformations influenced by weight
and bias parameters, paired with an activation function.
The transformations can be represented as:
h1 =
g(W1T x + b1 )
..
h5 = g(W5T h4 + b5 )
ŷ = g(W6T h5 + b6 )
0 for x < 0
g(x) =
x for x ≥ 0
(3.2)
(3.3)
where h is the hidden layer and g is the activation function. The objective is to minimise the error, calculated using the mean square error (MSE)
loss function, between the predicted and target values. The target values in
Paper I and II are wall temperature (Twall ) and void fraction (α). Once this
loss is computed, the error gradient relative to the weights (W ) and biases
(b) across all layers can be determined during a backward phase, using
the chain rule of differential calculus. The weights and biases are then updated with the Adaptive Moment Estimation (Adam) optimiser (Kingma
and Ba, 2014), known for handling large datasets and sparse gradients.
To prevent overfitting when handling a high volume of parameters,
a regularization term, specifically L2 norm (Ridge regression), is introduced into the loss function, making the network less complex by rendering some neurons negligible. This method shrinks the weight parameters
Figure 3.4: Deep neural network architecture.
Mälardalen University Press Dissertations
27
Machine Learning Techniques for Enhanced Heat Transfer Modelling
towards zero but not to zero, offering a balanced regularisation. This approach reduces the variance and improves the model’s generalisation to
new datasets.
N
L = MSE(y, ŷ) + λ ∑ w2i L2 norm
(3.4)
i=1
For an in-depth understanding of the DNN structure, Paper II offers
comprehensive insights into its architecture. It meticulously lays out the
hyperparameters employed in the training process of the DNN.
3.3.2
Physics informed neural network
The suggested strategy uses inverse PINN to tackle the heat transfer problem with unknown thermal boundary conditions (Raissi et al., 2019b). The
PINN framework utilises a two-part deep neural architecture, one trained
to represent the fields of velocities, pressure, and temperature, and another enforcing the Navier-Stokes and energy equations at random points
as shown in Figure 3.5. The network inputs are the non-dimensional coordinates and time of the given domain, mapping to expected outputs of
velocities in x and y components, pressure, and temperature.
(x∗ , y∗ ,t ∗ ) → (u∗ , v∗ , p∗ , θ ∗ )
(3.5)
The network’s effectiveness is enhanced by minimising a specified
loss term, and it comprises of four components:
L = Lr + Lub + Lθ b + LQR ,
(3.6)
where,
Lr =
1 4 Nr
∑ ∑ |e j (xi , yi ,t i )|2
Nr j=1
i=1
Lub =
1 Nub
∑ |u(xi , yi ,t i )) − uib |2
Nub Nt i=1
Lθ b =
1
Nθ b
∑
Nθ b Nt i=1
NQR
LQR =
1
(3.7)
|θ (xi , yi ,t i )) − θbi |2
∑
NQR Nt i=1
|Π(xi , yi ,t i )) − ΠiQR |2 .
Lr penalizes the continuity, momentum, and heat equation, computed
from residuals obtained through automatic differentiation. Lub is the loss
function for velocity boundary conditions on source walls, often known
28
Jerol Soibam
Chapter 3. Methodology
Figure 3.5: PINN framework for forced and mixed convection.
as no-slip conditions, are applied. Lθ b minimizes the inlet temperature.
LQR minimizes the difference between sensor values and predicted values. Automatic differentiation, which yields derivatives of the output relative to the inputs, is the crucial element of the PINN framework. This
mechanism forms the foundation upon which the PINN model operates.
In the training phase, the PINN model incorporates initial and inlet conditions from a numerical simulation and adjusts the parameters.
The optimisation process uses sensor locations identified by QR pivoting,
which serve both temperature and velocity measurements. PINN’s approach is different from traditional DNNs as it not only learns from data
but also incorporates known physical laws, enhancing its accuracy and
generalisation capabilities in solving intricate physical problems. An indepth formulation of the PINN architecture and sensor selection adopted
in this thesis can be found in Papers IV and VI.
3.3.3
Convolutional neural network
Convolutional neural networks are designed to handle multi-dimensional
data with spatial information. They require fewer parameters to optimise
than traditional feed-forward neural networks, making them more efficient. In object detection, two methods are generally employed: the onestage method and the two-stage method. The one-stage method, such as
you only look once (YOLO) (Redmon et al., 2016), detects the object’s
position, segment, and classifies it simultaneously, while the two-stage
method first predicts the position, and then classifies the object. The architecture in this thesis focuses on the one-stage method for its benefits of
speed and real-time prediction capabilities without compromising accuracy. The architecture’s strength lies in its division into three main components: the backbone, neck, and head networks as shown in Figure 3.6.
Mälardalen University Press Dissertations
29
Machine Learning Techniques for Enhanced Heat Transfer Modelling
Figure 3.6: Convolutional neural network for bubble segmentation and tracking.
Each part performs a distinct role within the overarching system, sequentially processing input data and passing the output onto the subsequent
component. This modular approach fosters a better understanding of the
object’s spatial and contextual information, thus facilitating more precise
object detection and segmentation. Proceeding further, a comprehensive
clarification of the distinct functionalities and intricacies of the three cardinal components within the YOLO architecture (Wang et al., 2022): the
backbone, neck, and head networks, is presented.
• Backbone: Extracts various features such as shape, size, edges, texture, and background from the input image through a series of operations. It outputs multi-scale feature maps which are then processed
by an extended-efficient layer aggregation network (E-ELAN) and
a max pooling module for further feature extraction.
• Neck: Ingests the output from the backbone and integrates these
multi-scale feature maps. It uses the path aggregate feature pyramid network (PAFPN) and the cross-stage partial spatial pyramid
pooling (SPPCSPC) to pool features from all levels, bridging the
gap between lower and upper feature levels and reducing computation costs.
• Head: Processes the features extracted by the backbone and neck
with a convolution layer before passing them to the head network.
The head network, based on yolact++ (Bolya et al., 2022), divides
the complex task of instance segmentation into two parallel tasks
that merge to form the final masks.
30
Jerol Soibam
Chapter 3. Methodology
The network is trained using four types of loss:
L = lbox + lob j + lcls + lmask
(3.8)
where L is the total loss function used to optimize the model. lbox is
the box regression loss, lob j is the objectness loss, lcls is the classification
loss, and lmask is the mask loss of the segmentation. It is of note that a
technique like transfer learning learning was adopted in this study for data
efficiency and also to improve the accuracy of the model. For a more detailed description of each component’s architecture within the backbone,
neck, and head, please refer to Paper V.
3.3.4
Gaussian process regression
Gaussian process regression (GPR) is a Bayesian technique employed for
nonlinear regression, offering flexibility and probabilistic modelling that
provides robust posterior. However, the traditional GPR tends to be computationally expensive and its complexity increases with the size of the
dataset, as it requires access to the full dataset during both the training and
prediction phases. For example, for a data size of N exact inference has a
computational complexity of O(N 3 ) with a storage demand of O(N 2 ).
To address these challenges, Parametric Gaussian process regression
(pGPR) (Raissi et al., 2019a) is used. In pGPR, the model is trained on
a hypothetical dataset, which is usually smaller than the actual data as
shown in Figure 3.7. This approach enhances data efficiency as it focuses on the hypothetical dataset, obtained via k-means clustering on data
points of the critical heat flux. Once the hypothetical dataset is established, pGPR is defined by the conditional distribution and its parameters are updated via the posterior distribution conditioned on a mini-batch
of observed data. The initial parameters of the pGPR model are trained
3
f (x)
3
6000 data points
2
2
8 Hypothetical data points
True function = x ∗ cos(4 ∗ 3.14 ∗ x)
Predicted function
±3σ
f (x)
f (x)
1
0
1
0
−1
−1
−2
−1.5
−1.0
−0.5
0.0
x
0.5
1.0
1.5
−2
−1.5
−1.0
−0.5
0.0
x
0.5
1.0
1.5
Figure 3.7: Framework for parametric Gaussian process regression.
Mälardalen University Press Dissertations
31
Machine Learning Techniques for Enhanced Heat Transfer Modelling
and then used for predictions on new data points. Like GPR, uncertainty
in pGPR is quantified by calculating the predicted variance, which provides a measure of the reliability of the model’s predictions. In essence,
it gives an indication of the spread of possible output values predicted by
the model. The smaller the predicted variance, the greater the confidence
in the model’s predicted output.
The choice of kernel function significantly influences a pGPR model’s
generalisation capabilities. In this study, the kernel function is assumed to
have a squared exponential covariance function, which is chosen for its
ability to model smooth functions. This kernel function includes variance
and length-scale hyperparameters. A more detailed formulation of the
pGPR for predicting critical heat flux can be found in Paper III.
3.4
Uncertainty quantification for robustness
Uncertainty quantification (UQ) in deep learning models is critical for
a variety of reasons. It aids in gauging the model’s confidence level in
its predictions, providing a safety measure against overconfident and potentially erroneous outputs. Additionally, it facilitates risk assessment in
practical applications where understanding the risk associated with each
prediction is paramount. Furthermore, it helps identify areas requiring
further training or data acquisition, thereby improving the model performance. Lastly, in scientific research, uncertainty quantification aids in
validating findings, enhancing the overall transparency and robustness of
the research process.
This thesis explores three distinct methods. The initial approach leverages the Bayesian methodology, elaborated in Section 3.3.4. The subsequent methods concentrate on UQ within the realm of deep learning, suitable for integration within any DNN or PINN architectures. More details
on these methods can be found in Papers II, III, and VI.
3.4.1
Monte Carlo dropout
Dropout is an effective technique that has been widely used to solve overfitting problems in DNNs just like the regularisation technique. It has
been shown that it can also be used for uncertainty estimation in a DNN,
using Monte Carlo dropout (Gal and Ghahramani, 2016). This technique
interprets a DNN trained with dropout as a Bayesian approximation of
a Gaussian process. It applies dropout not only during training but also
prediction as shown in Figure 3.8.
Given a trained network with an input feature X , it provides a predictive mean E(y) and variance Var(y∗ ). A length scale l is defined to cap32
Jerol Soibam
Chapter 3. Methodology
Figure 3.8: Monte Carlo dropout architecture with a dropout ratio of 20% for
training and testing.
ture the belief over data frequency, with shorter scales indicating higher
frequency. The Gaussian process precision (τ) is given as:
τ=
l2 p
2Nλw
(3.9)
where p is the probability of the units not dropped during training,
λw is the weight decay, and N is the dataset size. Dropout is also used
during the prediction phase. Using a dropout ratio of 20% during training
and testing, the prediction step is repeated several times (T ) with different
units dropped each time, giving the results ŷt (x). The predictive mean and
variance of the test data are given by:
E(y) ≈
Var(y) ≈ τ −1 ID +
3.4.2
1 T
∑ ŷt (x∗ )
T t=1
1 T
∑ ŷt (x)T ŷt (x) − E(y)T E(y∗ )
T t=1
(3.10)
(3.11)
Deep ensemble
Deep Ensemble is a non-Bayesian technique for uncertainty quantification in ML models. It involves training several neural networks, rather
than one, to enhance generalisation capabilities (Lakshminarayanan et al.,
2017). For a regression problem, the model assumes the target has a
normal distribution with mean and variance based on the input values
as shown in Figure 3.9. The loss function is adjusted to minimise the
difference between predictive and target distribution using Negative LogLikelihood (NLL) loss:
Mälardalen University Press Dissertations
33
Machine Learning Techniques for Enhanced Heat Transfer Modelling
Figure 3.9: Deep ensemble architecture for M = 2.
L = −logpθ
yn
logσ 2 (X) (y − μθ (X))2
=
+
+c
Xn
2
2σ 2 (X)
(3.12)
In a Deep Ensemble model, several networks are trained with different
random initialisations. Predictions are treated as a uniformly weighted
mixture model, approximated as a Gaussian distribution. The mean and
variance mixture are:
1
μθ m (X)
M∑
m
(3.13)
1
(σ 2 θ m(X) + μθ m2 (X)) − μ∗2 (X)
M∑
m
(3.14)
μ(X) =
σ 2 (X) =
The implementation involves defining a custom NLL loss function and
a custom layer to extract the mean and variance as network output.
34
Jerol Soibam
4
Results and discussion
This chapter describes the most important results from the appended papers. This chapter concludes with a discussion of the research questions
of the implications and potential contributions of these results.
4.1
Case study on boiling heat transfer
Boiling heat transfer plays a crucial role in many engineering and industrial processes such as power generation, cooling of electronics, and chemical processing, with its efficiency and effectiveness greatly influencing
the overall performance of these systems.
4.1.1
DNN for subcooled boiling
This subsection emphasises the outcomes garnered from the DNN model
implemented in Papers I and II. The purpose of this research was to ascertain if a data-driven approach, like DNN, can be harnessed to examine the
inherent physics in subcooled boiling from data, and if it can be used for
the prediction of quantities of interest (QoIs). Three architectures were investigated namely a traditional DNN, Monte Carlo dropout (MC dropout),
and deep ensemble (DE) model. The data used in all the models remains
consistent throughout the study as described in Section 3.2.1.
Validation of the models
The predictive and statistical performance for all the models was initially
evaluated using the validation dataset. The corresponding results are illustrated in Table 4.1. From the Table, it can be noted that all the models
Table 4.1: Performance of the DNN, MC dropout, and DE models on validation
dataset of the computational domain, VF: Void Fraction, Temp:
Temperature.
Case
Dataset
Validation
Models
DNN
MC Dropout
Deep Ensemble
R2
RMSEP
VF
Temp
VF
Temp
0.006
0.006
0.002
0.9982
0.9991
0.9998
0.995
0.997
0.998
Mälardalen University Press Dissertations
0.134
0.125
0.081
35
Machine Learning Techniques for Enhanced Heat Transfer Modelling
have low RMSE for both void fraction and temperature which are the
QoIs. However, it can be observed that the DE model outperforms the
other two models with a lower RMSE and a higher coefficient of determination (R2 ). The conclusion that can be drawn from this is that the DE
model fits better for the validation dataset.
Field prediction of the minichannel
The result presented here is based on an interpolation dataset as described
in Table 3.1 in Chapter 3, with a heat flux of q = 14,000 W m−2 and
an inlet velocity of u = 0.05 ms−1 . As was shown in the previous analysis that the DE model is superior compared to DNN and MC dropout,
hence the predicted fields of temperature and void fraction from the DE
model are plotted in Figure 4.1 and 4.2. The DE model’s predictions of
the minichannel temperature field closely align with the CFD values, as
depicted in Figure 4.1. Upon examination of the relative percentage error
field, it is observed that the majority of errors, lie under ± 0.2%, are concentrated where the heat flux is applied. Once again it can be seen that
the relative error is under 1% for the void fraction as shown in Figure 4.2.
The maximum error occurs at a non-dimensional arc length of 0.5 on the
y-axis. At this location, the void fraction increases suddenly, potentially
due to the formation of a larger bubble or a multitude of smaller bubbles
near the wall, which causes an abrupt increase in void fraction.
Figure 4.1: Temperature field prediction and the relative error between the CFD
and DE model of the minichannel.
36
Jerol Soibam
Chapter 4. Results and discussion
Figure 4.2: Void Fraction field prediction and the relative error between the CFD
and DE model of the minichannel.
Model sensitivity along the heated wall
The section of interest involves the near-wall region of a minichannel,
where heat flux causes sudden shifts in physics. An investigation into
model performance has been done with an extreme extrapolation dataset,
utilising a heat flux of q = 40,000 W m−2 and an inlet velocity of u = 0.2
ms−1 . Results, as seen in Figure 4.3, show the prediction of the void fraction in this area. The dimensionless arc length on the x − axis corresponds
to the minichannel height where heat flux is imposed. The DNN and MC
dropout models struggle to correctly predict the onset of nucleation sites.
Nonetheless, the MC dropout predicted values align well with the trend of
CFD void fraction, with uncertainty starting around 0.1 and peaks at 0.2
arc lengths due to liquid-to-bubble phase shifts. In contrast, the DE model
presents lower variation throughout the void fraction prediction and accu-
(a) DNN and MC dropout
(b) Deep Ensemble
Figure 4.3: Wall void fraction profile by the DNN, MC dropout and DE models
along the arc of the minichannel
Mälardalen University Press Dissertations
37
Machine Learning Techniques for Enhanced Heat Transfer Modelling
rately anticipates the onset of the nucleation site. However, the prediction
uncertainty grows with the void fraction, possibly due to the coalescing of
smaller bubbles into larger ones. The uncertainty then starts to decrease
around 0.7 arc length as the predicted void fraction from the DE model
converges with the CFD values. This suggests that the DE model can effectively recognise changes from liquid to vapour and increased bubble
presence, displaying greater uncertainty in such areas.
For a more comprehensive understanding of the predictive performance and robustness of the DNN, MC dropout and DE models, please
refer to Papers I and II. These studies delve into intricate details of the
models’ behaviour, addressing their capabilities, and discussing the uncertainties surrounding various aspects. They also further discuss the reasons
behind the uncertainties in their predictions and how these models handle
them in various situations.
4.1.2
Computer vision for subcooled boiling
This section illustrates the efficacy of the proposed CNN model in bubble identification within a constant heat flux rectangular mini-channel as
shown in Figure 3.2b. The model addresses the shortcomings of traditional methods, ensuring precise and dependable segmentation. The results underline the CNN model’s superior capability, enabling an in-depth
study of bubble behaviours.
Figure 4.4: Comparison of mask between ground truth (human eyes), classical
image processing and CNN model. a) Ground truth mask b)
Classical mask c) CNN mask d) Normal image with a surface void
fraction of 6.45% e) IoU for classical method and f) IoU for CNN
model
38
Jerol Soibam
Chapter 4. Results and discussion
Comparison and sensitivity analysis
The bubble segmentation mask from the CNN model is compared against
the ground truth masks and traditional image processing masks as shown
in Figure 4.4. The ground truth mask is converted into a binary image,
which sometimes lacks clear bubble boundaries due to uneven illumination and pixelation. Classical methods overestimate bubble area by not
distinguishing between bubbles and shadows and misinterpreting closely
located bubbles as one. Its intersection over union (IoU) score is 71%.
The CNN model performs better, as it can separate bubbles and shadows
and identify individual bubbles in close proximity. The model captures
tiny bubbles effectively and its IoU score is 88% against the ground truth
mask. The maximum pixel difference with the ground truth is 2 pixels,
consistent with the average uncertainty found in ground truth masks, indicating the model’s performance is as good as human labelling, despite
image downscaling.
The model’s sensitivity was evaluated under different conditions, as
depicted in Figure 4.5. Though the model was not trained with noise or
Figure 4.5: Masks given by the model with different noise conditions: a)
Original image with the mask, b) Contrast enhancement, c)
Sharpening, d) Top-hat filter with disk shape radius = 20, e)
Gaussian Blur σ = 0.8, f) Gaussian noise σ = 10, g) Gaussian noise
σ = 25, h) Black dead pixel 0.02% and i) IoU for noisy images
Mälardalen University Press Dissertations
39
Machine Learning Techniques for Enhanced Heat Transfer Modelling
enhancements in the raw image, it was able to handle the noisy conditions and predict the mask accurately. Enhancements such as increased
contrast and sharpness didn’t affect the model’s performance. A top-hat
filter, applied to one of the images, caused slight deviation, yet the model’s
overall predictive capability remained strong. The introduction of a Gaussian noise at σ = 10 saw the model performing robustly. It, however,
struggled slightly with tiny bubbles and shadow differentiation at higher
noise levels (σ = 25). The presence of dead pixels led to a few incorrect
predictions, showing the model’s sensitivity towards them. The model’s
performance under different noise conditions was evaluated using the IoU
measure. The IoU trend indicated the model’s ability to handle noise with
minor improvements or decreases in prediction accuracy. However, the
model showed a noticeable decline in performance under high Gaussian
noise at σ = 25.
Local bubble statistics
Figure 4.6: Results extracted from the binary mask given by CNN model to
study local bubbles behaviour: a) Trajectory of the bubbles with
velocity magnitude and growing bubble images (Note: axes are not
in scale), b) Bubble diameter with time, c) Bubble Reynolds number
with time.
40
Jerol Soibam
Chapter 4. Results and discussion
The mask generated by the CNN model is used for identifying bubble
nucleation sites, edges, and centroids. This information provides insights
into the bubbles’ trajectory, equivalent diameter, and velocity. Local bubble behaviours such as coalescence, collision, oscillations, and condensation, which lead to changes in size, shape, and speed, are examined by
tracking bubble centroids over time and correlating the images.
The mask generated by the CNN model and post-processed is shown
in Figure 4.6 for one specific nucleation site, undisturbed during the bubble life cycle. The bubble centroid’s position is used to calculate its trajectory, velocity, and magnitude. The growth of bubbles over time is shown
in Figure 4.6a, while the equivalent diameter is depicted in Figure 4.6b.
Figure 4.6c reveals the local Reynolds number during the bubble’s lifecycle until it condenses. The model immediately recognises a newly formed
bubble which grows on the nucleation site for a characteristic time tg of
5ms. In this phase, the diameter grows while the centroid and local velocity remain almost constant. After growth time, the bubble detaches and
moves with the flow, its diameter increasing until it reaches its final size,
driving up its velocity. Finally, the bubble detaches completely and condenses within the flow. A more detailed study and analysis can be found
in Paper V.
4.1.3
Critical heat flux
In this section, the predictive performance of pGPR for critical heat flux
(CHF) is discussed. The CHF represents the most efficient heat transfer regime when a liquid coolant undergoes a phase change on a heated
surface. However, there is a risk: should the heated surface temperature
rise significantly, it could reach the material’s melting temperature, causing a ’burn-out’ phenomenon which could result in catastrophic failure.
Therefore, particular efforts have been made to study the CHF from a data
perspective using pGPR for accurate prediction. A further intriguing aspect of this model is that it employs data efficiently by introducing the
concept of a hypothetical dataset, thus circumventing the need for a large
dataset.
The pGPR model and the look-up table (LUT) predictive performance
are showcased in Figure 4.7. The plot’s solid line signifies a perfect match
between predicted and experimental data, whereas the dashed line denotes
a ±10% error margin. It’s evident that most predictions fall within a ±5%
error, peaking at ±10%. Conversely, the LUT’s results diverge substantially from the optimal line, making accurate CHF predictions impossible.
The pGPR model surpasses the LUT predictions, and when compared to
several CHF correlations, previously evaluated by Kim et al. (2000), it
Mälardalen University Press Dissertations
41
Machine Learning Techniques for Enhanced Heat Transfer Modelling
Figure 4.7: Validation result of the CHF for pGPR and LUT predicted values
shows fewer errors. Most pGPR predictions fall on the optimal line, with
a few nearing the ±10% error line, significantly improving upon the average ±15% error observed in other models.
(a) Influence of the input parameters on the
CHF
(b) Prediction of CHF with varying mass flux
with uncertainty
Figure 4.8: Weighting factor and prediction obtained from pGPR model
The model illustrates the weighting factor for each input feature contributing to the prediction of CHF in Figure 4.8a. In this dataset, the heated
length (Lh ) carries the highest absolute weighting factor, superseded only
by the mass flux (G). In contrast, the lowest weights are ascribed to the
inlet subcooling and system pressure. These patterns align with the experimental data from Kim et al. (2000), which demonstrates a pronounced
dependency of CHF on Lh and G, but lesser sensitivity to flow pressure
(P) and inlet subcooling (Δhi ). This sheds light on how variations in the
input variables might impact the prediction of CHF.
The experimental data, pGPR model predictions, and associated uncertainty are presented in Figure 4.8b. The model proves effective in mapping the trends of the experimental data when the mass flux is adjusted and
42
Jerol Soibam
Chapter 4. Results and discussion
has been rigorously tested for a range of lengths, diameters, and pressures.
Despite minor discrepancies between the pGPR model’s predictions and
the experimental data, a notable level of uncertainty is evident across all
predictions, regardless of the conditions. This considerable uncertainty is
likely attributable to the high degree of nonlinearity and uncertainty encountered in the CHF dataset, which the model uses for its predictions.
Even with these challenges, the pGPR model demonstrates its value by
effectively distilling information from the experimental data and mirroring it with the testing dataset. This alignment with the underlying physics
further attests to the model’s capability and robustness. For a comprehensive examination and analysis of this study, please refer to Paper III.
4.2
Case study on convective heat transfer
In this section, particular attention is paid to cases where the thermal
boundary remains unknown, rendering numerical simulations infeasible,
regardless of the sophistication of the tools available. As a solution, inverse PINN is employed to address this problem, using certain sensor
points to identify the thermal boundary as well as to predict the full flow
field. The L2 relative error is used to evaluate the performance of the PINN
and it is given as:
ε=
||P − R||2
,
||R||2
(4.1)
where P represents the predicted quantities obtained from PINN model
and R is the corresponding reference values.
This technique has been tested for steady-state forced convection and
transient mixed convection.
4.2.1
Forced convection
The case study considered here is a flow around an enclosure with a heated
source. The numerical results are obtained from a CFD simulation with a
constant heat flux on the source wall for the conditions Re = 100 and Pec
= 75, and are used to compare with the prediction obtained from PINN.
Sensitivity of sensor placement
The PINN model’s predictive performance was tested in relation to sensor quantity and location through various case studies as shown in Figure
4.9. Four sensors were initially placed at the square source’s corners (case
study 1), producing around 9.2% temperature error compared to CFD.
Mälardalen University Press Dissertations
43
Machine Learning Techniques for Enhanced Heat Transfer Modelling
Figure 4.9: Case studies for different sensor placements to infer temperature
distribution in the enclosure domain
Active sensor placement added an extra point (case study 2), but this provided no significant accuracy improvement. This active sensor is achieved
based on the highest residual error when subject to the energy equation
(er = uTx + vTy − Pec−1 (Txx + Tyy )). Sensor count was increased to eight
in case study 3, slightly improving predictive accuracy, though this caused
some unrealistic temperature distribution predictions.
Sensor locations were revised for case study 4, yielding better performance than the first two case studies. Active sensor placement added
another sensor near the source wall’s right corner in case study 5, showing
immediate improvement. Another sensor, based on max residual error,
was added in case study 6, reducing the error to 1.4%. Thus, optimal
sensor location and minimum quantity, rather than sheer number, proved
crucial for accurate thermal distribution. Note that errors in u, v, and p
remained consistent across studies.
Temperature distribution
The thermal sensor positioning of case study 6 underpins the field predictions illustrated in Figure 4.10. The left portion of the figure conveys
the predictions made by the PINN model, the middle portion displays the
results of the CFD simulation, and the right portion shows the L2 relative error. A comparison of the two temperature profiles reveals the PINN
model’s ability to accurately reproduce the global temperature field, requiring only 6 sensor data points. Moving on to the examination of the
TError contour, the largest absolute errors (not presented as percentages)
are found to cluster near the source corner and the outlet. Despite this, the
44
Jerol Soibam
Chapter 4. Results and discussion
Figure 4.10: Case studies for different sensor placement to infer temperature
distribution in the enclosure domain
errors are fairly small in magnitude, suggesting the potential for further
accuracy enhancements through the addition of extra sensors.
Transfer learning
Training PINN models typically demands substantial computational power
due to the application of first-order optimisation techniques. However, using transfer learning, this training time can be significantly reduced. This
involves reusing the weights and biases from a previously trained network
for new configurations of a similar problem. For instance, weights and
biases from the case study of 6 sensor placements (Figure 4.9) have been
used as starting points for tuning a new network for different shapes and
two sources, utilising 8 sensor points on each source wall. The network
was then trained for 5000 iterations, with a learning rate of 1×10−6 . Compared to the previous case it showed an inferior predictive performance
with a maximum L2 relative error of 6.5%. It is noticeable that the transferred PINN model over-predicts around the source and under-predicts
between the two rectangular sources. Importantly, while training the inverse PINN from scratch took about 20 minutes on an nvidia RTX 3090
GPU. The use of transfer learning dramatically reduced the time needed
for fine-tuning parameters while retaining an acceptable accuracy.
Figure 4.11: Temperature distribution: Transfer learning case study for two
rectangular sources
Mälardalen University Press Dissertations
45
Machine Learning Techniques for Enhanced Heat Transfer Modelling
4.2.2
Mixed convection
The case study considered here is flow around a heated square cylinder.
Two cases have been studied, firstly, the heat is applied to the source using
constant heat flux at Re = 100, Pec = 100, and Ri = 0.4. In the second
case, constant temperature is imposed on the heated square for the flow
conditions Re = 50, Pec = 50, and Ri = 1.0.
QR sensor placement
The accuracy of inverse PINN simulations is strongly dependent on the
strategic placement of sensors within a given field. QR factorisation with
column pivoting is an effective method to optimise sensor positioning, as
it focuses on capturing the dominant modes of the field. The sensor location obtained from QR pivoting is then used to optimise the inverse PINN.
A performance comparison of PINN over time, utilising sensors derived
from both random selection and QR pivoting, is illustrated in Figure 4.12.
These figures indicate the L2 errors for u∗ and θ ∗ , thereby emphasising the
superior performance of PINN with sensors determined by QR pivoting,
compared to those selected randomly. This trend remains unaltered, despite the persistent changes in the flow’s temporal dynamics due to vortex
shedding from the square source.
The study investigates the relationship between the number of QR sensors used and the accuracy of PINN in reconstructing temperature distribution, as shown in Figure 4.13a. It demonstrates that as the number of
sensors increases, the L2 error decreases due to an enhanced ability to
capture detailed thermal characteristics of the flow. This increase in sensor count subsequently improves the performance of the PINN by offering more locations for optimising the energy and y-momentum equations,
which are vital for temperature field reconstruction. Nevertheless, a pattern of diminishing returns is observed, where further reduction in the L2
Figure 4.12: Comparision of L2 relative error between QR and randomly
selected sensors over time for the component u∗ and θ ∗
46
Jerol Soibam
Chapter 4. Results and discussion
(a) L2 relative error vs number of QR sensors
(b) L2 error for single and ensemble PINN
Figure 4.13: Sensitivity of QR sensors and uncertainty quantification of PINN
error tends to plateau beyond a certain sensor count. One of the primary
goals of this research is to pinpoint both the optimal and minimum necessary sensor locations for the PINN’s flow field reconstruction. Using
QR pivoting, 10 optimal sensors were chosen, offering a balance between
attaining the desired accuracy and constraining the number of sensors deployed.
Uncertainty quantification of PINN
This research compares single and ensemble PINN models. The single
model, trained on a minimal sensor dataset, is tested against the ensemble
model, which uses multiple models each starting from random initialisations. Both models use the same sensor parameters. The ensemble model,
although more computationally demanding, provides superior generalisability, and robust predictions by avoiding overfitting and underfitting as
shown in Figure 4.13b. In addition, it provides the mean predictions and
estimates the uncertainty, aiding in decision-making by indicating trustworthiness and ambiguity in the model’s predictions. Particularly in complex systems, these features are invaluable. The rest of the study concentrates on ensemble PINN predictions.
Constant heat flux
After training the PINN, it was then deployed to ascertain the temperature
profile surrounding the square, as depicted in Figure 4.14a. The graph
demonstrates a temperature profile from the PINN that aligns with the
trend of the values inferred from RANS around the square. The figure
shown was captured at a time of t ∗ = 40 and exhibits a peak L2 error of
8% on the right side where the flow interacts the least with the square.
Notably, the model’s uncertainty diminishes in areas with available sensor
Mälardalen University Press Dissertations
47
Machine Learning Techniques for Enhanced Heat Transfer Modelling
(a) Temperature profile around the square
cylinder
1
(b) Unknown parameters C1 = Ri and C2 = Pec
for
constant heat flux
Figure 4.14: Sensitivity of QR sensors and uncertainty quantification of PINN
data, which was expected. Yet, there is considerable uncertainty on the
left wall. A plausible reason for this could be an insufficiency of temperature data in that area for the PINN model. Likewise, there is notable
uncertainty on the right and bottom walls, potentially due to the impacts
of unstable flow and buoyancy.
During the training of the PINN, another assumption was that the nondimensional numbers, Richardson number (Ri) and Péclet number (Pec),
in the governing equation were unknown. Hence, the PINN was additionally tasked to predict these numbers during the training process. The
predicted values from the PINN and the actual values from RANS are pro1
vided in Figure 4.14b. Here, C1 stands for Ri and C2 represents Pec
. It’s
evident that at the start of training, both numbers are rather random and
exhibit substantial fluctuation and uncertainty. However, when the model
begins to stabilise around 100k epochs, the means of both C1 and C2 draw
closer to the actual values from RANS. On the flip side, the uncertainty
persists to be significant despite a consistent reduction. This can be ascribed to the uncertain thermal boundary condition in the square, given
both C1 and C2 play a role in the heat transfer in the governing equation.
Constant temperature
The analysis in this section focuses on a situation where the source wall’s
temperature remains constant. The insights for this case study have been
garnered from a DNS simulation, as indicated in (Fraigneau, Y., 2019).
Optimal sensor points were identified by applying QR pivoting based on
the temperature distribution. These sensors were then leveraged by the
PINN to reconstruct the flow field. The dynamics, as reconstructed by
the PINN at the 80th time-step, are shown in Figure 4.15. The left part
of the figure exhibits predictions made by the PINN, the centre shows the
48
Jerol Soibam
Chapter 4. Results and discussion
Figure 4.15: Comparison of flow field profiles for constant heat flux at t = 80, as
predicted by PINN and DNS, with corresponding L2 error for u∗ , v∗ ,
p∗ , θ ∗
outcomes from the DNS simulation, and the right side portrays the L2
relative error between the two sets of results. The figure demonstrates that
the error for u∗ , v∗ , and p∗ is below 3%. However, for the temperature θ ∗ ,
the error margin is around 10% close to the square wall, with the majority
of the error stemming from the square’s front side.
The PINN model’s predictive accuracy was evaluated at a random
wake-region point (x∗ =2.5, y∗ =0.1). This location was neither included
in the model’s training data nor was it a QR sensor point, making it a
reliable test of the model’s capacity to generalise from its training. As
depicted in Figure 4.16, the PINN model successfully predicted the components (u∗ , v∗ , p∗ , and θ ∗ ) of the flow dynamics at this point, and these
predictions exhibited strong alignment with the DNS data. This consistency with existing data reinforces the model’s predictive strength and reliability, demonstrating that the model performs well beyond its training
dataset. Despite its overall success, the model did show a higher level of
uncertainty in predicting the pressure component. Even with the increased
deviation in pressure prediction, the model’s uncertainty still manages to
encapsulate the range of the actual DNS values. This broad coverage provides an understanding of the model’s resilience and adaptability, despite
its fluctuating behaviour over time.
For the remaining components, the PINN model showed a notably low
level of uncertainty, highlighting its robustness and ability to maintain
Mälardalen University Press Dissertations
49
Machine Learning Techniques for Enhanced Heat Transfer Modelling
Figure 4.16: Predictive performance of the PINN at random point which was not
a part of QR sensor nor training
accuracy. The model also demonstrated resistance to overfitting, which
means it does not rigidly depend on the specific data it was trained on
and can accurately predict data beyond its training set. The versatility of
this model had been previously established when it was able to successfully reconstruct the full flow field using the QR sensor location data, and
this additional test further emphasises that characteristic. For more details
please refer to Paper VI.
4.3
Discussion of the research questions
The advent of DL, in fluid mechanics and heat transfer has opened a transformative shift in research methodologies. This thesis assesses the applicability of ML methods to navigate the complexities of heat transfer
problems. Although the research questions posed are general in nature,
the conclusions drawn are based on the specific studies conducted in this
study. Within this framework, the potential of ML models to accurately
represent and predict these dynamics is explored. Additionally, the integration of established physical principles, the challenges of limited data,
and the importance of uncertainty quantification in ensuring model transparency are examined.
50
Jerol Soibam
Chapter 4. Results and discussion
RQ 1: How effectively can deep learning techniques capture the intricate nature of heat transfer based on data, and how successfully can these
techniques extrapolate to unseen scenarios? (Papers I & II)
In recent years, ML techniques have been applied to fluid mechanics and
heat transfer problems, driven by an influx of data from experiments and
simulations. The study in Papers I and II showcased data-driven approaches for subcooled boiling heat transfer. Numerical data was used
to train the deep learning models for predicting temperature and void
fraction within the minichannel. Three distinct deep learning methods —
DNN, MC dropout, and deep ensemble — were explored. After training,
the models were validated against a validation dataset and subsequently
tested using the evaluation dataset, as highlighted in section 4.1. These
trained models demonstrated adeptness in capturing the complexities of
subcooled boiling heat transfer. They proficiently predicted the temperature and void fraction in the minichannel, aligning well with the numerical
values. The models further demonstrated their ability to predict the quantities of interest with acceptable accuracy, even under conditions of a high
heat flux of q” = 40,000 W m−2 and an inlet velocity of u = 0.2 ms−1 .
Notably, these conditions were outside the training dataset, indicating that
the model used in this study can extrapolate unseen datasets.
Beyond accurate predictions, the models discerned intrinsic patterns
in the data, thereby providing insights into the physical process of the
subcooled boiling process. For instance, an increased inlet velocity led
to a delay in the onset of nucleate boiling, for high heat flux conditions.
Despite this shift in physics, the models adeptly predicted the onset of
nucleate boiling. This lag can be linked to the slower attainment of saturated temperature at increased inlet velocities, causing a subsequent delay
in bubble formation within the wall of minichannel. The MC dropout and
deep ensemble are based on a probabilistic approach, further highlighting
the models’ predictive performance. As the void fraction approaches a
value of 0.1, a surge in uncertainty was observed, possibly due to the nucleation of bubbles near the wall, prompting a phase transition from liquid
to vapour. Of all the models, the deep ensemble consistently delivered
superior predictive performance for both interpolation and extrapolation
datasets, with a relative error under 2%. The models examined in this
thesis proved their ability to detect shifts in physical states, such as phase
transitions and delays in the onset of nucleate boiling, underscoring the
intricate dynamics of subcooled boiling in the minichannel.
RQ 2: How to incorporate known physics of heat transfer in deep learning
models to improve the generalisability in estimating unknown parameters
Mälardalen University Press Dissertations
51
Machine Learning Techniques for Enhanced Heat Transfer Modelling
and system behaviours? (Papers IV & VI)
Integrating well-understood physics into deep learning models, especially
in the domain of heat transfer, offers substantial improvements in their
ability to estimate unknown parameters, such as thermal boundaries and
non-dimensional parameters. One of the main obstacles in heat transfer
simulations is often the lack of accurate thermal boundary conditions, rendering traditional methods inadequate for certain problems.
In response, the research presented in this thesis utilises the capabilities of inverse PINN. The study delves into both forced and mixed convection heat transfer within enclosures, focusing on scenarios with unknown
thermal boundaries. To address this, the Navier-Stokes and energy equations are enforced at random points in the domain, adhering to known
physical laws. The key findings of the research highlight the efficacy of
the PINN model. Even with a limited set of sensor data, it effectively represents the entire flow field inside the domain and aligns closely with temperature profiles from numerical studies. Additionally, the PINN model
accurately predicted the Péclet and Richardson numbers, which govern
the flow behaviour within the domain. A standout observation concerns
the crucial role of sensor placement. The model’s accuracy and reliability
often depend on this, and methods based on the energy equation’s residual and QR pivoting proved particularly effective. To enhance the robustness and reliability of predictions, the research introduced an ensemble
approach to PINN, resulting in improved accuracy and predictive confidence. In summary, the findings underscore the importance of strategic
sensor placement and ensemble methodologies in ensuring accurate and
reliable predictions, vital for the complexities of industrial heat transfer
systems.
RQ 3: How can machine learning methods secure reliable prediction outcomes in situations where heat transfer data are scarce? (Papers III, IV
& V)
ML models rely on data to make informed predictions. Their predictive
performance depends on the quality and amount of data. When data is
limited, the model’s performance can be significantly compromised, potentially affecting its predictive accuracy and decision-making abilities.
Therefore, this thesis emphasises selecting appropriate ML algorithms and
optimal sensor placements to address data limitations. The thesis also explores the use of transfer learning to enhance model performance and to
reuse trained models without relying on extensive datasets.
52
Jerol Soibam
Chapter 4. Results and discussion
Methods such as pGPR, CNN with transfer learning, and PINN were
adopted to rely on smaller datasets without deviating from the physics
or sacrificing model accuracy. In the case of pGPR, hypothetical data
points are considered rather than using the entire dataset during training
and prediction. This method allows for efficient data use while also indicating the prediction’s degree of confidence. With inverse-PINN, only a
few sensors’ data are needed to estimate the flow field and to predict the
temperature around the source wall. The PINN model was further coupled
with transfer learning to estimate the flow field in the enclosure domain
when the source shape and number of sources change. Transfer learning
was similarly used in the CNN architecture to track and segment bubbles in subcooled boiling flow. Results from Paper V suggest that transfer learning can significantly enhance model accuracy. Additionally, the
computational demand for training the algorithm can be notably reduced,
yielding the same or higher accuracy levels depending on the application.
For instance, in tracking the bubbles, the computational cost was halved
compared to training the model from scratch, while producing more reliable predictions.
RQ 4: How can uncertainty quantification contribute to enhancing the
robustness and reliability of machine learning models utilised to simulate
and predict heat transfer? (Papers III, II & VI)
ML models, particularly deep learning architectures, often function as
"black boxes", making their internal logic elusive. In fields like heat
transfer, where it’s essential to understand the underlying mechanisms,
such opacity poses challenges. This thesis emphasises the importance of
quantifying uncertainty within these models to enhance their predictive
capabilities, thereby offering a layer of transparency and deeper insight
into the model’s behaviour.
In the case study of subcooled boiling, two probabilistic models (MC
dropout and deep ensemble) were examined. The study revealed that probabilistic models outperform the deterministic DNN. For example, while
predicting the onset of nucleate boiling using a deterministic DNN model,
it showed an early nucleation site compared to the numerical value. In
contrast, predictions from the probabilistic model (DE) aligned more closely with the numerical value, signifying its robustness. Moreover, this
probabilistic model displayed a heightened uncertainty in specific regions,
possibly linked to the abrupt shift in physics from liquid to vapour (generation of many nucleation sites). Such indications of uncertainty empower
engineers to gauge the model’s reliability and understand its behaviour in
particular scenarios. In another instance, a pGPR model, rooted in the
Mälardalen University Press Dissertations
53
Machine Learning Techniques for Enhanced Heat Transfer Modelling
Bayesian approach, was employed to predict the CHF. The model managed to predict the CHF within a range of ±10%, while most empirical correlations deviated by about ±15%. Comparisons with experimental values
indicated minor deviations in the model’s predictions. Yet, a pervasive
uncertainty was observed across all predictions, likely due to the inherent
nonlinearity and experimental uncertainty present in the CHF dataset.
To address potential issues like overfitting and underfitting, an ensemblePINN was introduced, resulting in predictions that were more robust and
adaptable than those derived from a single-PINN. Assessing the uncertainty of a PINN is crucial for understanding its generalisability, especially
when adapting to changing flow dynamics or when the model is trained
with limited sensor data. The L2 relative error from the ensemble-PINN
consistently surpassed that of the single-PINN. Notably, the ensemblePINN maintained consistent predictions even when the dynamic changes
due to vortex shedding over time. Furthermore, the model’s uncertainty
accurately covered the range of actual numerical values. This ability to
quantify uncertainty brings transparency and supports informed decisionmaking. In such scenarios, gauging the potential outcomes’ range and
likelihood is as vital as the prediction itself.
Limitations
The incorporation of ML into heat transfer research, especially in areas
like subcooled boiling and convective heat transfer, has offered transformative insights. However, several challenges and limitations exist in the
methodologies and outcomes of the studies presented in this thesis.
The efficacy of these models is intrinsically linked to the quality, diversity, and availability of data. In the absence of adequate data, the models
may falter, particularly in depicting subcooled boiling heat transfer. An
in-depth understanding of subcooled boiling necessitates a thorough analysis of the number of input features. Furthermore, these models might
struggle to adapt to scenarios outside their training data.
In terms of bubble detection, while the current method is competent,
it has its hurdles. If the boiling process alters or produces larger bubbles
(bubbly flow), the algorithm needs revising. There’s also a pressing need
for better ways to manage and process the overlapping bubble images from
the mask produced by the CNN model. To ensure the model’s stability, a
detailed sensitivity analysis is crucial, especially when faced with varied
noise conditions and experiments.
Integrating sensor data, though advantageous, brings about uncertainties. Accurate sensor placement, underpinned by preliminary data, is crucial, especially when employing techniques like QR pivoting. Incorrect
54
Jerol Soibam
Chapter 4. Results and discussion
sensor placements or data collection errors can adversely impact predictions. Additionally, while the studies underscore the potential of techniques like transfer learning in situations with scarce data, the efficacy
of such techniques is still bound by the quality of the base model and
the similarities between the source and target domains. Other elements,
such as the enforcement of specific boundary conditions or the selection
of appropriate model hyperparameters, are decisive for the model’s accuracy. While the PINN model excels in portraying patterns like vortex
shedding and heat transfer mechanisms in the domain, certain aspects remain opaque, highlighting the need for further clarity. Finally, although
the models identify uncertainties, comprehensively capturing all types of
uncertainty, especially in noisy data scenarios, remains a priority.
Mälardalen University Press Dissertations
55
5
Conclusions and future work
This chapter presents the major conclusions of the thesis.
The collective conclusion in this thesis, drawn from the six papers,
highlights the successful application of deep learning techniques and statistical methods in improving the understanding, modelling, and prediction of various aspects of boiling and convective heat transfer processes.
The following overarching insights have been gleaned from this extensive
body of work:
• The implementation of supervised DNN models, yielded effective
results in simulating subcooled boiling heat transfer. Despite the
observed limitations in the extreme extrapolation of void fraction
fields, the model displayed an impressive level of accuracy, especially when predicting temperature fields. The comparison of deterministic and probabilistic deep learning models further illustrated
the advantages of the latter. Probabilistic models demonstrated superior performance in accurately predicting the quantities of interest, and in accounting for uncertainties in the model. The results
indicated the successful capture and reproduction of the underlying
physics of the boiling process, even under extreme extrapolation
conditions and addressed in RQ 1&4.
• The application of pGPR for predicting CHF resulted in improved
prediction accuracy compared to and empirical method and lookup table based approaches. By harnessing information available in
experimental data, the pGPR model provided insights into the sensitivity of CHF to various operating conditions. Furthermore, this
method demonstrated the little need for data to accurately predict
the CHF which addresses RQ 3&4.
• A CNN model was successfully used to understand local bubble dynamics and track bubbles even under less favourable imaging conditions. The model demonstrated robustness against noise and fluctuations and was adaptable to different flow regimes. In summary,
the CNN model employed in this study demonstrates the considerable potential for understanding bubble dynamics in boiling processes and can serve as a robust foundation for future research in
Mälardalen University Press Dissertations
57
Machine Learning Techniques for Enhanced Heat Transfer Modelling
this field. Moreover, the model’s flexibility to adapt to different flow
regimes through transfer learning enhances its applicability and this
addresses RQ 1&3.
• In response to RQ 2&3, the use of inverse PINN has proven to
be a crucial tool in reconstructing the unknown thermal boundary
conditions. These networks successfully reconstruct thermal distributions and flow fields across a variety of conditions and shapes.
By applying transfer learning techniques to PINNs, it was able to
enhance the computational efficiency while retaining an acceptable
accuracy. This approach not only provided insights into the complex thermal dynamics but also opened up opportunities for optimisation and predictive modelling in heat transfer applications.
• Finally, the effectiveness of PINNs was further emphasised through
the study of mixed convection flows. Utilising a QR pivoting strategy for optimal sensor placement significantly improved prediction
accuracy, with ensemble models outperforming single PINN model
despite higher computational requirements. The ensemble models
also introduced the important aspect of predictive uncertainty quantification and addressed RQ 2,3, &4.
In conclusion, the findings from these studies strongly suggest that
ML methods, specifically deep neural networks (fully connected / convolutional), Gaussian process regression, and physics-informed neural networks, provide promising tools for capturing the intricate physics involved,
predicting quantities of interest accurately, and accounting for the inherent
uncertainties.
Future work
Based on the findings and results obtained from this thesis, some relevant
future work are summarised below:
• To investigate deeper the quality and impact of input features specific to subcooled boiling heat transfer. Further, explore how these
models are influenced by data availability and refine them by integrating known physics principles.
• Refine the existing CNN model for bubble segmentation to accommodate a wider range of experimental and flow conditions. To delve
further into the dynamics of subcooled boiling to establish correlations that can be instrumental for numerical simulations.
58
Jerol Soibam
Chapter 5. Conclusions and future work
• To integrate aleatoric uncertainty into the models to enhance their
understanding and adaptability in managing noisy data, thereby establishing a comprehensive and robust uncertainty framework.
• Integration of QR sensor with energy residuals to better optimise
sensor locations, while iteratively embedding this within the PINN
framework. Furthermore, to incorporate sensor noise to assess the
model’s sensitivity.
• Expand the existing methodology of PINN to 3D transient simulations and investigate its applicability in engineering systems of
HVDC electronic cooling.
Mälardalen University Press Dissertations
59
Bibliography
Alayil, R. and Balaji, C. (2015). Conjugate heat transfer in latent heat
thermal storage system with cross plate fins. Journal of Heat Transfer,
137(10):102302.
Albawi, S., Mohammed, T. A., and Al-Zawi, S. (2017). Understanding of
a convolutional neural network. In 2017 International Conference on
Engineering and Technology (ICET), pages 1–6.
Alifanov, O. M. (2012). Inverse heat transfer problems. Springer Science
& Business Media.
Ardabili, S., Mosavi, A., and Felde, I. (2023). Machine learning in heat
transfer: Taxonomy, review and evaluation. In 2023 IEEE 17th International Symposium on Applied Computational Intelligence and Informatics (SACI), pages 000433–000442. IEEE.
Baraldi, P. and Mangili, Fand Zio, E. (2015). A prognostics approach
to nuclear component degradation modeling based on gaussian process
regression. Progress in Nuclear Energy, 78:141–154.
Benard, N., Pons-Prats, J., Periaux, J., Bugeda, G., Braud, P., Bonnet, J.,
and Moreau, E. (2016). Turbulent separated shear flow control by surface plasma actuator: Experimental optimization by genetic algorithm
approach. Experiments in Fluids, 57(2):22.
Bhamare, D. K., Saikia, P., Rathod, M. K., Rakshit, D., and Banerjee, J.
(2021). A machine learning and deep learning based approach to predict
the thermal performance of phase change material integrated building
envelope. Building and Environment, 199:107927.
Bolya, D., Zhou, C., Xiao, F., and Lee, Y. J. (2022). Yolact++ better realtime instance segmentation. IEEE Transactions on Pattern Analysis and
Machine Intelligence, 44(2):1108–1121.
Breiman, L. (1996). Bagging predictors. Machine learning, 24(2):123–
140.
Brunton, S. L. and Noack, B. R. (2015). Closed-loop turbulence control:
Progress and challenges. Applied Mechanics Reviews, 67(5).
Mälardalen University Press Dissertations
61
Brunton, S. L., Noack, B. R., and Koumoutsakos, P. (2020). Machine
learning for fluid mechanics. Annual Review of Fluid Mechanics,
52:477–508.
Brunton, S. L., Proctor, J. L., and Kutz, J. N. (2016). Discovering governing equations from data by sparse identification of nonlinear dynamical systems. Proceedings of the national academy of sciences,
113(15):3932–3937.
Cai, S., Wang, Z., Fuest, F., Jeon, Y. J., Gray, C., and Karniadakis, G. E.
(2021a). Flow over an espresso cup: inferring 3-d velocity and pressure
fields from tomographic background oriented schlieren via physicsinformed neural networks. Journal of Fluid Mechanics, 915.
Cai, S., Wang, Z., Wang, S., Perdikaris, P., and Karniadakis, G. E.
(2021b). Physics-informed neural networks for heat transfer problems.
Journal of Heat Transfer, 143(6).
Celik, N., Kurtbas, I., Yumusak, N., and Eren, H. (2009). Statistical regression and artificial neural network analyses of impinging jet experiments. Heat and mass transfer, 45(5):599–611.
Colaco, M. J., Orlande, H. R., and Dulikravich, G. S. (2006). Inverse and
optimization problems in heat transfer. Journal of the Brazilian Society
of Mechanical Sciences and Engineering, 28:1–24.
Colvert, B., Alsalman, M., and Kanso, E. (2018). Classifying vortex wakes using neural networks. Bioinspiration & biomimetics,
13(2):025003.
Deng, J., Dong, W., Socher, R., Li, L.-J., Li, K., and Fei-Fei, L. (2009).
Imagenet: A large-scale hierarchical image database. In 2009 IEEE
Conference on Computer Vision and Pattern Recognition, pages 248–
255.
Di Natale, L., Svetozarevic, B., Heer, P., and Jones, C. (2021). Deep reinforcement learning for room temperature control: a black-box pipeline
from data to policies. In Journal of Physics: Conference Series, volume
2042, page 012004. IOP Publishing.
Dracopoulos, D. C. (2013). Evolutionary learning algorithms for neural
adaptive control. Springer.
Drikakis, D. (2003). Advances in turbulent flow computations using highresolution methods. Progress in Aerospace Sciences, 39(6-7):405–424.
62
Jerol Soibam
Duraisamy, K., Iaccarino, G., and Xiao, H. (2019). Turbulence modeling
in the age of data. Annual Review of Fluid Mechanics, 51:357–377.
Fleming, P. J. and Purshouse, R. C. (2002). Evolutionary algorithms in
control systems engineering: a survey. Control engineering practice,
10(11):1223–1241.
Fraigneau, Y. (2019). 2d flow around a heated square-cylinder, re=50,
ra=5×106 . Available at: https://datasetmeca.lisn.upsaclay.fr. LISN
Université Paris-Saclay, Accessed: 2023-03-31.
Fu, Y. and Liu, Y. (2019). Bubgan: Bubble generative adversarial networks for synthesizing realistic bubbly flow images. Chemical Engineering Science, 204:35–47.
Fukami, K., Nabae, Y., Kawai, K., and Fukagata, K. (2019). Synthetic
turbulent inflow generator using machine learning. Physical Review
Fluids, 4(6):064603.
Gal, Y. and Ghahramani, Z. (2016). Dropout as a bayesian approximation: Representing model uncertainty in deep learning. In international
conference on machine learning, pages 1050–1059. ICML’16.
Geng, Z. and Wang, Y. (2020). Automated design of a convolutional neural network with multi-scale filters for cost-efficient seismic data classification. Nature communications, 11(1):1–11.
Goodfellow, I., Bengio, Y., Courville, A., and Bengio, Y. (2016). Deep
learning, volume 1. MIT press Cambridge.
Haas, T., Schubert, C., Eickhoff, M., and Pfeifer, H. (2020). Bubcnn: Bubble detection using faster rcnn and shape regression network. Chemical
Engineering Science, 216:115467.
He, K., Gkioxari, G., Dollár, P., and Girshick, R. (2017). Mask r-cnn.
In 2017 IEEE International Conference on Computer Vision (ICCV),
pages 2980–2988.
Hobold, G. M. and da Silva, A. K. (2019a). Automatic detection of the
onset of film boiling using convolutional neural networks and bayesian
statistics. International Journal of Heat and Mass Transfer, 134:262–
270.
Hobold, G. M. and da Silva, A. K. (2019b). Visualization-based nucleate
boiling heat flux quantification using machine learning. International
Journal of Heat and Mass Transfer, 134:511–520.
Mälardalen University Press Dissertations
63
Hughes, M. T., Kini, G., and Garimella, S. (2021). Status, challenges,
and potential for machine learning in understanding and applying heat
transfer phenomena. Journal of Heat Transfer, 143(12):120802.
Iten, R., Metger, T., Wilming, H., Del Rio, L., and Renner, R. (2020).
Discovering physical concepts with neural networks. Physical review
letters, 124(1):010508.
Jambunathan, K., Hartle, S., Ashforth-Frost, S., and Fontama, V. (1996).
Evaluating convective heat transfer coefficients using neural networks.
International Journal of Heat and Mass Transfer, 39(11):2329–2332.
Jiang, B., Ren, J., Hu, P., and Zhao, F. (2013). Prediction of critical heat
flux for water flow in vertical round tubes using support vector regression model. Progress in Nuclear Energy, 68:210–222.
Jiang, B., Zhou, J., Huang, X., and Wang, P. (2020). Prediction of critical
heat flux using gaussian process regression and ant colony optimization.
Annals of Nuclear Energy, 149:107765.
Jiménez, J. and Moin, P. (1991). The minimal flow unit in near-wall turbulence. Journal of Fluid Mechanics, 225:213–240.
Karniadakis, G. E., Kevrekidis, I. G., Lu, L., Perdikaris, P., Wang, S., and
Yang, L. (2021). Physics-informed machine learning. Nature Reviews
Physics, 3(6):422–440.
Kim, H. C., Baek, W., and Chang, S. H. (2000). Critical heat flux of water
in vertical round tubes at low pressure and low flow conditions. Nuclear
Engineering and design, 199(1-2):49–73.
Kim, Y. and Park, H. (2021). Deep learning-based automated and universal bubble detection and mask extraction in complex two-phase flows.
Scientific reports, 11(1):1–11, 8940.
Kingma, D. P. and Ba, J. (2014). Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980.
Krizhevsky, A., Sutskever, I., and Hinton, G. E. (2012). Imagenet classification with deep convolutional neural networks. Advances in neural
information processing systems, 25:1097–1105.
Lagaris, I. E., Likas, A., and Fotiadis, D. I. (1998). Artificial neural
networks for solving ordinary and partial differential equations. IEEE
transactions on neural networks, 9(5):987–1000.
64
Jerol Soibam
Lakshminarayanan, B., Pritzel, A., and Blundell, C. (2017). Simple and
scalable predictive uncertainty estimation using deep ensembles. In
Advances in neural information processing systems, pages 6402–6413.
Lee, C., Kim, J., Babcock, D., and Goodman, R. (1997). Application of
neural networks to turbulence control for drag reduction. Physics of
Fluids, 9(6):1740–1747.
Lee, H. and Kang, I. S. (1990). Neural algorithm for solving differential
equations. Journal of Computational Physics, 91(1):110–131.
Ling, J., Jones, R., and Templeton, J. (2016a). Machine learning strategies for systems with invariance properties. Journal of Computational
Physics, 318:22–35.
Ling, J., Kurzawski, A., and Templeton, J. (2016b). Reynolds averaged
turbulence modelling using deep neural networks with embedded invariance. Journal of Fluid Mechanics, 807:155–166.
Ling, J. and Templeton, J. (2015). Evaluation of machine learning algorithms for prediction of regions of high reynolds averaged navier stokes
uncertainty. Physics of Fluids, 27(8):085103.
Lipton, Z. C. (2018). The mythos of model interpretability: In machine
learning, the concept of interpretability is both important and slippery.
Queue, 16(3):31–57.
Lucor, D., Agrawal, A., and Sergent, A. (2022). Simple computational
strategies for more effective physics-informed neural networks modeling of turbulent natural convection. Journal of Computational Physics,
456:111022.
Lumley, J. (1970). Stochastic tools in turbulence, academic press, new
york.
Maulik, R., San, O., Rasheed, A., and Vedula, P. (2019). Subgrid modelling for two-dimensional turbulence using neural networks. Journal
of Fluid Mechanics, 858:122–144.
Milano, M. and Koumoutsakos, P. (2002). Neural network modeling for
near wall turbulent flow. Journal of Computational Physics, 182(1):1–
26.
Mnih, V., Kavukcuoglu, K., Silver, D., Rusu, A. A., Veness, J., Bellemare,
M. G., Graves, A., Riedmiller, M., Fidjeland, A. K., Ostrovski, G.,
et al. (2015). Human-level control through deep reinforcement learning.
nature, 518(7540):529–533.
Mälardalen University Press Dissertations
65
Molland, A. F. and Turnock, S. R. (2007). 6 - theoretical and numerical
methods. In Molland, A. F. and Turnock, S. R., editors, Marine Rudders and Control Surfaces, pages 233–311. Butterworth-Heinemann,
Oxford.
Murata, T., Fukami, K., and Fukagata, K. (2020). Nonlinear mode decomposition with convolutional neural networks for fluid dynamics. Journal
of Fluid Mechanics, 882.
Otter, D. W., Medina, J. R., and Kalita, J. K. (2021). A survey of the
usages of deep learning for natural language processing. IEEE Transactions on Neural Networks and Learning Systems, 32(2):604–624.
Parish, E. J. and Duraisamy, K. (2016). A paradigm for data-driven predictive modeling using field inversion and machine learning. Journal of
Computational Physics, 305:758–774.
PHAN, M. Q., JUANG, J.-N., and HYLAND, D. C. (1995). On neural
networks in identification and control of dynamic systems. In Wave
Motion, Intelligent Structures And Nonlinear Mechanics, pages 194–
225. World Scientific.
Piatek,
˛
W., Oleksiak, A., and Da Costa, G. (2015). Energy and thermal
models for simulation of workload and resource management in computing systems. Simulation Modelling Practice and Theory, 58:40–54.
Podvin, B., Fraigneau, Y., Jouanguy, J., and Laval, J.-P. (2010). On selfsimilarity in the inner wall layer of a turbulent channel flow. Journal of
fluids engineering, 132(4).
Pollard, A., Castillo, L., Danaila, L., and Glauser, M. (2016). Whither
turbulence and big data in the 21st century? Springer.
Proctor, J. L., Brunton, S. L., and Kutz, J. N. (2016). Dynamic mode
decomposition with control. SIAM Journal on Applied Dynamical Systems, 15(1):142–161.
Rabault, J., Kuchta, M., Jensen, A., Réglade, U., and Cerardi, N. (2019).
Artificial neural networks trained through deep reinforcement learning
discover control strategies for active flow control. Journal of fluid mechanics, 865:281–302.
Rabault, J. and Kuhnle, A. (2019). Accelerating deep reinforcement learning strategies of flow control through a multi-environment approach.
Physics of Fluids, 31(9):094105.
66
Jerol Soibam
Rabhi, A., Aslanidou, I., Kyprianidis, K., and Fdhila, R. B. (2021). Onset of nucleate boiling model for rectangular upward narrow channel:
Cfd based approach. International Journal of Heat and Mass Transfer,
165:120715.
Raissi, M., Babaee, H., and Karniadakis, G. E. (2019a). Parametric
gaussian process regression for big data. Computational Mechanics,
64(2):409–416.
Raissi, M., Perdikaris, P., and Karniadakis, G. E. (2019b). Physicsinformed neural networks: A deep learning framework for solving forward and inverse problems involving nonlinear partial differential equations. Journal of Computational Physics, 378:686–707.
Raissi, M., Wang, Z., Triantafyllou, M. S., and Karniadakis, G. E. (2019c).
Deep learning of vortex-induced vibrations. Journal of Fluid Mechanics, 861:119–137.
Raissi, M., Yazdani, A., and Karniadakis, G. E. (2020). Hidden fluid mechanics: Learning velocity and pressure fields from flow visualizations.
Science, 367(6481):1026–1030.
Rassoulinejad-Mousavi, S. M., Al-Hindawi, F., Soori, T., Rokoni, A.,
Yoon, H., Hu, H., Wu, T., and Sun, Y. (2021). Deep learning strategies for critical heat flux detection in pool boiling. Applied Thermal
Engineering, 190:116849.
Reddy, G., Celani, A., Sejnowski, T. J., and Vergassola, M. (2016). Learning to soar in turbulent environments. Proceedings of the National
Academy of Sciences, 113(33):E4877–E4884.
Redmon, J., Divvala, S., Girshick, R., and Farhadi, A. (2016). You only
look once: Unified, real-time object detection. In 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages
779–788, Las Vegas, NV, USA.
Rowley, C. W. and Dawson, S. T. (2017). Model reduction for flow analysis and control. Annual Review of Fluid Mechanics, 49:387–417.
Rowley, C. W., MEZI?, I., Bagheri, S., Schlatter, P., Henningson, D., et al.
(2009). Spectral analysis of nonlinear flows. Journal of fluid mechanics,
641(1):115–127.
San, O. and Maulik, R. (2018). Machine learning closures for model order
reduction of thermal fluids. Applied Mathematical Modelling, 60:681–
710.
Mälardalen University Press Dissertations
67
Sarghini, F., De Felice, G., and Santini, S. (2003). Neural networks based
subgrid scale modeling in large eddy simulations. Computers & fluids,
32(1):97–108.
Scheiff, V., Aslanidou, I., Kyprianidis, K., and Bel Fdhila, R. (2023). Experimental study of nucleate boiling dynamics in a rectangular minichannel set-up. In 8th Thermal and Fluids Engineering Conference
(TFEC).
Schmidt, M. and Lipson, H. (2009). Distilling free-form natural laws from
experimental data. science, 324(5923):81–85.
Serra, P. L., Masotti, P. H., Rocha, M. S., de Andrade, D. A., Torres,
W. M., and de Mesquita, R. N. (2020). Two-phase flow void fraction estimation based on bubble image segmentation using randomized
hough transform with neural network (rhtn). Progress in Nuclear Energy, 118:103133.
Silver, D., Hubert, T., Schrittwieser, J., Antonoglou, I., Lai, M., Guez,
A., Lanctot, M., Sifre, L., Kumaran, D., Graepel, T., et al. (2018). A
general reinforcement learning algorithm that masters chess, shogi, and
go through self-play. Science, 362(6419):1140–1144.
Singh, A. P., Medida, S., and Duraisamy, K. (2017). Machine-learningaugmented predictive modeling of turbulent separated flows over airfoils. AIAA journal, 55(7):2215–2227.
Sinha, K. N. R., Kumar, V., Kumar, N., Thakur, A., and Raj, R. (2021).
Deep learning the sound of boiling for advance prediction of boiling
crisis. Cell Reports Physical Science, 2(3):100382.
Suh, Y., Bostanabad, R., and Won, Y. (2021). Deep learning predicts
boiling heat transfer. Scientific reports, 11, 5622(1):1–10.
Sun, L., Gao, H., Pan, S., and Wang, J.-X. (2020). Surrogate modeling for
fluid flows based on physics-constrained deep learning without simulation data. Computer Methods in Applied Mechanics and Engineering,
361:112732.
Taira, K., Brunton, S. L., Dawson, S. T., Rowley, C. W., Colonius, T.,
McKeon, B. J., Schmidt, O. T., Gordeyev, S., Theofilis, V., and Ukeiley,
L. S. (2017). Modal analysis of fluid flows: An overview. AIAA Journal,
55(12):4013–4041.
68
Jerol Soibam
Taira, K., Hemati, M. S., Brunton, S. L., Sun, Y., Duraisamy, K., Bagheri,
S., Dawson, S. T., and Yeh, C.-A. (2020). Modal analysis of fluid flows:
Applications and outlook. AIAA journal, 58(3):998–1022.
Traverso, T., Coletti, F., Magri, L., Karayiannis, T. G., and Matar, O. K.
(2023). A machine learning approach to the prediction of heat-transfer
coefficients in micro-channels. arXiv preprint arXiv:2305.18406.
Wang, C.-Y., Bochkovskiy, A., and Liao, H.-Y. M. (2022). Yolov7: Trainable bag-of-freebies sets new state-of-the-art for real-time object detectors. arXiv preprint arXiv:2207.02696.
Wang, Q., Li, X., Xu, C., Yan, T., and Li, Y. (2021). Bubble recognizing and tracking in a plate heat exchanger by using image processing
and convolutional neural network. International Journal of Multiphase
Flow, 138:103593.
Yang, L., Zhang, D., and Karniadakis, G. E. (2020). Physics-informed
generative adversarial networks for stochastic differential equations.
SIAM Journal on Scientific Computing, 42(1):A292–A317.
Zhang, D., Lu, L., Guo, L., and Karniadakis, G. E. (2019). Quantifying total uncertainty in physics-informed neural networks for solving
forward and inverse stochastic problems. Journal of Computational
Physics, 397:108850.
Zhu, Y., Zabaras, N., Koutsourelakis, P.-S., and Perdikaris, P. (2019).
Physics-constrained deep learning for high-dimensional surrogate modeling and uncertainty quantification without labeled data. Journal of
Computational Physics, 394:56–81.
Mälardalen University Press Dissertations
69
Download