A physics-driven and machine learning-based digital twinning approach to transient thermal systems

Purpose – In this study, the authors propose a novel digital twinning approach speci ﬁ cally designed for controlling transient thermal systems. The purpose of this study is to harness the combined power of deep learning (DL) and physics-based methods (PBM) to create an active virtual replica of the physical system. Design/methodology/approach – To achieve this goal, we introduce a deep neural network (DNN) as the digital twin and a Finite Element (FE) model as the physical system. This integrated approach is used to address the challenges of controlling an unsteady heat transfer problem with an integrated feedback loop. Findings – The results of our study demonstrate the effectiveness of the proposed digital twinning approach in regulating the maximum temperature within the system under varying and unsteady heat ﬂ ux conditions. The DNN, trained on stationary data, plays a crucial role in determining the heat transfer coef ﬁ cients necessary to maintain temperatures below a de ﬁ ned threshold value, such as the material ’ s melting point. The system is successfully controlled in 1D, 2D and 3D case studies. However, careful evaluations should be conducted if such a training approach, based on steady-state data, is applied to completelydifferent transient heat transfer problems. Originality/value – The present work represents one of the ﬁ rst examples of a comprehensive digital twinning approach to transient thermal systems, driven by data. One of the noteworthy features of this approach is its robustness. Adopting a training based on dimensionless data, the approach can seamlessly accommodate changes in thermal capacity and thermal conductivity without the need for retraining.


Introduction
In our rapidly evolving technological landscape, the field of artificial intelligence (AI) has gained paramount significance.AI is a discipline focused on imparting machines with pseudo-cognitive capacities to mimic human intelligence.Within this vast domain, machine learning (ML) is a sub-set concerned with the development of algorithms that allow systems to learn and adapt from data without explicit programming.Deep learning (DL), in turn, stands as a specialized branch of ML, characterized by the utilization of deep neural networks (DNNs) to enable high-level pattern recognition and complex feature extraction (Ongsulee, 2017).Alongside these advancements, the concept of the digital twin (DT) has emerged as a revolutionary framework that may amalgamate these technologies to offer a real-time digital copy of a physical entity or system.The DT concept embodies various definitions, each tailored to its application domain.However, a common thread among these definitions is that the DT is a digital mapping of the physical world aimed at facilitating analysis, prediction, and especially control of the physical world itself (Yao et al., 2023).From a theoretical perspective, a DT can exist independently of the utilization of AI and its sub-sets.A DT can be based on CAD (computer-aided design) models, CAE (computeraided engineering) simulations and the like.However, a challenge arises due to the fact that traditional 2D/3D simulations are still distant from being capable of exchanging real-time data with the physical system.To address this issue, AI, especially ML and DL, becomes a HFF highly useful tool for constructing predictive reduced models of the real system (Xuereb Conti et al., 2023;Grabe et al., 2023), optimization purposes (Lin et al., 2018;Guo et al., 2023;Keramati et al., 2022;Aldaghi et al., 2023;Zeeshan et al., 2023), solving inverse problems (Tamaddon-Jahromi et al., 2020;Löhner et al., 2020) and effectively mitigating the computational costs associated with conventional simulations (Zhu et al., 2023;Park and Kim, 2023;Zhang et al., 2023).It must be specified that the standalone CAE simulations do not represent a DT.Yao et al. (2023), in their review work, highlight the main difference between CAE simulations and DTs.Quoting from their work: "Simulation technologies partially reproduce the real world offline, primarily during the research and design stages.They usually do not perform analyses or optimization functions.However, DT reflects the state changes of physical objects in real-time and can be used to analyse and predict the decision optimization function of physical entities.Simulation technology relies on models and data to map the properties and parameters of the physical world.DT must sense, diagnose, and predict the state of physical entities in real time to optimize them".
DNNs, a cornerstone of DL, find relevance in the creation of DTs by enabling the synthesis of intricate models capable of replicating complex behaviours.Supervised learning (SL), a sub-set of ML, contributes to training these networks, establishing a connection between AI and DTs.Furthermore, the realm of deep reinforcement learning (DRL) extends this interconnection, where AI agents learn to make decisions through interactions with their environment, mirroring the way control strategies can be developed and optimized within the DT framework.
The history of DT traces back several decades, with roots in aerospace engineering and industrial automation.Historically, the first DT was created by NASA in 1970 during Apollo 13 programme to train personnel and troubleshoot for space exploration (Ferko et al., 2022).Over time, the application domains of DTs have expanded, encompassing sectors such as health care (Liu et al., 2019;Pizzolato et al., 2019;Chakshu et al., 2019Chakshu et al., , 2021;;Chakshu and Nithiarasu, 2022;Tamaddon Jahromi et al., 2022), energy Zhou et al. (2020);Francisco et al. (2020), civil engineering (Kumar et al., 2018;Shim et al., 2019) and manufacturing (Abadías Llamas et al., 2020;Gupta and Basu, 2019).In the area of additive manufacturing, very recent research articles and reviews show that the state of the art is still focusing on reduced order modelling.The obvious final aim is to control, in real-time, the 3D printing process (Afazov et al., 2022;Wei et al., 2021;Li and Polydorides, 2023;Montoya-Zapata et al., 2021).
In the realm of heat transfer, the DT is not always intended as a compound system where there is a continuous data exchange between the physical system and its digital copy, according to the definition reported by Yao et al. (2023).Below, we provide a more specific literature survey of DT in the heat transfer field to support the said claim.We provide examples of actual DTs and others that can be categorized as "modelling" and not DT according to the definition used so far.Yang et al. (2023) propose an actual DT to optimize the secondary cooling and final electromagnetic stirring in a continuous casting process.They do that via an offline and then online calibration of a 2D heat transfer model (of an equivalent solid medium) coupled to a neural network (NN) to enhance the accuracy of the (stationary) boundary conditions.In this case, the final aim is to reduce the macro-segregation phenomena in the semi-finished product.
In the work by Nget et al. (2023), a multi-physics model of microwave cooking is presented.An optimization to maximize the uniformity of the temperature field inside food is planned as future development in a DT, but no more details are described.Another example of DT is instead shown by Leeming et al. (2023).They delve into a DT to effectively Digital twinning approach to transient thermal systems control the supersaturation in batch crystallization.This is an example where no AI/ML/DL algorithm is exploited because the DT is only based on a one-dimensional population balance model.Chen et al. (2023) build a thermodynamic model and an algorithm for solving temperatures in a plate heat exchanger with variable fluid properties.Even though the authors claim to deal with a DT, no strategy allows the digital copy to control, for example, the outlet temperature of the heat exchanger.Similarly, Spateri et al. (2023)'s work presented a validated simplified electro-thermal model to manage thermoforming.Even if the model potentially fits well for its very low computational costs (it is a zero-dimensional model) for a real-time implementation in a DT, no real-time control/optimization is proposed.Cui et al. (2023) reproduce a DT of a supercritical coal-fired boiler for detecting temperature anomalies in advance.The DT is fully driven by experimental data coming from temperature sensors and it is based on Particle Swarm Optimization-XGBoost machine learning algorithms.Hachem et al. (2021) developed a DRL model to minimize the Nusselt number in 2D/3D fluid-dynamics cases.For example, in a square cavity natural convection case, they vary the distribution of the hot temperature boundary condition keeping the cold one at a constant value.Although the DRL is intrinsically close to a DT concept, the exchange of information between the CFD model (considered as the physical world, the environment in the DRL language) and the NN occurs only during the learning phase.The model is stationary and is aimed at finding the temperature distribution to be applied as a boundary condition in order to minimize the Nusselt number.Similarly, Beintema et al. (2020) focus on the same case study, comparing two different control strategies.The first one is the linear-proportional controller, and the second one is based on a DRL model.The results show that the second one is able to achieve a lower Nusselt number.Another similar example of DRLbased heat transfer control is provided by Renault et al. (2023).The latter three works (Hachem et al., 2021;Beintema et al., 2020;Renault et al., 2023) share the same learning algorithm: the single step proximal-policy-optimization.In these cases, also known as open-loop control problems, the policy optimization does not depend on the environment response.On the other hand, a closed-loop control of a heat transfer problem is realized by Wang et al. (2023).They confirm the DRL-based control strategy, made by an oscillatory flow rate, is the best one to minimize the maximum temperature in the system.Furthermore, even though the authors never mention it, this work may be considered a DT as the NN actions rely on the current state of the environment (CFD simulation).
The above literature review illustrates DT in heat transfer is still a new and emerging research area compared to other sectors, where it is more mature.It emerges that sometimes the concept of DT does not correspond to that one proposed by Yao et al. (2023).The other articles, that effectively develop a DT according to the general definition given in this article, build DTs fully driven by experimental data or DTs that deal with stationary or pseudostationary heat transfer systems.
The main objective and main novelty over already published papers of this work is to propose a digital twinning of an intrinsically unsteady heat transfer system in a solid medium.More specifically, this work aims at addressing the following question.Given a thermal system with a known transient inward heat flux boundary condition, what is the value of cooling (in terms of heat transfer coefficient boundary condition) we should apply to dynamically control the maximum temperature below a threshold temperature value?In cases where a physical system is not available to provide experimental data, a numerical model can serve as a surrogate system.As demonstrated in a previous work HFF (Samaniego et al. (2020)), this surrogate system can also be a DL model that represents the governing equations of the considered physics, eliminating the need for numerical discretization.However, we assume a finite element method (FEM) model to be our surrogate physical system, while a DNN the digital copy of the system.The DNN model acts as a DT and provides the heat transfer coefficient to meet our goal.Note that this clearly does not represent a reduced order optimization model of the FEM simulation or a simple Inverse Thermal Problem (ITP) aimed at determining a heat transfer coefficient.We couple them, guaranteeing a continuous exchange of data to control the maximum temperature of the system.We adopt a training based on steady-state data to reach our goal.We adopt a training based on steady-state data to reach our goal.Although the results shown later are satisfactory, such results cannot be guaranteed for completely different transient thermal systems.
The rest of the article is organized as follows.Section 2 introduces the DT of the proposed thermal system, Section 3 demonstrates the proposed digital twinning through a number of examples and Section 4 draws some conclusions (Table 1).
2. Digital twin of a thermal system 2.1 The finite element method modelthe surrogate physical system: governing equation and dimensionless form Without a physical system ready to produce continuous experimental data, a FEM model is assumed to be our surrogate physical system.Figure 1(a) shows the basic thermal problem chosen to demonstrate the proposed DT.An unsteady heat flux boundary condition q(t) is applied on the top surface E3.On the opposite bottom boundary, a convective heat transfer boundary condition h c t ð Þ is assumed with an unknown heat transfer coefficient (E1).It has to be determined by the DT.The left E4 and right E2 vertical surfaces are considered adiabatic.In this case, the problem is reduced to a one-dimensional model.A two-dimensional setup is built by switching the right adiabatic boundary condition with the heat flux condition on the top and vice-versa (see "Step 2. DL model training -Python", left).Finally, a 3D problem is constructed by adopting an isothermal boundary condition on F6 as shown in Figure 1(b)  and "Step 2. DL model training -Python" (right).The maximum element size used in the calculations is 0:1L where L is a characteristic length in the problems of interest.Secondorder polynomial functions for discretization, as well as a second-order implicit time scheme with an adaptive time step, are used to carry out all the numerical simulations.The governing equation is the heat conduction in a solid medium, which reads as follows: where k is the thermal conductivity of the solid medium, rc its thermal capacity, and their ratio a is the thermal diffusivity.The Neumann and Robin boundary conditions representing the inward heat flux and the cooling convective boundary are described by the following equations: where C q is the heat flux boundary, C c is the convective boundary and n is the unit normal vector.A dimensionless form of the transient thermal system is provided below.
The physical (dimensional) variables are: density r, specific heat c, temperature T, time t, thermal conductivity k, heat flux q, heat transfer coefficient h c and length L. According to the Buckingham theorem (Misic et al., 2010), after a careful evaluation of the relevant physical parameters involved in the problem, as well as the number of fundamental units (kg; m; s; K), we can build the minimum number of dimensionless groups required to describe the unsteady thermal system.The non-dimensional form of the heat equation is: coordinateÞ; Non-dimensional quantities of the Neumann boundary conditions are: T i and T 0 are the initial and the atmospheric temperatures, respectively.The dimensionless time, also known as Fourier number (Fo), takes into account the diffusion and thermal inertia phenomena.In the present work, Fo is varied between 0 and 150.The dimensionless heat flux boundary condition (q Ã ) is assumed to vary between 0 and 100.Similar to Fo, the Biot number (Bi) is generally introduced in unsteady heat transfer problems to make a convective type boundary condition dimensionless.It is varied between 1 and 15 in the FEM-based data set generation, while it is an unknown (determined by the trained NN) in the full DT as we will see later.Along with these limits, the dimensionless melting temperature is fixed at 100.The FEM model has two different goals.They are: (1) Generating (stationary) data to train the NN (DL).
Digital twinning approach to transient thermal systems (2) Testing the DT (trained DL model) to establish its accuracy in predicting the correct heat transfer coefficient (Biot number) for a given threshold temperature and transient heat flux.

The deep learning model
As mentioned above, the objective of the DL model is to solve the ITP and obtain the dynamic heat transfer coefficient for a given time-dependent heat flux and threshold temperature.We trained in total three different NNs for 1 D, 2 D and 3 D case studies, respectively.The first two DL models have two inputs (the maximum temperature calculated over the domain and the heat flux applied as a boundary condition) and one output (the heat transfer coefficient).The third one has an additional input as the temperature of the isothermal boundary condition is also varied.The first two DL models have three hidden layers with 32, 8 and 4 neurons respectively while the NN for the 3D cases has an intermediate hidden layer with 16 neurons.The activation function for all inner layers is the hyperbolic tangent (tanh) function.Given that the heat transfer coefficient has no upper bound, a rectified linear unit function ("ReLu") is chosen as output activation function.No particular strategy is used to standardize/normalize the input and output data.The loss function is the classical mean square error (MSE) function.A maximum of 2,000 epochs are used to train the DL models, while 1,064 is the batch size.Adam is the chosen optimizer due to its wide range of applicability, little memory requirements and computational efficiency (Kingma and Ba, 2014).Furthermore, it has been used in other heat transfer DL models as can be seen in the work of Tamaddon-Jahromi et al. (2020).To prevent classical problems of ML algorithms such as overfitting/underfitting and local minima, a sensitivity analysis on hyperparameters was carried out to test the robustness of the models.Furthermore, a K-fold cross-validation is implemented, partitioning the data into four subsets (or folds).In each iteration, one sub-set is used for validation, and the remaining three are used for training.The process repeats such that each sub-set gets used once as the validation set.The reliability of the model is confirmed by numerical values of root mean square error (RMSE) of both training and validation data sets and their average (for each fold) that stand at the same order of magnitude (see Table 2).

Finite element method -deep learning coupling (digital twinning)
Assuming we have already trained, validated, and tested the DL (see Section 3), its coupling with the real system (FEM) is essential to control the maximum temperature in real time.
The coupling between the FEM (the physical system) and DL is illustrated in Figure 2.
Figure 2(a), schematically represents the DT with continuous exchange of data with the FEM model.The DT (the trained DL model) receives time-dependent heat flux and the maximum allowed temperature data and returns a time-dependent heat transfer coefficient to the FEM model.In Figure 2(b) the proposed coupling strategy is presented: the total physical time of the simulation is divided into N intervals.In the last step of k-th interval, the DT is invoked to determine the heat transfer coefficient to be used in the interval k þ 1 of the FEM simulation.The inputs of the DT are the maximum allowed temperature, reduced by a constant safety factor and the maximum applied heat flux in the k-th interval N k .It is important to remember that, in this article, we train the DL with steady-state data.On the other hand, the real system for the DT is unsteady by definition.Nevertheless, steady-state data can comprehensively establish an effective relationship between heat transfer coefficient, heat flux and maximum temperature.As the heat transfer coefficient (Biot number) calculated by the DT is based on the maximum heat flux recorded during the time HFF interval, it is higher than what would be produced by a DT trained on transient data (see Appendix 1).As confirmed by the results in the next section, the proposed approach works effectively.A safety factor is applied to the threshold temperature to take into account the fact that the heat transfer coefficient is not updated instantaneously with heat flux and maximum temperature variations.A further analysis shown in Appendix 1 aims to clarify that the correlation between average heat flux, maximum temperature, and Biot number is the same in steady and transient simulations.The adopted strategy based on steady-state data allows us to drastically reduce the computational costs in both dataset generation and the DT training.

Steady-state finite element method data for training
As mentioned above, the dimensionless heat flux on the top surface (see Figure 1) is assumed to be between 0 and 100, while the Biot number on the bottom varies between 1 and 15.For every sample (q Ã ,Bi), the maximum temperature over the domain is calculated.The three variables (dimensionless heat flux, Biot number and maximum dimensionless temperature) are obviously physically related.As seen in Figure 3, for specific heat flux, the higher the Biot number, the lower the maximum temperature and vice-versa.For a specific heat transfer coefficient, an increase in heat flux corresponds to an increase in the maximum temperature.These trends are summarized by contours of the 1D results in Figure 4, where the Biot number and heat flux are represented along the x and y axes, respectively.Examples of 2D and 3D temperature maps are provided in Figure 5.In these simulations, the boundary conditions are specified in the Figure 5 caption.Digital twinning approach to transient thermal systems

Digital twin training, validation and testing
The validation-split ratio is 0.4 and the classical mean-squared-error (MSE), equation ( 6) is chosen as loss function as specified in Table 2.As an additional metric parameter to evaluate the training accuracy, the mean-absolute-percentage-error (MAPE) is calculated equation ( 7).MSE and MAPE are given as:  2.
The testing is performed on completely new/unseen data adopting the strategy shown by the flowchart in Figure 7.
To evaluate the performance of the DL, we assume that the prediction is "true" when MAPE between the heat transfer coefficient predicted by the DL and that of the one used in the FEM model as boundary condition is less than 2.5%.With this assumption, a user- HFF defined accuracy (like for a classification problem) is calculated as the ratio between true predictions and all data (see Figure 8).The confusion matrix shown in Figure 8, is generally adopted for such kinds of problems."Positive" means that prediction is higher than the real heat transfer coefficient, while "Negative" represents the opposite case.The results are satisfactory because the MAPE is larger than 2.5% only in 21 out of 1,000 samples.Among such samples, only in four the predicted heat transfer coefficient is lower than the real one.The accuracy, calculated as the ratio of the numbers of true/false positive cases over the total cases is 97.9%.

Digital twinning results and discussion
The digital twinning of the thermal system includes the following steps.In the first step, a comprehensive physics-based data set, covering different physical parameters, is generated.The second step involves training, validation and testing of a DNN to generate a DT of the Sinusoidal distribution Compound function with a tan and sin Linear heat flux

HFF
Here, the norm operator normalizes a function between 0 and 1.Now we dynamically couple the FEM solution and DT model in time as shown in Figure 2.This results in continuous data exchange between the physical system and its DT.The physical system provides timedependent heat flux applied on it and the threshold temperature (e.g.material failure or melting point temperature) to the DT.The DT returns an appropriate, time-dependent heat transfer coefficient that restricts the maximum physical system temperature just below the threshold maximum temperature.Since there could be a time lag between the timedependent heat flux applied and the evolution of the maximum temperature that depends on the transient response, we may need a safety factor to reduce the threshold temperature supplied to the DT.We fix it at 0.95 for all simulations.
The results are shown in Figures 9(a)-(c), respectively, for the applied heat fluxes of equations ( 8), ( 9) and ( 10).It is apparent that in all the cases the maximum temperature is below the melting temperature and the system is safe.In the first interval, the heat transfer coefficient is initialized with a maximum value of 15.Then, after every interval the DT tries to find the minimum heat transfer coefficient to guarantee a temperature below the maximum temperature constraint, but at the same time, very close to the maximum allowed temperature.It is obvious but essential to specify that it is practically impossible to obtain a constant maximum temperature if the boundary conditions are unsteady.
In Figure 9(a), we can see how the heat transfer coefficient sharply decreases from the maximum value and then follows the trend of the applied sinusoidal heat flux.As seen, the maximum temperature starts from the zero initial condition and increases until it reaches a value close to but below the maximum allowed temperature.Later, the dimensionless heat transfer coefficient flattens, reaching a stable value of unity (minimum value of the Biot number allowed).As observed in Figure 3(a), increasing the heat transfer coefficient above a certain value will have little effect on the maximum temperature for a given heat flux value.On the other hand, the heat transfer coefficient cannot be zero to avoid an infinite value of temperature.Therefore, the maximum and minimum heat transfer coefficient limits are imposed by physics, not by the DT implementation or the NN.For this reason, they are carefully taken into account when generating data, as well as in the testing and the DT architecture.
In Figure 9(a), the heat transfer coefficient initially drops sharply but gains momentum to follow the heat flux variation.When the applied heat flux drops during the latter part of the calculation, the heat transfer coefficient flattens to the minimum value allowed (unity).In Figure 9(b), we can see that, before and after the pulse, the heat transfer coefficient is constant (minimum allowed value) as the heat flux is at its pseudo-stationary state.The heat flux pulse clearly produces an increase in the maximum temperature.The heat transfer coefficient follows the same pattern, even with a low sampling frequency.We choose, as also in the other cases, to split the total 900 time steps of the simulations into 100 intervals.It means that the heat transfer coefficient is a piece-wise function, changing its value every nine time steps.The third case, shown in Figure 9(c), presents a heat flux that is steeply increasing from zero to about 90 with high-frequency oscillation given by the sinusoidal component.Even in this case, the heat transfer coefficient starts from its maximum value.Then, it reaches a plateau as heat flux is very low and then goes up with an obvious delay with respect to heat flux.In the last case shown in Figure 9(d), a linear heat flux is assumed.It results that when Fo is less than 50, the Biot number stands at its minimum allowed value while the maximum temperature linearly increases.Later, a non-linear increase in the Biot number determined by the trained DT enables to reach an approximately constant maximum temperature.

Digital twinning
approach to transient thermal systems  3.3.3Digital twinning of a non-linear case: temperature-dependent thermal conductivity.Here the reliability of the same DT for a non-linear heat transfer case with a temperaturedependent thermal conductivity is investigated.The analytical expression of our choice takes into account the previous physical limitations.We consider the following formula to express non-linearity: where T melt is the maximum allowed temperature.Figure 11(a) shows the results obtained using equation (10).As seen, the heat transfer coefficient required to have a maximum Digital twinning approach to transient thermal systems temperature below the maximum allowed temperature is smaller than the linear case (see Figure 9) due to the overall higher thermal conductivity.3.3.42D case.As mentioned above, to make the problem two-dimensional, it is sufficient to switch the heat flux and right side adiabatic boundary condition [see Figure 1(a)].In this case, the number of features and labels correspond to those of the NN used in the previous DT.Also, the retraining of the NN is done starting from the 1D model to accelerate the process.Furthermore, in this case, using a safety factor equal to 0.95, the DT is able to maintain the maximum temperature below the melting point, as shown in Figure 11.i is fixed at 1, the maximum temperature is located closer to the isothermal boundary (F2, 1 (b)).Later on, the location of the maximum temperature moves towards the heat flux boundary (F2), when Fo is approximately 5. To prove that, the dynamic position of maximum temperature for the case shown in Figure 12(a), in terms of the three dimensionless coordinates x Ã max ; y Ã max ; z Ã max , is pictured in Figure 12(c).For easier reading of Figure 12(c), it is important to remember that the plane equations where the isothermal (F6) and heat flux boundary (F2) conditions [Figure 1(b)] were applied are y Ã ¼ 0 and z Ã ¼ 1, respectively.Nevertheless, the 3D proposed digital twinning approach maintains the maximum temperature below the threshold temperature assumed.

Conclusions
In this article, a digital twinning approach to thermal systems was presented.The literature review demonstrated how digital twinning is still an emerging topic in this field.The work

HFF
presented here is therefore among the first to focus on a digital twinning approach with a clear aim to control a transient thermal system.The digital twinning approach here means a digital copy of a real system able to exchange information continuously with the physical world and influence each other.Although we assumed a surrogate FEM model for the physical system, demonstrating such a digital twinning process on a real physical system is the ultimate aim and the results presented here are the first step towards such an aim.
The objective of the digital twinning process in the presented work was to control the maximum temperature inside the system in the presence of an unsteady heat input into the system.However, the digital twin training was based on stationary data.The goal of the digital twinning process was to limit the maximum temperature below the maximum allowed temperature by allowing the DT to choose an appropriate convective heat transfer coefficient.Once the relevant and independent dimensionless variables were identified Digital twinning approach to transient thermal systems through a systemic dimensional analysis, a sensitivity analysis based on their variations was carried out.Furthermore, 1 D, 2 D and 3 D FEM setups were investigated to test the DT.
The training of the DL model was made separately for the three cases.The results clearly showed that limiting the maximum temperature below the maximum allowed temperature was possible even in the two and three-dimensional cases.In all cases, a safety factor was applied on the threshold maximum temperature before being passed on to the DT.The limitations of the proposed work include a static DT, trained on steady-state data.Even if the results shown in the present article demonstrate that the DT works effectively to control the maximum temperature of the system, similar results in completely different systems (in terms of dimensionless variables, their ranges, and boundary conditions) cannot be established a priori.Another limitation is the lack of availability of a real physical system.In conclusion, further studies should be aimed at developing a continuously changing DT, via a reinforcement learning platform to rapidly represent any new data.
Figure 3. DL training and validation data Figure 5. DL training and validation data: 2 D and 3 D examples Figure 6.Training the DT Figure 8. Testing: real vs predicted heat transfer coefficient, confusion matrix and accuracy thermal diffusivity(Fo).The objective of this sub-section is to demonstrate that the DT, being trained on dimensionless data, effectively works also with different physical thermal diffusivity and therefore different transient responses of the thermal system.In practical terms, the physical time window is fixed and different thermal properties are chosen by varying the Fourier number.We adopt equation (10) as test case.Four different Fo are simulated Fo ¼ 3; 15; 60; 150.A low Fo indicates a stronger internal energy stored compared to conductive transport and the opposite happens at higher Fo.The results are shown in Figure10.Figure10(a)shows the results for lower Fo while results for the highest Fo are illustrated in Figure10(d).At the lowest Fo numbers, the time shift between boundary conditions and maximum temperature indicates a higher thermal inertia and a slower thermal transport compared to the other three cases.From case (b) and (c) in Figure10, the maximum temperature profile becomes more noisy as the thermal inertia of

Figure 9 .
Figure 9. Digital twinning of a 1D heat transfer system Figure 10.Digital twinning of a 1D heat transfer system 3.3.5 3D cases.The 3 D cases are the final cases presented in this work.The additional isothermal boundary condition included in the computational domain (see "Step 2. DL model training -Python") is considered an added feature to clearly improve the accuracy of the NN.The dimensionless temperature value on the isothermal boundary varies between 0 and 70.The higher the temperature the higher will be the Bi number to maintain the maximum temperature under control.Two cases are presented in Figure 12 with isothermal boundary values of 50 [Figure 12(a)] and 70 [Figure 12(b)].It can be noted in both cases that the maximum temperature location changes with time [Figure 12(c)].At the beginning of the simulations, as the initial condition T Ã

Figure 11 .
Figure 11.Digital twinning of non-linear (1D) and two-dimensional cases Figure 12. 1D cases without safety factor