Abstract
The technology of visual servoing, with the digital twin as its driving force, holds great promise and advantages for enhancing the flexibility and efficiency of smart manufacturing assembly and dispensing applications. The effective deployment of visual servoing is contingent upon the robust and accurate estimation of the vision-motion correlation. Network-based methodologies are frequently employed in visual servoing to approximate the mapping between 2D image feature errors and 3D velocities, offering promising avenues for improving the accuracy and reliability of visual servoing systems. These developments have the potential to fully leverage the capabilities of digital twin technology in the realm of smart manufacturing. However, obtaining sufficient training data for these methods is challenging, and thus improving model generalization to reduce data requirements is imperative. To address this issue, we offer a learning-based approach for estimating Jacobian matrices of visual servoing that organically combines an extreme learning machine (ELM) and a differential evolutionary algorithm (DE). In the first stage, the pseudoinverse of the image Jacobian matrix is approximated using the ELM, which solves the problems associated with traditional visual servoing and is resistant to outside influences such as image noise and mistakes in camera calibration. In the second stage, differential evolution is utilized to select input weights and hidden layer bias and to determine ELM’s output weights. Experimental results conducted on a digital twin operating platform for 4-DOF robot with an eye-in-hand configuration demonstrate better performance than classical visual servoing and traditional ELM-based visual servoing in various cases.
1. Introduction
The digital twin (DT) technology endeavors to fabricate a simulated model of a tangible device, coupled with a comprehensive analysis of its life cycle and implementation model, to enhance the safety and manufacturing efficiency of the application system. In this regard, the vision servoing (VS) technology is a fundamental enabler, wherein the DT and VS are synergistically employed to achieve optimal system performance. VS is a closed-loop control approach that integrates robot motion control with visual information to rapidly process images and minimize error towards the desired position. The dynamic responsiveness and environmental adaptability of VS’s closed-loop architecture make it a versatile technique with diverse applications, including industrial robots [1, 2], automated guided vehicles [3, 4], unmanned aerial vehicles (UAVs) [5, 6], underwater robots [7, 8], and medical robots [9, 10], among others.
Closed-loop feedback control is the defining feature of VS. Based on the type of visual information utilized to provide feedback error, VS can be classified into three categories: (1) position-based visual servoing (PBVS) [11, 12], which defines the error in cartesian space; (2) image-based visual servoing (IBVS) [13, 14], which defines the error in image space; and (3) Hhybrid visual servoing (HBVS) [15, 16], which combines both of these characteristics but is computationally intensive. Among these approaches, IBVS has gained popularity as a mainstream research method due to its balance of robustness and accuracy, which is the primary focus of our work.
One of the primary challenges in IBVS is the acquisition of the image Jacobian matrix and its pseudoinverse. The image Jacobian defines the mapping relationship between the error in image features in 2D space and velocities in 3D space, which is constructed using calibration parameters, feature projection models, and depth information of the features relative to the camera frame. The analytic expressions for the image Jacobian corresponding to various types of image features have been derived in [17]. However, as the complexity of image features increases, obtaining the pseudoinverse of the image Jacobian for use in VS control law can be challenging. Furthermore, stability and convergence issues can arise in certain VS scenarios, such as camera retreat when the goal is rotated about the z-axis [18].
To address these challenges, an offline or online scheme can be employed to estimate the numerical value of the pseudoinverse. In the online scheme, the image jJacobian matrix is viewed as an optimization problem that can be solved using optimization methods. For example, a dynamic Broyden’s method has been proposed for model-independent visual servo control without precise calibration of kinematic and camera models in [19]. Similarly, a novel algorithm for multicamera pose estimation using virtual visual servoing has been proposed [20]. However, these iterative approaches may face issues such as unexpected camera motion due to poor selection of initial conditions or updated speed. In contrast, the offline scheme addresses the mapping problem using neural networks, which can avoid issues related to training well models. However, the computational complexity and error-proneness of vision-based robot positioning techniques often limit the number of controllable degrees of freedom. A novel method based on global picture descriptors and neural learning has been introduced in [21] to address these issues. An adaptive neural network module has also been constructed to approach unknown dynamics, allowing for control of robots with nonlinear and structurally unknown dynamics in [22]. However, the control parameters (learning rate, learning epochs, etc.) of these methods need to be tuned manually, which can be challenging when faced with local minima, and training can be time-consuming when there is a lot of training data.
The preamble machine learning techniques are used in the digital twin to analyze physical and virtual data while enhancing the design properties and operational characteristics of key aspects of the visual servo system based on a data-driven approach to further improve the system performance. In recent times, the learning efficiency of the extreme learning machine (ELM), a new neural algorithm, has been found to surpass traditional offline methods significantly. Remarkably, the ELM algorithm necessitates solely the determination of the number of hidden nodes as its solitary parameter. Previous works have attested to the benefits of ELM in this regard [23, 24]. ELM, in combination with fuzzy logic (FL), has been proffered as a viable solution for addressing three common problems in visual servoing (VS), which entail obtaining the interaction matrix and its pseudoinverse for defined feature points, selecting an appropriate gain value for the VS controller, and ensuring that features remain within the field of view (FOV) for VS permanency [25]. In addition, a novel estimation technique based on ELM and Q-learning has been put out in [26] to address complex modeling problems and achieve effective servo gain selection.
Notwithstanding the extensive research conducted on the matter, a significant oversight in the literature pertains to the inherent challenges associated with obtaining a considerable quantity of industrial field data sets to ensure a large range convergence of VS. Moreover, the weights pertaining to the input and hidden layer Q, along with the hidden layer bias B, remain static during the training phase. This characteristic may lead to suboptimal network performance when a substantial disparity exists between the training and validation data. Thus, if we are able to devise a means to update Q and B under specific circumstances while using the same or fewer training data as traditional ELM, it may be possible to enhance the prediction accuracy of the model. The Moore–Penrose (MP) generalized inverse of ELM is used to find the output weights analytically, and the differential evolutionary algorithm is used to choose the input weights. This hybrid learning technique is detailed in [27]. However, it should be noted that, to the best of our knowledge, this approach has not yet been implemented in a visual servoing application.
This paper proposes a novel digital twin-based approach to visual servoing that integrates extreme learning machine (ELM) and evolutionary algorithm (EA). The proposed approach leverages the capabilities of ELM to approximate the pseudoinverse of the image Jacobian matrix, which effectively addresses matrix singularity and provides robustness against image noise and camera calibration errors. Compared to other classification techniques [28, 29], ELM significantly reduces the training time since it avoids slow gradient-based learning algorithms and eliminates the need for iterative parameter calculation. Furthermore, because EA is frequently used as a global optimization seeking strategy, the combination of analytical methodologies and EA is anticipated to be successful for network training. In the suggested method, EA is used to pick the best input weights and hidden biases, allowing the network parameters to be adaptively changed to raise the model’s prediction accuracy. Furthermore, in the context of platform development, we have devised a digital twin infrastructure grounded on the EtherCAT real-time Ethernet technology, tailored to physical robotic entities, with the aim of guaranteeing the safety of scenario deployment and the efficacy of performance validation. Overall, the proposed visual servoing approach with ELM and EA is expected to provide better performance than conventional techniques in challenging robotic applications.
The present article is structured as follows: in Section 2, an exposition of the sophisticated approximation methodology for IBVS is presented. The fundamental principles of ELM and EA are introduced and subsequently employed in tandem within the visual servoing module. Section 3 presents empirical results based on a 4-axis robot digital twin operating platform, which helps to validate the effectiveness of the proposed approach. Finally, Section 4 outlines the conclusion derived from the study.
2. A Novel Visual Servoing with ELM and EA
The present section furnishes an exhaustive exposition of the proposed methodology for IBVS, with a specific focus on point features, designed for robot manipulators operating in an eye-in-hand configuration.
The present exposition commences by expounding upon the process of image plane projection. Specifically, the pinhole model of a camera, depicted in Figure 1, is introduced. It is assumed that the camera captures K pixels, denoted as and that the origin is located at the top left-hand corner of the image plane, in accordance with convention. The proposed methodology proceeds to consider feature points, denoted as in the image frame, and corresponding desired image coordinates , which are typically obtained through a priori knowledge. In general, the camera projection can be expressed in the following form:where

The notation denotes the homogenous expression of the image point , while represents the features’ depth information relative to the camera frame. The camera parameter matrix is denoted by , and is the project matrix. The pixel dimensions are given by and for width and height, respectively, and denotes the pixel coordinate of the principal point. Furthermore, represents the feature points expressed in the camera frame, with denoting its homogenous expression. is a homogenous transformation matrix that relates the camera frame and the end effector frame, and denotes the feature points expressed in the world frame. Feature points expressed in different frames are obtained by (1). The goal of IBVS system is to minimizewhere is the vector of the features error signal. To convert image space error signals to manipulator pose space, we use the following relationship:where is the time derivative of features error and is the camera spatial velocity. The link between the variation of and the camera velocity is described by the image Jacobian matrix . According to the classical kinematic theory,
Take the derivative of in (1)
Injecting (5) in (6), where related to is
The final general control law can be calculated by using as the input to the low level controller and obtainingwhere is a control gain to ensure an exponential decrease of the error, is the estimation of Moore–Penrose pseudoinverse of the image matrix.
As shown in (8), the primary objectives of visual servoing control are to approximate the relationship between the feature error signal and the camera spatial velocity , which is primarily determined by the control gain and the estimation of the Moore–Penrose pseudoinverse of the image matrix . However, as image attributes become increasingly intricate, computing becomes a laborious and complicated task. To address this challenge, ELM is employed to learn the nonlinear coupling relationship between and by randomly selecting the input weights and hidden bias. Furthermore, DE is utilized to enhance the performance of the visual servoing controller by tuning the input weights and hidden bias.
The accurate estimation of is essential to determine the performance of the visual servoing system, but it is a challenging task due to its dependence on camera calibration parameters and object model parameters, which are difficult to precisely determine. Considering the lack of extensive research on the estimation form of , a novel method to approximate it, is proposed. ELM is a type of single-hidden layer feedforward neural network algorithm (SLFN), and its algorithm framework is presented in Figure 2. Distinguishing itself from conventional gradient-based learning algorithms such as Levenberg–Marquardt (LM) and backpropagation (BP), ELM employs randomly generated or artificially set weight coefficients and biases for the hidden layer, which obviates the need for gradient descent method (GDM) for error backpropagation and reduces the requirement for tuning. The objective of ELM is to determine the weights of the output layer. Based on these properties, ELM offers faster computing efficiency and stronger computing power than traditional methods.

In the context of the IBVS system, assuming a given set of known input samples, , where is the size of the input vector for a single sample, and , where is the size of the output vector from a single sample. Accordingly, we can express each sample as and .
In this paper, we employ AprilTag [30, 31] as the target and extract four corners as the image features. Subsequently, we utilize ELM to estimate the Cartesian space velocity of the camera, where the dimensions of the input and output vectors are and , respectively. The hidden layer’s output function iswhere is weight matrix, which maps the hidden layer to output layer, is weight matrix, which maps the input layer to hidden layer. is defined as
In this regard, the function is commonly known as the neuron activation function or feature mapping function. It can be defined as various functions such as the sigmoid function, RBF function, and others. In addition, stands for the matrix, which is defined at random as the initial value for the hidden layer’s input. Furthermore, is a representation of the hidden nodes’ bias.
The primary objective of this process is not to learn and , but rather to minimize the error between and . Therefore, the objective function can be defined as
The Moore–Penrose pseudoinverse matrix of is calculated in this context, and it is multiplied by to produce the lowest norm least squares solution of the objective function . In other words, can be expressed as . Once the matrix has been trained, it can be utilized to predict the camera space velocity for a given feature error input vector using (9).
The random assignment of input weights and hidden bias can potentially impact the accuracy of the model’s predictions. To mitigate this issue, differential evolution (DE) is employed. DE is a type of evolutionary algorithm (EA) first proposed by Storn and Price [32]. The fundamental concept behind utilizing DE in ELM can be described as follows: Step 1, Population Initialization: here, NP represents the population size, and D is the dimension of each individual. In our particular design, each individual consists of input weights and hidden bias, with a total dimension of . Thus, a population can be generated in the following manner: Step 2, Mutation: like living creatures, the individuals in the population (denoted as ) are subject to mutation. Consequently, mutant individuals can be described as follows: The current investigation revolves around the selection of indices, specifically referred to as , , and , from a given population. It is essential that these indices are chosen in a random manner and are distinct from each other. To ensure adherence to this requirement, the population size, denoted as , must surpass a value of four. Furthermore, the variable , which is constrained within the interval , represents the zoom factor that governs the extent of scaling for differential variation. Step 3, Crossover: to broaden the population’s diversity, it is imperative to carry out crossover operations, and the D-dimensional of crossover individuals is defined as follows: where Herein, , where , represents the jth value that is randomly generated and conforms to the range of . The crossover probability factor , which resides in the interval , is subject to determination by the user. Additionally, denotes an index that is randomly selected from the set of integers . This index guarantees that at least one parameter from is incorporated into the vector . Step 4, Selection: regarding the selection of individuals for the ensuing generation, the mutant individual denoted by is subjected to comparison with . If the former displays a higher level of performance in relation to the latter, the former shall be designated for the succeeding generation. In contrast, should outperform , the original individual , will be preserved.
Based on the above basis, a novel approach named EA-ELM-IBVS is proposed and its framework is shown in Algorithm 1.
|
Before testing the accuracy of the ELM network on the validation set, we know that is of size and is of size in (10). As a result, we create the population at random. Every member of the population possesses all and parameters as a vector
All values of and are initialized randomly within the interval . Subsequently, we calculate the output weights for each individual using the least squares solution in accordance with (11). The activation function adopted in this context is as follows:
The fitness function is described as follows:where is the root mean squared error (RMSE) to evaluate each individual.
Conventionally, the RMSE is employed as the fitness metric, derived from the complete training set. However, given that the ELM optimizes β through the least-squares solution, evaluating the fitness metric based on the entire training set may result in overfitting. Therefore, in order to achieve enhanced performance of the visual servoing controller while mitigating computational overhead, we exclusively assess the RMSE on the validation set. Subsequently, once the fitness of each individual has been ascertained, we proceed to delve into the core concept of DE, encompassing mutation, crossover, and selection operations.
The user often assigns a fixed value to the zoom factor during mutation, which has an impact on the ability for global optimization. In the event, where is set to a smaller value, the model will tend to escape from the local minimum, leading to a slower convergence rate. To circumvent this issue, an adaptive approach is described below for adjusting the value of :wherewhere denotes the initial value of the zoom factor, and ranges between and . During the early stage of evolution, is assigned a relatively large value to preserve the diversity of individuals and forestall the occurrence of premature convergence. Conversely, during the later stages of evolution, is assigned a smaller value to retain valuable information, prevent the degradation of the optimal solution, and augment the likelihood of discovering the global optimum.
While performing crossover, if the zoom factor assumes a large value, the mutated individuals may acquire values outside of the permissible range. As a result, a limit function is added that has the following definition to get around this problem:where and are the lower and upper bounds. Here, we define and .
During the selection process, the mutated population is compared with the original population based on their fitness. To achieve this, the output weights of the mutated population are computed and their corresponding fitness is evaluated on the validation set. The individuals with better fitness are then chosen for the next generation. A minor validation error does not necessarily imply a small testing error, which depends mainly on the distribution of the validation data. Therefore, choosing individuals only on the basis of fitness may not be the best course of action. In order to overcome this problem, and motivated by [27], a new criterion, the norm of output weights, denoted as , is added to the selection process. The individual who results in a smaller value of is chosen when the difference in fitness between several individuals is minimal. Therefore, the selection strategy for the next generation is defined as
In this formula, is the fitness function, and is the predetermined tolerance rate. After selecting a fresh population, determine its fitness before starting the DE process again and repeating it until the desired iteration number or range of convergence is obtained.
3. Experiment Result
3.1. Numerical Simulations
Three distinct systems’ performances were compared using simulations: the suggested IBVS system (EA-ELM-IBVS), the extreme learning machine-based IBVS (ELM-IBVS), and the classical image-based visual servoing (C-IBVS). The volume and distribution of training data have a significant impact on the mapping model’s capacity to learn the link between image attributes and the end effector’s spatial velocity. However, acquiring training data that cover the entire workspace of the robot can be highly time-consuming. Therefore, improving the generalization ability of the mapping model with fewer training data can expand the application capability of the system. To demonstrate the advantages of the proposed method, experiments were designed and conducted using MATLAB and the Machine Vision Toolbox [18]. The camera was mounted at the end of the end effector, and the homogeneous transformation matrix between the camera and the end of the manipulator was assumed to be the identity. This can be expressed as
The camera resolution is specified as , with the pixel coordinates of the camera’s origin being . The camera frame rate is identified as , and the focal length is determined to be . Additionally, the desired feature points in both Cartesian space, denoted as , and the pixel frame, denoted as , are provided as
The successful completion of the designated task is contingent upon the attainment of a feature error of less than pixels. In the current study, the number of hidden nodes, denoted by , is set to , whereas the population number, denoted by , is established at . Furthermore, the algebra of population evolution, represented by , is determined to be . These technical specifications are critical in guiding the implementation and optimization of the utilized algorithm, emphasizing their importance in achieving the desired outcomes.
The forthcoming subsections adopt two distinct scenarios to facilitate a comprehensive comparative evaluation of C-IBVS, ELM-IBVS, and the proposed EA-ELM-IBVS. In the first scenario, the validation data are presumed to be in close proximity to the training data, and the camera’s spatial velocity is examined across the three aforementioned methods. Subsequently, in the second scenario, a greater disparity is introduced between the initial poses of the validation and training data, wherein the superiority of the EA-ELM-IBVS approach in terms of generalization performance becomes evident. This comparison analysis sheds important light on the efficacy and robustness of the various visual servoing techniques under various circumstances, highlighting their individual advantages and disadvantages.
3.1.1. Case 1: Validation Data Training Data
In the first scenario, is designated as the initial pose of the end effector frame relative to the world frame. Subsequently, the end effector is directed to move towards the desired pose until the image feature error converges. The calculated error and spatial velocity at each iteration of the process are utilized as the inputs and outputs of the training data, respectively. The corresponding information pertaining to is designated as the validation data. The homogeneous transformation matrix of and is of critical importance in this experimental setting, which is shown as follows:
It is apparent that a minute discrepancy of solely in the direction exists between the initial poses of the training data and the validation data. The forecasted spatial velocity, , is depicted in Figure 3, with the red curves signifying the desired output. As inferred from the experimental outcomes, the disparity in prediction between the two techniques in the first scenario is not statistically noteworthy. Since the validation and training data are proximate, the ELM technique has computed the optimal prediction via norm least squares approximation, while the differential evolution approach has had a subtle influence.

(a)

(b)

(c)

(d)

(e)

(f)
3.1.2. Case 2: Validation Data Training Data
In the second scenario, the process information of the homogenous transformation matrix is utilized as the training dataset. Nevertheless, an alternative initial pose, , associated with the world frame is designated as the validation dataset. Notably, the disparity in the x-direction is augmented by a factor of ten. The value attributed to is thereby expanded as follows:
Observing the results presented in Figure 4, it is evident that the discrepancy between the initial poses of the training and validation datasets is in the x-direction. This value is substantially larger than the corresponding disparity in the first case. Notably, the curve denoted in green and generated by the ELM-IBVS exhibits a pronounced phenomenon of significant deviation. This trend implies that the predicted spatial velocity curve markedly increases in both slope and extreme value. Notably, if the velocity predicted by the ELM-IBVS is fed into the underlying velocity controller under identical experimental conditions, the motor will experience more significant vibrations, potentially leading to damage to the robot. In contrast, EA-ELM-IBVS manages to retain its predictive ability even when confronted with substantial differences between the testing and validation datasets.

(a)

(b)

(c)

(d)

(e)

(f)
3.2. Real Experiments
In order to ensure the safety and efficiency of physical device experimentation, we have developed a digital twin operation platform that integrates a 4-axis robot virtual model with its corresponding physical entity. This organic combination enables the concurrent evolution of both the virtual and physical components and facilitates system-level analysis and performance verification. The effectiveness of the proposed method is validated through experiments conducted on the platform, as illustrated in Figure 5, thus demonstrating its practical applicability in real-world settings. This platform encompasses a testbed composed of a high-level visual servoing controller as well as a low-level motion controller. To acquire image data, a Realsense D435 camera is affixed to the end of the SCARA robot. Subsequently, the environmental data is conveyed to the visual servoing controller, while the motion controller receives motion commands following certain processing steps.

The fundamental aspect of the construction of a digital twin framework lies in the capacity to procure accurate data from tangible apparatuses contemporaneously. In consideration of the system depicted in Figure 5, the data flow within the system is illustrated in Figure 6. To facilitate real-time activities, we opted to implement the xenomai patch on a Linux IPC (Intel i3-4150, 3.5 GHz). Subsequently, the following three programs are installed on the IPC: (1) The visual servoing controller, which processes the visual servoing algorithm, robot kinematics, and Jacobian conversion at a rate of 30−50 Hz. (2) The motion controller, which enables real-time layer control and facilitates communication of motion tasks at a frequency of 1 KHz. (3) The data collector, which generates a data buffer designed to receive pulse values for each axis in real time. Additionally, four Panasonic MADHT1507BA1 EtherCAT servo drives are configured to operate in cyclic synchronous (CSP) mode, functioning as slaves of the EtherCAT network.

In the subsequent part, we proceed to carry out experiments on actual SCARA robots in relation to cases 1 and 2, respectively. We present a scenario wherein a SCARA robot is tasked with conducting an IBVS eye-in-hand static tracking experiment. The outcomes of the experiment are showcased in Figures 7 and 8. In the context of case 1, it is evident that ELM-IBVS exhibits a consistent bias along with larger fluctuations in amplitude. In contrast, the proposed EA-ELM-IBVS successfully achieves convergence at the desired speed and maintains smaller fluctuations. Moving on to case 2, the fluctuations observed in ELM-IBVS are more pronounced. Additionally, a peculiar phenomenon arises where the predicted velocity direction contradicts the expected velocity direction, which is also observed in EA-ELM-IBVS. To address this issue, the most straightforward solution entails augmenting the training data samples. However, merely increasing the number of training samples falls short in meeting the requirements of robots operating in a wide range, thereby presenting a significant research problem that necessitates further investigation in subsequent studies.

(a)

(b)

(c)

(d)

(e)

(f)

(a)

(b)

(c)

(d)

(e)

(f)
Ultimately, within the domain of error analysis, there exist two quintessential calculated indicators which hold great import, namely the sum of error squares (SSE) and the mean square error (MSE). These indicators, SSE and MSE, are commonly described aswhere denotes the predicted value, and refers to the desired value. , on the other hand, denotes the total number of samples under consideration. It is pertinent to note that and are four-dimensional vectors in the current experimental setup. Therefore, in order to measure the extent of improvement brought about by the experimental intervention, we propose the utilization of two indicators, namely the average of the sum of error squares (AS-SSE) and the average of the sum of mean square errors (AS-MSE), which are defined as follows:
Tables 1 and 2 present the pertinent data indicators for the aforementioned four experiments. Upon conducting an in-depth analysis of the simulation data, it was observed that the EA-ELM-IBVS method outperformed the other methods with respect to AS-SSE and AS-MSE. These findings corroborate the notion that the proposed methodology exhibits superior generalization ability, particularly when there is a significant divergence between the training and verification data. Moreover, an examination of the actual robot data also revealed that the EA-ELM-IBVS approach demonstrated superior performance, thereby providing evidence for its heightened robustness in the face of environmental factors such as image noise.
4. Conclusion
Given that the efficacy of learning-based methods for approximating the correlation between image feature errors and the spatial velocity of the end effector is primarily contingent on the volume of data contained within the training set, it can be a time-consuming and challenging task to amass a sufficient amount of data for the robot to operate effectively across a broad range of scenarios. To address this issue, the present study proposes an EA-ELM-IBVS approach that has been shown to exhibit superior generalization ability in comparison to the ELM-IBVS method. This signifies that the suggested method can perform better with a comparatively smaller training dataset. Meanwhile, the built digital twin operation system can better realize the organic interaction between virtual models and physical entities, which facilitates the efficiency and security of algorithm deployment and performance verification. Nevertheless, it is crucial to acknowledge that there are notable hurdles that persist in situations where the training and validation data showcase substantial dissimilarities concerning image noise. In light of this circumstance, forthcoming investigations will concentrate on undertaking algorithmic and methodological interventions and improvements that render the learning-based approach more resilient and dependable in the realm of image Jacobi estimation within the visual servo system. Furthermore, efforts will be made to incorporate additional constraints when implementing the proposed approach on tangible devices.
Data Availability
The data supporting the findings of this study are available from the corresponding author upon reasonable request.
Conflicts of Interest
The authors declare that they have no conflicts of interest.
Authors’ Contributions
Minghao Cheng and Hao Tang equally contributed to the study.
Acknowledgments
This work was supported by the Princess Nourah bint Abdulrahman University Researchers Supporting Project number PNURSP2023TR140, Princess Nourah bint Abdulrahman University, Riyadh, Saudi Arabia.