Introduction

The choice to use drilling processes in the upstream petroleum sector instead of more expensive activities is a crucial decision. This strategic shift holds significant importance and could lead to a substantial reduction in operational costs1,2. There are different parameters in the drilling rig operation (i.e., technical, economic, and human parameters) that have a high influence in this regard3. In the drilling industry, the relationship between drilled depth and time can be highly changed using different parameters. This relationship arises from the rate of penetration in the rock2,4. In the case of the drilling program, the meticulous selection of parameters that influence the rate of penetration into the formation is crucial and can potentially augment the penetration rate of the drill into the rock. The higher drilling rate can be due to the higher rate of penetration, which can decrease drilling time and finally decrease the total operation cost (e.g., Improving and predicting the impressive parameters can enhance the penetration rate for a specific zone until around 50 m that drilling rate is around 37 m per day. Since the wells of Iran are classified as deep wells, improvement and prediction of penetration rates in drilling wells are very significant5.

The most appropriate penetration rate within the formation can be obtained by utilization of calculating approaches that can increase the drilling rate and may lead to lower costs by comparing and studying the impressive parameters by taking into account the mud utilized and the drill bit and hydraulics type as well as the rig power6. The least error can be achieved by considering impressive factors of the volume flow rate of drilling fluid, the formation rate of penetration, density of drilling fluid drill string rotational speed (RPM), the pressure of standpipe, size/type of drill bit along with kind of formation the relation among the weight on the drill bit and the drilling factors.

The above-mentioned parameters consist of the following factors:

  1. (1)

    It is known that the higher used weight does not presently mean a higher drilling rate. Weight on a bit is the weight exerted on the drill bit that is named weight on bit (WOB). The most important factor of the drilling rate is when the driller attempts to identically use the optimized weight of the drill string for the bit under the well bottom (e.g., in some sticky substrates, the higher weight leads to the bit falling within the layer. It can barricade the dig of the formation and the drill string rotation, and sometimes causes the drill string traps in the form or even cutting the drill string.

  2. (2)

    The rate of Penetration (ROP) into the formation shows the drilling rate. In this way, the drill adjustment attempts to retain it at the highest feasible level to the lowest cost along with the lowest risk of the operation. ROP in an operation team is commonly called and introduced as the per meter depth drilled (with the unit of time).

  3. (3)

    The drilling depth (Depth) along with the type and hardness of the formation have been classified as the factors that affect the size and type of bit (Rock Bit or PDC Bit).

  4. (4)

    The volumetric flow rate of drilling fluid (GPM) is used with the driller based on drilling states. It includes the size of the drill bit as well as the diameter of the driller nozzles, depth, and pressure of the standpipe. Researchers are trying to retain it at a constant amount.

  5. (5)

    The density of drilling fluid has been considered as one of the factors that is commonly used when hydrostatic pressure in the well depth is more than the formation pressure that prohibits the fluid formation flow of the well. It is commonly attempted to identify and use the optimized fluid density during drilling and determining the optimized fluid density while taking into account the formation layer and well situations.

  6. (6)

    The density of drilling fluid is calculated as the unit of pound per cubic foot. In addition, for drilling operations, the fluid density is named mud weight (i.e., MW).

  7. (7)

    Considering drilling conditions such as depth and formation type, one of the crucial parameters that is sought to be utilized optimally is the rotational speed of the drill string, measured in rounds per minute (rpm).

  8. (8)

    The formation type that is being drilled varies in the case of geology. This leads to the enhancement of the drilling depth. The layers of formation regulation are usually similar in distinct points of a particular oilfield and even sometimes at the same depths, various wells commonly have the same and unchanged structures.

  9. (9)

    Another factor that affects the drilling rate and states is the fluid pressure pumped in the drill string. The drilling fluid can be pumped by taking into account the reservoirs in the drilling string. The drilling fluid, upon exiting the drill bit, should return to its initial level in the reservoirs, both in the annular space surrounding the drill string and within the well. The fluid pressure exerted by the drill depends on several factors, such as the density of the drilling fluid, the effectiveness of the pumps, the pressure drop across the drill nozzles, the pressure drop within the drill string, and the volumetric flow rate determined by the driller.

Many scholars are concerned about the optimization and simulation of the procedure utilizing the common mathematical methods due to the high variation of the parameters involved7,8,9,10,11. It is important to note that simulating systems by usual mathematical patterns like differential equations for complex systems that contain uncertainties is known as a relatively good method without high effective performance. Hence, many studies have used artificial intelligence and deep learning models to approximate key variable of different processes, including the drilling operations12,13,14,15,16.

Furthermore, some drawbacks of typical intelligent models have driven scholars to employ hybrid evolutionary algorithms to achieve a better approximation of the desired variable17. In this sense, Anemangely and Ramezanzadeh optimized the performance of an artificial neural network (ANN) by using a cuckoo optimization algorithm (COA) and particle swarm optimization algorithm for predicting the drilling rate18. They concluded that the combination of the neural network with COA could achieve high accuracy in the mentioned utilization. Moraveji and Naderi successfully employed a bat algorithm for identifying the optimal range of parameters for maximizing the drilling rate of penetration8.

This research applies machine learning techniques to predict the WOB as a function of different combinations of depth, RPM, ROP, and GPM as the input variables. Firstly, the relevance test sorted the impact of each independent variable on the WOB. Then, two wise optimization algorithms, namely grey wolf optimization (GWO) and biogeography-based optimization (BBO), are applied to improve the prediction capability of artificial neural networks in estimating the WOB. Although ANNs have been promisingly used, the authors came across a few studies that conduct the optimization of ANNs for prevailing its computational drawbacks in the mentioned field. Therefore, two novel ensembles of GWO-ANN and BBO-ANN are designed, and their best structure is determined by executing a sensitivity analysis. The effect of input variables dimension on the accuracy and training time of the best neural-evolutionary framework has also been investigated in this study. The prediction accuracy of the best neural-evolutionary framework is compared with the traditional ANN, multiple linear regression (MLR), support vector regression (SVR), adaptive neuro-fuzzy inference systems (ANFIS), group method of data handling (GMDH).

Data preparation and statistics

2250 data samples, comprising four independent variables (i.e., drilling depth, RPM, ROP, and GPM), as well as WOB as the dependent variable, are used to develop the intelligent models of this study. The key statistical parameters related to this databank are summarized in Table 1.

Table 1 Statistical analyses of the WOB and effective parameters.

The histogram of all independent and dependent variables as well as their scatterplots (known as the pair plot) is presented in Fig. 1.

Figure 1
figure 1

Pair plot of the involved variables in the present study.

1800 samples (80% of the whole dataset) are used to perform the learning phase of the proposed models. Indeed, this learning phase helps the intelligent tools to understand the relationship between the WOB and its influential factors. Then, the capability of the developed networks is evaluated by the remaining 450 samples (20% of the whole dataset), which were not seen by the trained models in the learning phase. This stage, which is known as the testing phase, monitors the generalization ability of the trained models against some unseen data samples.

Feature importance analysis

Feature importance analysis is a very important step before the development of predictive models to estimate the target WOB. This analysis helps understand the level of WOB sensitivity on independent factors. The results of applying Pearson method19 on the available databank to reveal the importance level of features have been depicted in Fig. 2. This figure clarifies that the Depth and RPM are the most important features for determining the target, i.e., WOB.

Figure 2
figure 2

Results of the feature importance analysis.

The p-values between WOB and each independent variable are reported in Table 2. This analysis claimed that depth and RPM are highly significant statistically. In addition, ROP and GPM have been identified as very significant and not significant variables, respectively.

Table 2 The observed p-values between WOB and the input variables.

Methodology

After providing a proper dataset, it is divided into two parts for training and testing the proposed ANN, GWO-ANN, and BBO-ANN networks. To develop the mentioned ensembles, the GWO and BBO optimization techniques are coupled with the ANN. Then, the best structures are used to estimate the WOB. The results are evaluated utilizing two statistical criteria, i.e., mean square error (MSE) and mean absolute error (MAE). Equations (1) and (2) formulate the MSE and MAE. Finally, the results of the ensembles are compared with the typical ANN to examine the effect of the applied metaheuristic algorithms. Moreover, the predictive mathematical equation of the most capable network is presented21.

$$MSE = \left( {1/N} \right)\sum\limits_{i = 1}^{N} {\left( {WOB_{i}^{act} - WOB_{i}^{cal} } \right)^{2} }$$
(1)
$$MAE = \left( {1/N} \right)\sum\limits_{i = 1}^{N} {\left| {WOB_{i}^{act} - WOB_{i}^{cal} } \right|}$$
(2)

in which the superscripts of act and cal denote the actual and calculated values of WOB, respectively. The term N represents the number of instances.

Artificial neural network

The ANN algorithm can be used to solve different non-linear engineering problems22. McCulloch and Pitts23 suggested the idea of this algorithm, which mimics the biological neural procedures. As other machine learning methods, two different groups of data are needed to develop the ANN. The first one includes the most data named training dataset that is employed for pattern recognition via developing mathematical relations. The second group analyzes the performance of the developed network named testing database. Multi-layer perceptron (MLP) is known as one of the most appropriate methods between various types of ANNs that were employed in many studies24. In the present paper, we have used the algorithm of LM (Levenberg–Marquardt)25, GWO, and BBO to accomplish the training phase of MLP neural networks. A typical MLP structure related to our study is shown in Fig. 3.

Figure 3
figure 3

The typical MLP structure.

After performing the training method in the process of BP, the efficiency error is calculated based on the difference between the exact and estimated response variable. For updating the computational weights (W) as well as biases (b), the efficiency error is propagated backward. We consider Out, T = {\({Depth}_{i}\), \({RPM}_{i}\), \({ROP}_{i}\), \({GPM}_{i}\); i = 1, 2, …, N}, and F as output vectors, input matrix, and the activation function of the MLP model. After that, the jth computational unit performance (i.e., the neurons of the MLP) can be described as follows:

$$Out_{j} = F \, \left( {\sum\limits_{m = 1}^{M} {T_{m} W_{mj} + b_{j} } } \right)$$
(3)

It should be mentioned that the current study uses the tangent sigmoid (tansig) and linear activation functions in the hidden and output layers, respectively26.

Grey wolf optimization

The grey wolf optimization has been suggested as a progressive evolutionary algorithm27. GWO follows the behavior of grey wolf herds, which is a type of social hierarchy. For the first time, this algorithm is suggested by Mirjalili et al.28. A flowchart of this algorithm is shown in Fig. 4. A group of wolves in the wolf flock (named alpha, α) is the head of the herd involving several male and female wolves that are decision-makers flocks for hunting and searching. A group of wolves in the wolf flock (named beta, β) assist and obey the first group of decision-makers. The Beta wolves establish the discipline of the herd, and this is their main duty. When alpha (α) wolves retire or die, beta (β) wolves can substitute instead of them. The next group of wolves that act as scouts, sentinels, hunters, etc., and have the weakest relationships and babysitters are named delta (δ) (omega (ω) wolves). Omega (ω) wolves are the weakest wolves in the flock, and it is likely to see internal fights without omega (ω) wolves. These wolves can participate in hunting, and this is their main social behavior with a social hierarchy29,30.

Figure 4
figure 4

The working procedure of the GWO algorithm.

Three base stages of the algorithm of GWO are (a) approaching the target, following, detecting, and (b) circumambienting the target, as well as (c) attacking the prey31. Alpha (α) is used as the highest-fitted solution, and after that β, δ, and ω. The encompassing formula is obtained as below:

$$\overrightarrow {D} = \left| { \, \overrightarrow {S} \cdot \, \overrightarrow {P}_{p} (t) - \overline{P} (t) \, } \right|$$
(4)
$$\overrightarrow {P} (t + 1) = \overrightarrow {P}_{p} (t) - \overrightarrow {F} \, \cdot \, \overrightarrow {G}$$
(5)

where t stands for the time of iteration. A new position for a wolf is proposed by the vector of \(\overrightarrow{D}\). The \(\overrightarrow{F}\) and \(\overrightarrow{S}\) stand for the coefficient vectors, and \(\overrightarrow{{P}_{p}}\) and \(\overrightarrow{P}\) show the prey and wolf positions, respectively.

We assume that δ, α, and β are more trustworthy knowledge around the prey place, the most appropriate solutions for considered positions might be updated. After that, other wolves can then register the positions.

$$\overrightarrow {D}_{\alpha } = \left| { \, \overrightarrow {S}_{1} \, \cdot \overrightarrow {P}_{\alpha } - \overrightarrow {P} \, } \right|$$
(6)
$$\overrightarrow {D}_{\beta } = \left| { \, \overrightarrow {S}_{2} \cdot \, \overrightarrow {P}_{\beta } - \overrightarrow {P} \, } \right|$$
(7)
$$\overrightarrow {D}_{\delta } = \left| { \, \overrightarrow {S}_{3} \cdot \, \overrightarrow {P}_{\delta } - \overrightarrow {P} \, } \right|$$
(8)
$$\overrightarrow {P}_{1} = \overrightarrow {P}_{\alpha } - F_{1} \cdot \overrightarrow {D}_{\alpha }$$
(9)
$$\overrightarrow {P}_{2} = \overrightarrow {P}_{\beta } - F_{2} \cdot \overrightarrow {D}_{\beta }$$
(10)
$$\overrightarrow {P}_{3} = \overrightarrow {P}_{\delta } - F_{3} \cdot \, \overrightarrow {D}_{\delta }$$
(11)
$$\overrightarrow {P} (t + 1) = \frac{{\overrightarrow {P}_{1} + \overrightarrow {P}_{2} + \overrightarrow {P}_{3} }}{3}$$
(12)

Finally, in this method, when a target stands the attack is accomplished. Note that the F includes random variables among − 2α and 2α. This shows that the wolves attack the target (\(\left|A\right|\) < 1) or look for a better one (\(\left|A\right|\) > 1)29.

Biogeography-based optimization

Simon32 introduced biogeography-based optimization that comes from the biogeography knowledge and dispensation of various species. This model has been classified as a population-based search method, and Mirjalili et al.33 applied this algorithm to MLP to optimize its efficiency. Figure 5 denotes the flowchart of the BBO. This model begins using the generation of a so-called random population called “habitat.” These parameters show possible solutions that have been obtained from the habitat suitability index (HSI). Moreover, the suitability index variable (SIV) can be used to evaluate the habitability of the habitats and zones. More precisely, an SIV shows the population of the candidate solutions and is classified as a group of real numbers. The BBO extracts two distinct practices (i.e., migration and mutation). The basic purpose is to improve the quality of the possible solutions by enhancing them according to other existing solutions. In this stage, λg, an immigration rate, is considered for deciding about the need for correction of each SIV34. In addition, emigration rates (μg) can be used to choose the solution that migrates. It should be noted that other metaheuristic methods are retained constant away from this act to pass probabilistic corruption35.

Figure 5
figure 5

The operation procedure of the BBO algorithm.

Abrupt variations in the case of HSI values can be due to some normal hazards threatening geographical zones. The habitat likely strays due to equilibrium HIS. This method is called mutation, and the probability of each species count determines its rate. This probability is expressed as below:

$$P_{g}^{f} = \left\{ {\begin{array}{*{20}l} { - (\lambda_{g} + \mu_{g} )P_{g} + \mu_{g + 1} P_{g + 1} } \hfill &\quad {S = 0,} \hfill \\ { - (\lambda_{g} + \mu_{g} )P_{g} + \lambda_{g - 1} P_{g - 1} + \mu_{g + 1} P_{g + 1} } \hfill &\quad {1 \le S \le S_{\max } - 1,} \hfill \\ { - (\lambda_{g} + \mu_{g} )P_{g} + \lambda_{g - 1} P_{g - 1} } \hfill &\quad {S = S_{\max } .} \hfill \\ \end{array} } \right.$$
(13)

in which S stands for the species number for a region.

To decide whether the relation will change or not, any population relation becomes a possibility. This additionally shows whether the relation is a solution to the existing issue or not. The proximity of the solution to the final solution can be obtained by probability value. A higher probability value means that the solution is near to the overall solution36.

Stopping criteria

The involved optimization algorithms follow an iterative procedure to minimize the deviation between actual and predicted values of the WOB in the learning phase. The weights and biases of the ANN model are continuously updated by the backpropagation method to reduce this deviation as much as possible. Reaching the maximum number of iterations, converging to a prespecified accuracy, and experiencing a minimum error gradient are the most well-known criteria for stopping the learning phase.

Results and discussion

As explained, this study applies two novel optimization algorithms to perform the learning stage of artificial neural networks for estimating the WOB. To this end, after providing the required data, utilizing the programming language of Matlab (Version, 2019b), the GWO and BBO metaheuristic algorithms are coupled with an MLP neural network to find the best computational parameters of this tool. More specifically, the mathematical equation of the MLP is introduced as the main problem with the weights and biases as the variables. The mentioned algorithms aim to find the most appropriate values of the weights and biases, which produce the most consistent responses. Then, the optimized ANN is reconstructed by applying the new parameters.

Finding the optimal structure of intelligent models

It is well understood that diverse structures of intelligent models yield distinct results. Therefore, in order to pinpoint the most dependable configurations for the proposed GWO-ANN and BBO-ANN ensembles, a comprehensive process of trial and error was conducted. As explained supra, both GWO and BBO are population-based techniques. Ten different structures, i.e., with ten population sizes varying from 50 to 500 with 50 intervals, were designed and tested within 1000 repetitions. Remarkably, the MSE error criterion was defined as the cost function (CF) to evaluate the accuracy of the models. Over time, the individuals of the optimization algorithms aim to find a more fitted solution, which results in a decrease in error. It is worth noting that to evaluate the repeatability of the used models, each network was performed five times. The convergence curves of the implemented GWO-ANN and BBO-ANN are presented in Fig. 6a,b, respectively.

Figure 6
figure 6

Sensitivity analysis for the (a) GWO-ANN and (b) BBO-ANN.

As the charts illustrate, both GWO and BBO reduced the majority of errors in the first 300 iterations. Comparing the accuracies, it was revealed that the lowest MSE was obtained for the GWO-ANN and BBO-ANN with population sizes of 150 and 450, respectively. However, other population sizes presented a close MSE. Taking into account the computational time, the elite GWO-ANN took about 37.13 min to find the most proper structure of the ANN. This time was obtained about 107.62 min for the best BBO-ANN network.

Model assessment

Then, the MSE and MAE of the prediction responses of the implemented LM-ANN, GWO-ANN, and BBO-ANN in the training and testing stages were calculated to compare their efficiency. The results are shown in Fig. 7a–f. In this figure, the graphical comparison between the actual and predicted WOBs is presented, as well as the calculated error (i.e., indicating the difference between the targets and outputs) and the histogram of the errors.

Figure 7
figure 7

The results obtained for the LM-ANN model in (a) training and (b) testing stages. The results obtained for the GWO-ANN model in (c) training and (d) testing stages. The results obtained for the BBO-ANN model in (e) training and (f) testing stages.

As is seen, all three models performed satisfactorily in both recognizing and predicting the pattern of the WOB influenced by depth, RPM, ROP, and GPM. Moreover, the calculated MSEs indicate that the GWO and BBO algorithms reduced the learning error of the LM-ANN by 14.62% and 24.90% (from 16.96 to 14.48 and 12.73), respectively. Also, the obtained MAEs attest to these results with 7.70% and 10% (from 3.07 to 2.85 and 2.70) decrease in the mean absolute value of the ANN training error, respectively, by applying the GWO and BBO algorithms.

As for the testing phase, it was revealed that the reinforced ANN enjoys more prediction capability (in comparison with the typical one), regarding the observed decrease in the calculated MSEs and MAEs. Accordingly, the testing MSE declined by 9.86% and 9.41% (from 15.52 to 13.99 and 14.06). Additionally, these values were obtained as 2.05% and 5.82% (from 2.92 to 2.86 and 2.75) for the testing MAEs. Furthermore, the cross-plots associated with the estimated WOBs by the LM-ANN, GWO-ANN, and BBO-ANN are presented in Fig. 8a–c, respectively.

Figure 8
figure 8

The cross-plots of (a) LM-ANN, (b) GWO-ANN and (c) BBO-ANN models.

From a comparison point of view, it is deduced that the BBO surpassed the GWO in optimizing the ANN. In other words, the solution provided by the BBO algorithm for solving the mathematical equation of the ANN developed (for estimating the WOB) was more efficient than what was found by the GWO technique.

As mentioned, the BBO-ANN is the most accurate model of this study, but for answering the question “Which one of the neural-based ensembles is more efficient?” the complexity and time-effectiveness should be considered as well. It was shown that the BBO needs more time (about 2.9 times) and a larger population (3 times) compared to GWO to achieve the best solution. Notably, the difference between the learning error (i.e., the MSE) was approximately 2 units. Here, when time or computational ease comes out as determinant factors, utilizing the GWO is a better choice for optimizing the ANN, especially when it has provided satisfying accuracy. However, when the accuracy of the prediction is a more crucial factor, employing the BBO seems more reasonable.

Investigating the effect of input variable dimension

Pearson’s method has previously been applied to sort the strength of WOB dependency on the depth, RPM, ROP, and GPM. This section investigates the effect of different combinations of input variables on both the accuracy and training time of the BBO-ANN model. As Table 3 shows, the BBO-ANN accuracy toward WOB predictions decreases by decreasing the input variable dimension from four (Depth, RPM, ROP, GPM) to two (Depth, RPM). It can also be seen that decreasing the input variable dimension has a slight impact on the BBO-ANN training time. Indeed, the training stage of the BBO-ANN model with four, three, and two input variables takes 31.17, 30.72, and 30.42 min, respectively.

Table 3 Checking the effect of input variable dimension on the BBO-ANN model performance.

Comparison with other techniques

This section compares the BBO-ANN accuracy toward the WOB prediction with MLR37, SVR38, ANFIS39, and GMDH40. Equation (14) introduces the linear correlation between WOB and all four input variables.

$${\text{WOB }} = \, - {16}.0{7 } + \, 0.00{748 } \times {\text{ Depth }} + \, 0.0{262 } \times {\text{ RPM }} - \, 0.{1217 } \times {\text{ ROP }} + \, 0.0{126 } \times {\text{ GPM}}$$
(14)

Table 4 relies on the presented MAEs and MSEs by the BBO-ANN, MLR, SVR, ANFIS, and GMDH in the training and testing steps to compare the models’ accuracy. This comparison analysis clarifies that the BBO-ANN accuracy in both the training and testing stages is better than all the other checked techniques. After the BBO-ANN, the adaptive neuro-fuzzy inference system with a cluster radius of 0.5 is the best intelligent tool to predict WOB from Depth, RPM, ROP, and GPM. This ANFIS model estimates the training dataset with MAE = 2.96 and MSE = 15.51. It also predicts the testing group with MAE = 2.95 and MSE = 15.32. The support vector regression with the linear kernel is also the worst model among the tested tools to estimate the WOB. It has MAE = 3.65 and MSE = 21.48 in the training step and MAE = 3.60 and MSE = 21.42 in the testing step.

Table 4 Comparing the BBO-ANN reliability for the WOB prediction with other techniques.

Extracting the neural predictive formula

Lastly, due to the superiority of the BBO-based ensemble, the governing formula of this network, composed of the ANN optimized parameters, is extracted and presented in Eq. (15). This equation addresses the mathematical rule established in the last layer of the used MLP (i.e., the output layer). Also, the optimized weights and biases of the middle layer (i.e., hidden layer) are presented in Table 5.

$$\begin{aligned} {\text{WOB}} & { = }1.3784 \times {\text{Z}}_{1} + 2.4171 \times {\text{Z}}_{2} + 0.99494 \times {\text{Z}}_{3} \\ & \quad - 1.3479 \times {\text{Z}}_{4} - 2.0371 \times {\text{Z}}_{5} \\ & \quad + 3.0183 \times {\text{Z}}_{6} - 4.1156 \times {\text{Z}}_{7} \\ & \quad - 9.0499 \times {\text{Z}}_{8} - 1.4611 \times {\text{Z}}_{9} + 15.6774 \\ \end{aligned}$$
(15)

where Z1, Z2, …, and Z9 are calculated as follows:

Table 5 Optimized weight and biases of the BBO-ANN model.

Model interpretability

The drilling industry will greatly benefit from applying this novel neural-evolutionary framework for determining WOB in drilling operations. In the upstream oilfield industry, where cost-effective drilling operations are vital, an accurate calculation of WOB is essential for improving drilling processes. The model could represent the complex interactions affecting WOB by including drilling depth, drill string rotational speed, rate of penetration, and volumetric flow rate.

With the use of this knowledge, drilling program decisions may be made with great consideration, resulting in parameters that considerably impact the rate of penetration into the formation. Its adaptability results from the framework's success in predicting WOB under various conditions, including those where mud qualities, drill bit types, and hydraulics are taken into account. Using such sophisticated prediction models can result in better-informed and optimized drilling procedures, which will eventually help the drilling industry succeed as it seeks out cost- and efficiency-saving opportunities.

The application of our study holds tremendous potential for operational companies looking for cutting-edge solutions in the complex and expensive world of drilling operations. Automated drilling rigs, often known as "smart" drilling rigs, are one example. These rigs use cutting-edge technology to automate several drilling operations, such as wellbore navigation and directional drilling41. These systems frequently use real-time data analysis and machine learning to improve drilling circumstances and make decisions. Real-time data analytics tools are also being integrated more and more frequently. These platforms analyze data from various sensors and drilling equipment to deliver quick insights into drilling performance, empowering operators to make decisions based on data immediately42.

Conclusions

The main objective of this study was to investigate the capability of the evolutionary optimization algorithm to accomplish the training process of artificial neural networks to predict the WOB as a function of different combinations of depth, RPM, ROP, and GPM. To this end, grey wolf and biogeography-based optimization algorithms were coupled with the ANN to adjust weights and biases. Based on the executed sensitivity analysis, the BBO-ANN model needed a larger population size and more computational time than the GWO-ANN. The obtained error criteria of MSE and MAE revealed that both applied algorithms perform satisfactorily in enhancing the learning capability of the ANN. Moreover, the BBO surpassed GWO in prevailing the computational drawbacks of the ANN, and also, the BBO-ANN produced the most consistent results, followed by GWO-ANN and LM-ANN. The effect of input variable dimension on the accuracy and training time showed that the BBO-ANN provided better WOB prediction when it was constructed based on all four input variables instead of utilizing those with the highest linear correlation (i.e., depth and RPM or depth, RPM, and ROP). In addition, the constructed BBO-ANN predicted the actual WOB data with better accuracy than the MLR, ANFIS, SVR, and GMDH. The BBO-ANN formula was presented to be used for directly estimating the WOB, influenced by four effective factors, including drilling depth, drill string rotational speed, rate of penetration, and volumetric flow rate.