Professional Documents
Culture Documents
DOI 10.1007/s00521-016-2721-x
ORIGINAL ARTICLE
123
Neural Comput & Applic
methods are better at handling the nonlinearities situations, of the LEA model are shown in Sect. 2; the details of the
they do not perform well at times while attempting to solve modelling procedures are described in Sect. 3; numerical
highly nonlinearity time series data [19]. The ones used for experiments and analysis are shown in Sect. 4, and con-
multivariate regression are often used to build static mod- clusions are drawn in Sect. 5.
els, which are only input–output boxes or black boxes for
the influence factors and the petroleum production. But not
a single petroleum production system is static as it is well 2 The Arps decline model and its linear extension
known that even if any influence factor does not change,
the petroleum production will still decline along with time. 2.1 The Arps decline model
Thus, these models are also limited in applications.
Above all, the decline curve methods and the intelligent The classical decline curve equation was presented by Arps
methods have their own advantages and disadvantages, and [1] as follows:
this research is aiming at building a new prediction model 1 dqðtÞ
for the petroleum production by combining the advantages ¼ aqðtÞd ð1Þ
qðtÞ dt
of the decline curve methods and the intelligent methods.
The main advantage of the decline curve methods is that it where a and d are empirical constants to be determined
reflects the overall declining trend of the petroleum pro- based on production data. When d ¼ 0, Eq. (1) degenerates
duction and that of the intelligent methods is their ability to to an exponential decline model, and when d ¼ 1, Eq. (1)
deal with the nonlinearities. To combine these advantages, yields a harmonic decline model. When 0\d\1, Eq. (1)
we need to use the so-called kernel trick. derives a hyperbolic decline model. These three types are
The kernel trick used in this paper is similar to the least not only applicable to analyse the oil and gas wells per-
squares support vector machines (LS-SVM), which was formance, but also available to predict the oil and gas
introduced by Suykens [20]. The LS-SVM employs the reservoir performance.
ridge regression form [21] considering the equality con-
straints instead of the inequalities of the standard SVM of 2.2 The linear extension of Arps decline model
Vapnik [22, 23]. In recent researches, many classical linear
methods have been transformed into nonlinear ones based Obviously, the production is influenced by the other factors,
on both the LS-SVM formulation [24–26] and the standard such as the pressure difference for the oil and gas wells, the
SVM formulation [27–29], and the LS-SVM formulation water injections and the number of active production wells
has been presented to be easier to use as it only involves for the oil or gas fields. If these factors are considered, one
solving a set of linear equations while the standard SVM can add a linear combination of them into Eq. (1) with d ¼ 0
formulation involves solving a quadric programming (the exponential decline equation), which is
problem using the gradient-based training methods, which dqðtÞ
often suffer from many local minima and high computa- ¼ aqðtÞ þ b1 u1 ðtÞ þ b2 u2 ðtÞ þ þ bN uN ðtÞ þ c
dt
tional cost [30]. ð2Þ
The main idea of this paper starts with extending the
Arps equation for exponential decline to a multivariate where the ui ðtÞ represents the variable corresponding to the
ordinary equation by adding the linear combination of the influence factor, i.e. the pressure difference, number of
input parameters, and its discrete form becomes a linear active oil or gas wells and amount of water injections, bi
recursive multivariate model, which is called the linear are the system parameters to be determined, and c is a bias
extension of the Arps model (LEA). The kernel trick is term. The corresponding difference equation of Eq. (2) can
then employed to transform this linear model to a nonlinear be written as
prediction model, which is called the nonlinear extension qðkÞ ¼ aqðk 1Þ þ b1 u1 ðkÞ þ b2 u2 ðkÞ þ þ bN uN ðkÞ þ c:
of the Arps model (NEA). This model is different from the
past kernel regression models, as the one-step linear ð3Þ
recursion is known, while the past models are totally black- The Eq. (3) is called the linear extension of Arps decline
box models, such as the LS-SVM regression [31–33], model, which is abbreviated as LEA model. Given a
Recursive LS-SVM [34–36]. Being similar to the LS-SVM, sample set as fu1 ðkÞ; u2 ðkÞ; . . .; uN ðkÞ; qðkÞgðk ¼ 1; 2;
the solution of the NEA also follows directly from solving . . .; nÞ, the parameters of LEA in (3) can be obtained using
a set of linear equations, which makes the new model to be the least squares method as follows:
easy-to-use. 1
The rest of this paper is organized as follows: a brief ½a; b1 ; b2 ; . . .; bN ; cT ¼ BT B BT Y; ð4Þ
overview of the Arps decline model and the representation
123
Neural Comput & Applic
where
The nonlinear function g is difficult to determine, and in
0 1
qð2Þ this paper we use a method by which we can find an
B qð3Þ C equivalent expression of g.
B C
Y ¼B C; One defines the nonlinear mapping
@ ... A
qðnÞ u: RN ! F ; ð10Þ
123
Neural Comput & Applic
123
Neural Comput & Applic
123
Neural Comput & Applic
Table 1 continued
Time Production Injection Time Production Injection Time Production Injection
1
In this case, we are interested in forecasting the oil pro-
The source code are available at http://cn.mathworks.com/matlab
duction using the historical data of the water injection and
central/fileexchange/58918-the-source-code-of-the-kernel-regular
ized-extension-of-the-arps-decline-model-knea-. the oil production. The water injection is regarded as the
123
Neural Comput & Applic
3
4
8
4
0 50 100 150 200 250
Time
influence factor. According to the reservoir engineering, it 4.1.2 Case study 2: predicting the oil production of an oil
is difficult to determine the relationship between the water field in Cambay Basin in India
injection and the oil production due to the complex
underground conditions [39]. In this study, we will build In this case, we will build the LEA model and the NEA
the prediction models based on the LEA model and the model using the raw data from [40]2 as shown in Table 3,
NEA model, respectively. which contains five input series and one output series. The
The raw data are collected from the Block-1 in Huabei raw data shown in Table 3 are the smoothed data of the
oil field, which is located in north China. This block was monthly production of five oil wells and the cumulative
developed in January 1992, and its oil production started production of an oil field in Cambay Basin in India, where
to decline since 1994. In the same period, the oil field the five input series are corresponding to the monthly
company started to develop the water injection wells in production of five oil wells, and the output is corresponding
order to enhance the oil production. The data set contains to the cumulative production of this oil field. And it was
227 pairs of the historical data, in which the first 200 reported by Chakra et al. [40] the relationship between the
pairs are used to build the prediction models, and the last input and output presented to be highly nonlinear.
27 ones are used for testing the effectiveness of these The raw data contain 51 groups, the first 34 groups are
two models (Table 1). used to build the prediction models, and the rest 17 groups
In this case, we set the parameters r ¼ 0:1 and c ¼ 10 are used for validation. In this case, we set the parameters
for the NEA model. The evaluation criteria are shown in r ¼ 0:5 and c ¼ 100 for the NEA model. The evaluation
Table 2, and the predicted series by LEA and NEA along criteria are shown in Table 4, and the predicted series by
with the production data are plotted in Fig. 1. LEA and NEA along with the production data are plotted in
It can be seen in Table 2 that the all the training error Figs. 2 and 3, respectively.
and the testing error of NEA are much smaller than those The training error and testing error of LEA are very
of LEA, which indicates that the NEA is much more large as shown in Table 4, which indicates that the LEA
accurate than the LEA in predicting the oil production. model can not describe the nonlinear relationship between
The LEA model can only roughly describe the overall the input and output series; thus, it is not available to
tendency of the oil production as shown in Fig. 1, and predict the cumulative production in this case. But the
the predicted values by LEA are quite far away to the oil errors of NEA are small, which indicates that the NEA is
production points. The predicted values by NEA are very efficient to predict the cumulative production of this oil
close to the oil production points, and the NEA can also field. It can be clearly seen in Figs. 2 and 3 that the pre-
describe the tendency of the oil production accurately as diction results of LEA are much larger than the real values
shown in Fig. 1. The LEA model can not perform well of cumulative production, and the results of NEA are very
because the relationship between the oil production and close to the real values. In summary, the NEA model still
water injection is not linear, and on the other hand the outperforms the LEA in this case.
NEA model, which can describe the nonlinear relation-
ship between the inputs and outputs, can perform well in
this study. 2
The raw data are listed in the Table 4 in [40].
123
Neural Comput & Applic
Table 3 Smoothed data of monthly production of the oil field in Table 3 continued
Cambay Basin in India [40]
Months Input-1 Input-2 Input-3 Input-4 Input-5 Target
Months Input-1 Input-2 Input-3 Input-4 Input-5 Target
49 0.047 0.028 0.567 0.046 0.029 0.786
1 0.111 0.072 0.309 0.09 0.237 0.775 50 0.048 0.048 0.577 0.05 0.035 0.8
2 0.107 0.068 0.303 0.084 0.233 0.758 51 0.048 0.068 0.579 0.05 0.041 0.827
3 0.106 0.065 0.299 0.081 0.224 0.752
4 0.099 0.068 0.3 0.075 0.216 0.752
5 0.099 0.064 0.304 0.074 0.211 0.745 Table 4 RMSPE (%) by NEA
Model Training Testing
6 0.101 0.061 0.309 0.073 0.207 0.736 and LEA in case study 2
7 0.091 0.059 0.319 0.074 0.202 0.722 NEA 3.9983 4.2216
8 0.081 0.057 0.326 0.078 0.195 0.708 LEA 192.5 8878.3
9 0.074 0.048 0.324 0.082 0.195 0.685
10 0.062 0.044 0.326 0.085 0.191 0.676
4.2 Performances of NEA with different parameters
11 0.05 0.04 0.327 0.081 0.187 0.67
r and c
12 0.051 0.036 0.332 0.065 0.192 0.637
13 0.057 0.032 0.322 0.069 0.189 0.624
In this subsection, we will present two experiments to show
14 0.061 0.028 0.307 0.072 0.168 0.628
the impact of the kernel parameter r and the regularized
15 0.069 0.027 0.295 0.079 0.154 0.621
parameter c.
16 0.074 0.026 0.291 0.089 0.147 0.601
17 0.078 0.025 0.281 0.1 0.137 0.622
4.2.1 Results by NEA with different kernel parameters r
18 0.074 0.023 0.279 0.091 0.133 0.625
19 0.073 0.023 0.297 0.082 0.147 0.616 This experiment is carried out to test the impact of the
20 0.068 0.022 0.311 0.065 0.158 0.641 kernel parameter r for the Gaussian kernel (24) to the
21 0.063 0.022 0.312 0.061 0.158 0.671 performance of the NEA model. The raw data in the
22 0.058 0.023 0.35 0.06 0.15 0.675 Table 1 are used in this experiment. The parameter r is set
23 0.055 0.025 0.389 0.059 0.143 0.663 to be 0.01, 0.1, 1, 10, 100, and c is set to be c ¼ 10,
24 0.049 0.025 0.41 0.062 0.129 0.659 respectively. The RMSPEs for training and testing are
25 0.043 0.024 0.421 0.065 0.111 0.612 listed in Table 5 and also plotted in Fig. 4. It can be seen
26 0.041 0.024 0.437 0.057 0.1 0.554 that the training error is growing with larger r, while the
27 0.04 0.024 0.402 0.051 0.094 0.571 testing error is decreasing. The predicted values by NEA
28 0.038 0.027 0.357 0.045 0.086 0.598 with r being 0.01, 1, 100 are plotted in Figs. 5, 6 and 7,
29 0.045 0.031 0.363 0.048 0.084 0.618 respectively. It is shown that the predicted series are closer
30 0.052 0.037 0.376 0.051 0.082 0.621 to the oil production with smaller r, and the predicted
31 0.055 0.043 0.386 0.058 0.076 0.675 series are flatter with larger r. In this experiment, the NEA
32 0.055 0.042 0.393 0.058 0.073 0.714 model with best testing accuracy is the one with r ¼ 0:1.
33 0.054 0.04 0.444 0.062 0.075 0.738 This implies that the optimal kernel parameter is often
34 0.048 0.04 0.493 0.057 0.076 0.77 obtained in the mid-size values.
35 0.042 0.038 0.53 0.052 0.078 0.824
36 0.036 0.031 0.574 0.046 0.083 0.747
37 0.035 0.029 0.624 0.048 0.088 0.722 4.2.2 Results by NEA with different regularized
38 0.033 0.025 0.566 0.041 0.083 0.704 parameters c
39 0.033 0.021 0.548 0.038 0.083 0.685
40 0.034 0.017 0.534 0.036 0.083 0.67 This experiment is carried out to test the impact of the
41 0.03 0.017 0.528 0.033 0.076 0.756 regularized parameters c to the performance of NEA
42 0.03 0.02 0.519 0.033 0.067 0.746 model. The raw data in the Table 1 are also used. The
43 0.034 0.024 0.59 0.04 0.067 0.725 parameter c is set to be 0.01, 0.1, 1, 10, 100, and r is set to
44 0.033 0.027 0.586 0.04 0.06 0.712 be r ¼ 0:1, respectively. The RMSPEs for training and
45 0.034 0.029 0.579 0.04 0.043 0.713 testing are listed in Table 6 and also plotted in Fig. 8. It
46 0.041 0.03 0.569 0.041 0.031 0.711 can be seen that the training error is decreasing with larger
47 0.042 0.026 0.57 0.042 0.033 0.716
c, but the testing error is not monotonic. The details of the
48 0.043 0.022 0.57 0.043 0.033 0.757
results can be seen more clearly in the Figs. 9, 10 and 11,
123
Neural Comput & Applic
100
50
0
0 10 20 30 40 50 60
Month
0.7
0.65
Table 5 RMSPE by NEA with which describe the plots of the predicted values. The
r Training Testing smaller c is corresponding to the smoother predicted series,
different kernel parameters r
0.01 1.5990 4.9007 and larger c leads to more accurate training results. The
0.1 3.9983 4.2216 results are coincidence with the optimization problem (15).
1 7.1460 4.9339 The optimal performance of NEA is obtained with c ¼ 10,
10 8.8670 4.5510 which is also not too large or too small, which implies that
100 11.9723 2.9737 the optimal c would also be obtained in the mid-size
values.
Training
8
RMPSE (%)
6
Testing
0
−2 −1.5 −1 −0.5 0 0.5 1 1.5 2
lg(σ)
123
Neural Comput & Applic
4
0 50 100 150 200 250
Time
4
0 50 100 150 200 250
Time
9
3
4
0 50 100 150 200 250
Time
123
Neural Comput & Applic
Table 6 RMSPE by NEA with which can describe the nonlinear relationship between
c Training Testing
different regularized parameters the system inputs and output. It has been shown in the
c 0.01 16.3535 7.6375 case studies that the NEA outperforms the linear
0.1 13.8975 6.4731 extension of Arps decline model (LEA), which indi-
1 7.8722 4.7022 cated that the NEA was eligible to be applied in the
10 3.9982 4.2216 nonlinear forecasting problems in the petroleum
100 2.1993 4.9803 engineering.
The impact of the kernel parameter r and the regular-
ized parameter c has also been tested in this study. The
5 Conclusions results show that the smaller r makes the NEA model to be
more nonlinear, and the larger r makes the NEA model to
In this paper, we proposed a novel prediction model for approach to a linear model. And the NEA model produced
the petroleum production based on the Arps decline smoother points with larger c, and it was represented to be
model and the kernel method, which is called the NEA more accurate in training with larger c. These results all
model. The introduction of the kernel method can indicated that the optimal r and c would be obtained in the
extend the Arps decline model to a nonlinear model mid-size values.
18
16
14
Training
12
RMPSE (%)
10
Testing
2
−2 −1.5 −1 −0.5 0 0.5 1 1.5 2
lg(γ)
9
3
4
4
0 50 100 150 200 250
Time
123
Neural Comput & Applic
4
0 50 100 150 200 250
Time
4
0 50 100 150 200 250
Time
The limitations of this study should also be noticed that Compliance with ethical standards
we have not proposed an available algorithm to compute
Conflict of interest All authors declare that they have no conflict of
the optimal parameters r and c. In fact, selection of these interest.
two parameters is an open problem in the machine learning
researches, and there is no available algorithm to select the
optimal parameters up to now. But it will be interesting and References
might be possible to find an optimal interval which con-
tains the optimal parameters r and c, and once this interval 1. Asps JJ (1945) Analysis of decline curves. Trans Am Inst Min
is found, it will be more computationally efficient to select Metall Pet Eng 160:228–247
2. Li K, Horne RN et al (2003) A decline curve analysis model
the optimal parameters for the NEA model. based on fluid flow mechanisms. In: SPE western regional/AAPG
pacific section joint meeting. Society of Petroleum Engineers
Acknowledgements The authors thank the editors and anonymous 3. Velazquez Camacho R, Fuentes-Cruz G, Vasquez-Cruz MA et al
referees for their useful comments and suggestions, which have (2008) Decline-curve analysis of fractured reservoirs with fractal
helped to improve this paper. This work was supported by the Doc- geometry. SPE Reserv Eval Eng 11(03):606–619
toral Research Foundation of Southwest University of Science and
Technology (No. 16zx7140).
123
Neural Comput & Applic
4. Sureshjani MH, Gerami S (2011) A new model for modern 20. Suykens JAK, Vandewalle J (1999) Least squares support vector
production-decline analysis of gas/condensate reservoirs. J Can machine classifiers. Neural Process Lett 9(3):293–300
Pet Technol 50(7–8):10–23 21. Golub GH, Van Loan CF (2012) Matrix computations, vol 3. JHU
5. Ling K, Jun H et al (2012) Theoretical bases of Arps empirical Press, Baltimore
decline curves. In: Abu Dhabi international petroleum conference 22. Cortes C, Vapnik V (1995) Support-vector networks. Mach Learn
and exhibition. Society of Petroleum Engineers 20(3):273–297
6. Mohaghegh SD et al (2005) Recent developments in application 23. Harris D, Burges CJC, Kaufman L, Smola A, Vapnik V (1997)
of artificial intelligence in petroleum engineering. J Pet Technol Support vector regression machines. Adv Neural Inf Process Syst
57(04):86–91 9:155–161
7. Sun Q, Liebmann D, Nejad A, Bansal Y, Ayala LF, Balogun O, 24. Suykens JAK, Van Gestel T, Vandewalle J, De Moor B (2003) A
Suleen F, Ertekin T, Karpyn Z et al (2013) Forecasting well support vector machine formulation to PCA analysis and its
performance in a discontinuous tight oil reservoir using artificial kernel version. IEEE Trans Neural Netw 14(2):447–450
neural networks. In: SPE unconventional resources conference— 25. Alzate C, Suykens JAK (2008) A regularized kernel CCA con-
USA. Society of Petroleum Engineers trast function for ICA. Neural Netw 21(2):170–181
8. Aizenberg I, Sheremetov L, Villa-Vargas L, Martinez-Muñoz J 26. Xu J-H, Zhang X-G, Li Y-D (2004) Regularized kernel forms of
(2016) Multilayer neural network with multi-valued neurons in minimum squared error methods. Acta Autom Sin 30(1):27–36
time series forecasting of oil production. Neurocomputing 27. Schölkopf B, Smola A, Müller K-R (1997) Kernel principal
175:980–989 component analysis. In: Artificial neural networks ICANN’97.
9. Theodosiou M (2011) Disaggregation & aggregation of time Springer, pp 583–588
series components: a hybrid forecasting approach using gener- 28. Scholkopft B, Mullert K-R (1999) Fisher discriminant analysis
alized regression neural networks and the theta method. Neuro- with kernels. Neural Netw Sig Process IX:1
computing 74(6):896–905 29. Tipping M (2003) Relevance vector machine. US patent
10. Sheremetov L, Cosultchi A, Martı́nez-Muñoz J, Gonzalez-Sán- 6,633,857, 14 Oct 2003
chez A, Jiménez-Aquino MA (2014) Data-driven forecasting of 30. Wang H, Hu D (2005) Comparison of SVM and LS-SVM for
naturally fractured reservoirs based on nonlinear autoregressive regression. In: International conference on neural networks and
neural networks with exogenous input. J Pet Sci Eng brain, 2005. ICNN&B’05, vol 1. IEEE, pp 279–283
123:106–119 31. De Brabanter K (2011) Least squares support vector regression
11. Zhong Y, Zhao L, Liu Z, Xu Y, Li R (2010) Using a support with applications to large-scale data: a statistical approach. PhD
vector machine method to predict the development indices of thesis, Katholieke Universiteit Leuven
very high water cut oilfields. Pet Sci 7(3):379–384 32. Hou L, Yang Q, An J (2009) An improved LSSVM regression
12. Rammay MH, Abdulraheem A et al (2014) Automated history algorithm. In: International conference on computational intelli-
matching using combination of adaptive neuro fuzzy system gence and natural computing, 2009. CINC’09, vol 2. IEEE,
(ANFIS) and differential evolution algorithm. In: SPE large scale pp 138–140
computing and big data challenges in reservoir simulation con- 33. Cao S-G, Liu Y-B, Wang Y-P (2008) A forecasting and fore-
ference and exhibition. Society of Petroleum Engineers warning model for methane hazard in working face of coal mine
13. Knabe SP, Goel H, Al-Jasmi AK, Rebeschini J, Querales MM, based on LS-SVM. J China Univ Min Technol 18(2):172–176
Adnan FMd, Rivas F, Rodriguez JA, Carvajal GA, Villamizar M 34. Suykens JAK, Vandewalle J (2000) Recurrent least squares
et al (2013) Building neural-network-based models using nodal support vector machines. IEEE Trans Circuits Syst I: Fundam
and time-series analysis for short-term production forecasting. In: Theory Appl 47(7):1109–1114
SPE middle east intelligent energy conference and exhibition. 35. Xie J (2009) Time series prediction based on recurrent LS-SVM
Society of Petroleum Engineers with mixed kernel. In: Asia-pacific conference on information
14. Gupta S, Fuehrer F, Jeyachandra BC et al (2014) Production processing, 2009. APCIP 2009, vol 1. IEEE, pp 113–116
forecasting in unconventional resources using data mining and 36. Qu H, Oussar Y, Dreyfus G, Xu W (2009) Regularized recurrent
time series analysis. In: SPE/CSUR unconventional resources least squares support vector machines. In: International joint
conference—Canada. Society of Petroleum Engineers conference on bioinformatics, systems biology and intelligent
15. Harris C et al (2014) Potential pitfalls in exploration and pro- computing, 2009. IJCBS’09. IEEE, pp 508–511
duction applications of machine learning. In: SPE western north 37. Smola AJ, Schölkopf B (2004) A tutorial on support vector
American and rocky mountain joint meeting. Society of Petro- regression. Stat Comput 14(3):199–222
leum Engineers 38. Espinoza M, Suykens JAK, De Moor B (2005) Kernel based
16. Al-Fattah SM, Startzman RA et al (2001) Predicting natural gas partially linear models and nonlinear identification. IEEE Trans
production using artificial neural network. In: SPE hydrocarbon Autom Control 50(10):1602–1606
economics and evaluation symposium. Society of Petroleum 39. Nield DA, Bejan A (2013) Mechanics of fluid flow through a
Engineers porous medium. Springer, Berlin
17. Al-Fattah SM, Startzman RA et al (2001) Neural network 40. Chakra NC, Song K-Y, Gupta MM, Saraf DN (2013) An inno-
approach predicts us natural gas production. In: SPE production vative neural forecast of cumulative oil production from a pet-
and operations symposium. Society of Petroleum Engineers roleum reservoir employing higher-order neural networks
18. Nakutnyy P, Asghari K, Torn A et al (2008) Analysis of water- (HONNs). J Pet Sci Eng 106:18–33
flooding through application of neural networks. In: Canadian
international petroleum conference. Petroleum Society of Canada
19. Tokar AS, Johnson PA (1999) Rainfall-runoff modeling using
artificial neural networks. J Hydrol Eng 4(3):232–239
123