Combining Global and Local Optimization Techniques for Automatic History Matching Production and Seismic Data
- S. Mantica (ENI-Agip) | A. Cominelli (ENI-Agip) | G. Mantica (INFN-INFM, U. dell'Insubria)
- Document ID
- Society of Petroleum Engineers
- SPE Journal
- Publication Date
- June 2002
- Document Type
- Journal Paper
- 123 - 130
- 2002. Society of Petroleum Engineers
- 5.5.8 History Matching, 5.1.5 Geologic Modeling, 5.1.9 Four-Dimensional and Four-Component Seismic, 5.6.3 Deterministic Methods, 5.5 Reservoir Simulation, 5.6.9 Production Forecasting
- 2 in the last 30 days
- 305 since 2007
- Show more detail
- View rights & permissions
|SPE Member Price:||USD 12.00|
|SPE Non-Member Price:||USD 35.00|
Gradient-based optimization techniques are increasingly adopted by oil industries for computer-aided history matching because of the great timesaving they can offer over conventional trial-anderror approaches. However, these methods lead to the identification of a single set of parameters, thus neglecting the inherent nonuniqueness of the solution of the underlying inverse problem. In this paper we propose a new approach that couples a chaotic sampling of parameter space with a local minimization technique. Through the evolution of a nonlinear dynamical system, we identify several points to be successively used as initial guesses for a local gradient-based optimizer. This provides a series of alternative matched models with different production forecasts that improve the understanding of the possible reservoir behaviors. The validity of this approach has been proven on a synthetic reservoir derived from a real west Africa field.
The search of an optimal set of simulation parameters to match reservoir production and time-lapse seismic data is usually a significant problem. Typically, it requires the minimization of a nonconvex, least-squares objective function in a parameter space populated by many local minima. Starting from a physically reasonable point, gradient-based methods1,2 allow a fast descent to the closest minimum. The drawback of this approach is potentially twofold. First, an unsatisfactory reduction in the cost function may result. This could only be improved with a redefinition of some of the initial parameters. Second, even when the match is acceptable, only a single forecast scenario is produced.
The nonconvex nature of the history-match problem can be conceptually better tackled using stochastic global optimization techniques, in which parameter space is explored by randomly generated trajectories until a satisfactory minimum is reached. In this framework, entrapment around local minima is avoided by ad hoc hill-climbing rules. A typical approach of this kind is the well-known simulated annealing3 (SA) method, in which the uphill moves are accepted in accordance with a thermally driven Metropolis rule.4 A cooling of the fictitious statistical mechanical system defined in this way guarantees the convergence toward the global minimum.
Many authors have already proposed global minimization approaches, either stochastic or deterministic, in the field of reservoir simulation (see Ouenes et al.5 for a review). In particular, Simulated Annealing,6 the Tunneling Method,7 Genetic Algorithms,8 and hybrid approaches9 all seem to be very promising.
Unfortunately, however, global convergence - even to an approximation of the solution - usually requires a huge number of iterations. As a matter of fact, this price is often too high for the reservoir history-match problem, in which the computation of the objective function is mostly expensive.
Intuitively at least, some sort of mating between local and global techniques should lead to a reasonable compromise between the slow convergence of the latter and the fast, but locally trapped, nature of the former. We might think of generating some (short) trajectories in parameter space, which would hopefully get close to some minima of the history-match problem. A reasonably selected sample of points on these trajectories could then be used to start a good local gradient-based optimizer. Certainly, this is not an entirely new idea; many minimization libraries employ the technique of multiple starting points. What is new in our approach is the attempt to improve dynamically - and not stochastically - on the "quality" of these starting points. The way in which we hope to achieve this goal will be evident in a moment.
SA, surely the first candidate as global counterpart in a coupled approach, is in our opinion inadequate for our aim, at least in the known implementations. At the beginning of the evolution, when the annealing temperature is high, SA samples parameter space merely as an undirected random walk. It is only the slow descent of the temperature that drives the path toward the global minimum at later times. But even so, the Markovian property of SA implies that all the information on the structure of the objective function, obtained all along the minimization path, is discarded. This waste of information is critical when the number of function evaluations is considered a key issue for the problem under study.
We then propose a totally different approach: the initial seeds for the local minimization should be provided by a nonlinear dynamical system, driven by the values of the objective function. This method, originally developed in the context of a fractal reconstruction problem,10 has been applied to simple synthetic history- match problems in a previous paper.11
A trajectory is generated in parameter space according to a nonlocal rule, by which the point xn+1 is obtained as a function, B?, of all the previously generated points
In the above, ? = an external parameter, which will be defined later on. Of course, B? is constructed also taking into account the values of the objective function f in these points. As it stands, B? is a deterministic process, as is the gradient method. The difference is that we do not want this process to aim straight at the supposed minimum, but rather we would like it to wander in space, spending larger spans of time where the objective function is lower, thereby learning its structure, and effecting better and better guesses of where the global minimum might lie. These guesses should then provide us with the initial points of the gradient technique.
We propose to carry out such a complex task via a "chaotic" system, that is, one with trajectories that have maximal algorithmic complexity, and which can therefore encode the structure of the objective function. Within these bounds, the actual implementation, and the definition of the rule B?, is a matter of problem engineering. We present in this paper a particular realization of this idea that has performed satisfactorily on the problem at hand.
|File Size||1 MB||Number of Pages||8|