skip to main content

Title: Application of Machine Learning Methods to Well Completion Optimization: Problems with Groups of Interactive Inputs.

In unconventional reservoirs, optimal completion controls are essential to improving well productivity and reducing costs. In this article, we propose a statistical model to investigate associations between shale oil production and completion parameters (e.g., completion lateral length, total proppant, number of hydraulic fracturing stages), while accounting for the influence of spatially heterogeneous geological conditions on hydrocarbon production. We develop a non-parametric regression method that combines a generalized additive model with a fused LASSO regularization for geological homogeneity pursuit. We present an alternating augmented Lagrangian method for model parameter estimations. The novelty and advantages of our method over the published ones are a) it can control or remove the heterogeneous non-completion effects; 2) it can account for and analyze the interactions among the completion parameters. We apply our method to the analysis of a real case from a Permian Basin US onshore field and show how our model can account for the interaction between the completion parameters. Our results provide key findings on how completion parameters affect oil production in that can lead to optimal well completion designs.

more » « less
Award ID(s):
Author(s) / Creator(s):
; ; ;
Date Published:
Journal Name:
SPE Annual Technical Conference and Exhibition
Medium: X
Sponsoring Org:
National Science Foundation
More Like this

    The near-surface seismic structure (to a depth of about 1000 m), particularly the shear wave velocity (VS), can strongly affect the propagation of seismic waves and, therefore, must be accurately calibrated for ground motion simulations and seismic hazard assessment. The VS of the top (<300 m) crust is often well characterized from borehole studies, geotechnical measurements, and water and oil wells, while the velocities of the material deeper than about 1000 m are typically determined by tomography studies. However, in depth ranges lacking information on shallow lithological stratification, typically rock sites outside the sedimentary basins, the material parameters between these two regions are typically poorly characterized due to resolution limits of seismic tomography. When the alluded geological constraints are not available, models, such as the Southern California Earthquake Center (SCEC) Community Velocity Models (CVMs), default to regional tomographic estimates that do not resolve the uppermost VS values, and therefore deliver unrealistically high shallow VS estimates. The SCEC Unified Community Velocity Model (UCVM) software includes a method to incorporate the near-surface earth structure by applying a generic overlay based on measurements of time-averaged VS in top 30 m (VS30) to taper the upper part of the model to merge with tomography at a depth of 350 m, which can be applied to any of the velocity models accessible through UCVM. However, our 3-D simulations of the 2014 Mw 5.1 La Habra earthquake in the Los Angeles area using the CVM-S4.26.M01 model significantly underpredict low-frequency (<1 Hz) ground motions at sites where the material properties in the top 350 m are significantly modified by the generic overlay (‘taper’). On the other hand, extending the VS30-based taper of the shallow velocities down to a depth of about 1000 m improves the fit between our synthetics and seismic data at those sites, without compromising the fit at well-constrained sites. We explore various tapering depths, demonstrating increasing amplification as the tapering depth increases, and the model with 1000 m tapering depth yields overall favourable results. Effects of varying anelastic attenuation are small compared to effects of velocity tapering and do not significantly bias the estimated tapering depth. Although a uniform tapering depth is adopted in the models, we observe some spatial variabilities that may further improve our method.

    more » « less

    Analysis of tectonic and earthquake-cycle associated deformation of the crust can provide valuable insights into the underlying deformation processes including fault slip. How those processes are expressed at the surface depends on the lateral and depth variations of rock properties. The effect of such variations is often tested by forward models based on a priori geological or geophysical information. Here, we first develop a novel technique based on an open-source finite-element computational framework to invert geodetic constraints directly for heterogeneous media properties. We focus on the elastic, coseismic problem and seek to constrain variations in shear modulus and Poisson’s ratio, proxies for the effects of lithology and/or temperature and porous flow, respectively. The corresponding nonlinear inversion is implemented using adjoint-based optimization that efficiently reduces the cost function that includes the misfit between the calculated and observed displacements and a penalty term. We then extend our theoretical and numerical framework to simultaneously infer both heterogeneous Earth’s structure and fault slip from surface deformation. Based on a range of 2-D synthetic cases, we find that both model parameters can be satisfactorily estimated for the megathrust setting-inspired test problems considered. Within limits, this is the case even in the presence of noise and if the fault geometry is not perfectly known. Our method lays the foundation for a future reassessment of the information contained in increasingly data-rich settings, for example, geodetic GNSS constraints for large earthquakes such as the 2011 Tohoku-oki M9 event, or distributed deformation along plate boundaries as constrained from InSAR.

    more » « less
  3. This paper explores an inverse approach to the problem of characterizing sediment sources' (“source” samples) age distributions based on samples from a particular depocenter (“sink” samples) using non-negative matrix factorization (NMF). It also outlines a method to determine the optimal number of sources to factorize from a set of sink samples (i.e., the optimum factorization rank). We demonstrate the power of this method by generating sink samples as random mixtures of known sources, factorizing them, and recovering the number of known sources, their age distributions, and the weighting functions used to generate the sink samples. Sensitivity testing indicates that similarity between factorized and known sources is positively correlated to 1) the number of sink samples, 2) the dissimilarity among sink samples, and 3) sink sample size. Specifically, the algorithm yields consistent, close similarity between factorized and known sources when the number of sink samples is more than ∼3 times the number of source samples, sink data sets are internally dissimilar (cross-correlation coefficient range >0.3, Kuiper V value range >0.35), and sink samples are well-characterized (>150–225 data points). However, similarity between known and factorized sources can be maintained while decreasing some of these variables if other variables are increased. Factorization of three empirical detrital zircon U–Pb data sets from the Book Cliffs, the Grand Canyon, and the Gulf of Mexico yields plausible source age distributions and weights. Factorization of the Book Cliffs data set yields five sources very similar to those recently independently proposed as the primary sources for Book Cliffs strata; confirming the utility of the NMF approach. The Grand Canyon data set exemplifies two general considerations when applying the NMF algorithm. First, although the NMF algorithm is able to identify source age distribution, additional geological details are required to discriminate between primary or recycled sources. Second, the NMF algorithm will identify the most basic elements of the mixed sink samples and so may subdivide sources that are themselves heterogeneous mixtures of more basic elements into those basic elements. Finally, application to a large Gulf of Mexico data set highlights the increased contribution from Appalachian sources during Cretaceous and Holocene time, potentially attributable to drainage reorganization. Although the algorithm reproduces known sources and yields reasonable sources for empirical data sets, inversions are inherently non-unique. Consequently, the results of NMF and their interpretations should be evaluated in light of independent geological evidence. The NMF algorithm is provided both as MATLAB code and a stand-alone graphical user interface for Windows and macOS (.exe and .app) along with all data sets discussed in this contribution. 
    more » « less

    Seismic tomography is a cornerstone of geophysics and has led to a number of important discoveries about the interior of the Earth. However, seismic tomography remains plagued by the large number of unknown parameters in most tomographic applications. This leads to the inverse problem being underdetermined and requiring significant non-geologically motivated smoothing in order to achieve unique answers. Although this solution is acceptable when using tomography as an explorative tool in discovery mode, it presents a significant problem to use of tomography in distinguishing between acceptable geological models or in estimating geologically relevant parameters since typically none of the geological models considered are fit by the tomographic results, even when uncertainties are accounted for. To address this challenge, when seismic tomography is to be used for geological model selection or parameter estimation purposes, we advocate that the tomography can be explicitly parametrized in terms of the geological models being tested instead of using more mathematically convenient formulations like voxels, splines or spherical harmonics. Our proposition has a number of technical difficulties associated with it, with some of the most important ones being the move from a linear to a non-linear inverse problem, the need to choose a geological parametrization that fits each specific problem and is commensurate with the expected data quality and structure, and the need to use a supporting framework to identify which model is preferred by the tomographic data. In this contribution, we introduce geological parametrization of tomography with a few simple synthetic examples applied to imaging sedimentary basins and subduction zones, and one real-world example of inferring basin and crustal properties across the continental United States. We explain the challenges in moving towards more realistic examples, and discuss the main technical difficulties and how they may be overcome. Although it may take a number of years for the scientific program suggested here to reach maturity, it is necessary to take steps in this direction if seismic tomography is to develop from a tool for discovering plausible structures to one in which distinct scientific inferences can be made regarding the presence or absence of structures and their physical characteristics.

    more » « less
  5. Abstract

    Shock waves in geological materials are characterized by a sudden release of rapidly expanding gas, liquid, and solid particles. These shock waves may occur due to explosive volcanic eruptions or be artificially triggered. In fact, underground explosions have often been used as an engineering solution for large‐scale excavation, stimulating oil and gas recovery, creating cavities for underground waste storage, and even extinguishing gas field fires. As such, hydrocodes capable of simulating the rapid and significant deformation under extreme conditions can be a valuable tool for ensuring the safety of the explosions. Nevertheless, as most of the hydrocodes are often formulated in an Eulerian grid, this setting makes it non‐trivial to track the deformation configuration of the materials without a level set. The objective of this paper is to propose the use of the material point method equipped with appropriate equation of state (EOS) models as a hydrocode suitable to simulate underground explosions of transverse isotropic geomaterials. To capture the anisotropic effect of the common layered soil deposits, we introduce a new MPM hydrocode where an anisotropic version of the Mie‐Gruneisen EOS is coupled with a frictional Drucker‐Prager plasticity model to replicate the high‐strain‐rate constitutive responses of soil. By leveraging the Lagrangian nature of material points to capture the historical dependence and the Eulerian calculation of internal force, the resultant model is capable of simulating the rapid evolution of geometry of the soil as well as the high‐strain‐rate soil mechanics of anisotropic materials.

    more » « less