The present disclosure describes quantitative hydraulic fracturing surveillance from fiber optic sensing using machine learning.
Generally, fiber optic sensing is implemented in a wellbore environment via a logging tool that includes at least one fiber optic sensor for capturing a measurement in the wellbore environment. A fiber optic line is in optical communication with the fiber optic sensor. The data captured by the fiber optic sensor is transmitted through the fiber optic line in real time, and the fiber optic sensor can be a passive sensor not requiring electrical or battery power.
Embodiments described herein enable quantitative hydraulic fracturing surveillance from fiber optic sensing using machine learning. The present techniques include fiber-optic distributed sensing. Fiber optic distributed sensing, such as distributed acoustic sensing (DAS), distributed temperature sensing (DTS) and microseismic, has been increasingly used in unconventional fields for intelligent completions, production monitoring, and optimization. Advancement in fiber-optic distributed sensing technology in the past decades has made it possible to reveal critical operational information in situ and in real time. The analysis of a large volume of fiber-optic sensing data and their association with operation states remains mostly qualitative, correlative and after-fact descriptive. The present techniques include deep learning based methods that directly predict operation states and variables, including the pumping variables, the production flow pressure and rates, and the fracking cluster locations, from all the available fiber-optic measurements. Additionally, the present techniques establish an automated quantitative framework for intelligent completion and production monitoring, with minimal manual interpretation or intervention. When combined with efficient pre-processing of the raw measurement data, this will enable DAS/DTS based field monitoring to improve real-time operation decision making. For example, real time decisions can be made to correct issues discovered based on the predicted data.
DAS measurements tend to accumulate significant amount raw data and even simple spectral analysis can be computationally costly and complex, due to the large channel numbers, high sampling frequency, and long time duration over which the measurements are taken. Traditionally, associating the processed DAS information with operation variables and states of interest are generally qualitative, correlative and after-fact descriptive, requiring significant amount of human intervention and interpretation. This can be cumbersome and even infeasible for applications involving long duration monitoring. Accordingly, the present techniques enable automated and quantitative processing frameworks capable of predicting operation states or variable values directly from fiber optic DAS and/or DTS data, with minimal manual intervention. When combined with efficient pre-processing of the raw measurement data, this DAS/DTS based field monitoring to improve real-time operation decision making in enabled. In some embodiments, time traces of DAS at each depth location are transformed into the time-spectral domain. The magnitude of the transformed data in certain frequency bands most correlated with operation variables of interest are identified and collated with operation states as the so-called frequency band extracted (FBE) signals. The present techniques enable direct modeling from DAS and/or DTS data to hydraulic fracturing characteristics or production variables.
DAS and DTS is used to record vibration and temperature around a fiber, respectively. To determine operational and completion design efficiency, DAS/DTS and microseismic data are monitored during the perforation and the actual hydraulic fracturing pump phases. In some embodiments, the pump data (e.g., slurry rates, pressures) is predicted using the DAS/DTS and microseismic measurement over the monitored stages. The inputs to the deep learning models 108 are the preprocessed data samples 106 partitioned from these measurements and their transformed results. In examples, the DAS and microseismic data are converted into spectrogram and then time segmented during preprocessing 104. The present techniques include, at least in part, three types of deep learning models 108: I) A multimodal ResNet network, which maps time snapshots of these measurement samples to the synced pump data independently; II) a multimodal ResNet followed by convolutional LSTM sequence learning model maps time segments of these measurement samples into the synced hydraulic fracturing flow data; and III) constrained version of I and II by enforcing the prediction to be consistent with the learned relationship between the flow pressure and rates. In examples, the models and their constrained versions are trained over a randomly partitioned subset of samples before applied to the remaining testing samples.
In examples, the models and their constrained versions are trained and tested over a DAS/DTS and microseismic dataset acquired over one hour duration with known flow data during hydraulic fracturing process and production phase. The trained models perform robustly over the testing samples and produce accurate prediction. The LSTM sequence-learning model as described herein produces a smoother and more consistent pressure and slurry rate prediction. The trained deep learning models enable an accurate prediction of the pump data from the fiber optic measurements, and also enable an automated and quantitative way to analyze and predict stage isolation state, cluster locations, and determine the fluid profile. In examples, stages here refer to a specific hydraulic fracking process which starting from bottom of the well, perforates and fracks over a certain well depth range (along the well, not necessarily vertical) before sealing it and moving up to perforate and frack the next depth range above; each one of these depth range is called stage. Stages are different well depth ranges and their operations are also separated over time.
In examples, the input data includes DAS input, DTS input, and microseismic input. In examples, the DAS data is sampled at a sample rate fs,das (e.g. 2 kHz), from a total of Ndas (e.g. 493) DAS channels recorded vibrations along the lateral with a spacing ddas (e.g. of approximately 16.74 ft (5.1 m)). The DAS data can be in sgy files, HDF5, or other data file format. Each of the files store a certain time duration Tdas (e.g. 3000 second long) signal of Ndas (e.g. 493) traces. In examples, the DTS data is recorded at a sample rate of fs,dts (e.g. 30 Hz). The space resolution ddts (e.g. approximately 1 ft or 30.48 cm) during hydraulic fracture stimulation. The DTS data can be from a csv file, an HDF5 file or some other format, where each row represents one depth value with total number of channels N, and each column represent one time point, with the total number of columns Tdts. In examples, the micro-seismic data is sampled at a sample rate of fs,ms (e.g. 2 kHz). A total of Nms (e.g. 36) channels are recorded in the monitoring well. The micro-seismic data can be from sgy files, HDF5, or other data file format, where each file stores a certain time duration Tms (e.g. 3-second) long signal of Nms (e.g. 36) traces.
In examples, the output data includes injection or pumping data output.
Referring again to
For the DAS and micro-seismic data in the offline mode, each file only contains a small segment of data. Data is iteratively read from the files within the determined time window, the data segments are concatenated. DTS, injection, pumping, and production data are also read from files within a determined time window. However, for the DTS injection/pumping or production data, due their significantly smaller data size compared to DAS or microseismic, the entire data set can be directly loaded if each is provided in a single file for the entire duration.
At data processing 104, a workflow is implemented. In examples, the workflow is based on the type of data being processed. An example preprocessing workflow is provided in
Referring again to
Referring again to
Here (ω,) is the spectrogram of the ith channel DAS data at frequency ω and sampled at every Δ seconds for m=1, . . . M. (n) is the window function such as the hamming window or blackman window of length L.
Take the average of the spectral signal from traces that have the strongest signal:
Generally raw DAS data has very high sampling frequency but the most informative frequency band are under certain cutoff frequency f max. If the frequency sampling interval is Δf, then the number of frequency samples is Nf=f max/Δf. The resulting DAS spectrogram is of dimension Nf×M×Ndas, the averaged DAS spectrogram is of dimension Nf×M.
For ease of explanation, the following Example 1 is provided. Example 1, consider a set of Ndas=493 channel DAS data of duration Tdas=3000 s, sampled at fs,das=2 kHz, for spectrogram with a window length L=2000, time step every 10 samples or 5 ms, f max=100 Hz, and Δf=1 Hz. In this example, spectrogram of dimension 101×(200×3000) for a single channel is determined, with a total 493 of these spectrograms. This contributes a significant amount spectral data.
In some embodiments, the DTS data is directly used as the network input without the Fourier transformation. For example, the DTS data in the study range is averaged along the measured depth, resulting in one dimensional (1D) vectors as the input data. The DTS data is recorded in a much lower frequency and compared to the DAS/microseismic data, so the DTS is linearly interpolated to have the same time resolution as the DAS and micro-seismic spectrograms. In examples, linear interpolation is applied to the DTS data such that is has the same number of time points as the Short-time Fourier transform (STFT) of the DAS and micro-seismic data. Then the DTS data according to the depths of chosen traces is determined, the average value computed as follows:
In embodiments, the DTS vector is expanded to 2D matrices to be consistent with the DAS and micro-seismic spectrograms. This is done by the outer product:
DTS2D=υ ⊗ DTS1D
where DTS1D∈Nt is the DTS data after the linear interpolation, υ∈Nf is a randomly initialized vector and Nf is the number of frequency points of the STFT results. After the outer product, the DTS data has size of Nf×Nt. The outer product vector v will be updated during the training process.
Referring again to
Referring again to
The DTS data, and the output data will be sampled at the same rate as the DAS and microseismic spectrograms, synchronized, before they are broken into snapshot of the same length every time step of Δ seconds which, after converting into images as described above, will yield sample images of the same dimension and numbers as that of DAS above.
During preprocessing 104 of
During preprocessing 104 of
In the Example I above, with 60%, 10% and 30% ratio this will generate 1800 training samples, 300 validation samples, and 900 testing samples in the i.i.d. case. Data samples 106 are illustrated in
In examples, a ResNet is a network based on a structure called a “residual block” that uses skip connections, or shortcuts to jump over some layers, as shown by the residual block 800A. In particular, a skip connection 802 bypasses Layer I-2, directly connecting Layer I and Layer I-2. The ResNet extracts features from the input data. In examples, the ResNet-18 is a network based on a residual block with a total of eighteen layers. Multiple ResNets can be defined and used as well such as ResNet18, ResNet34, ResNet50 and etc.
In
In some embodiments, there are strong temporal dynamics in both the DAS, DTS and microseismic data as governed by the event physics (e.g. pumping, fracturing, injection or production). To explore the temporal dynamics, an RNN model is used to model sequential memories and dynamics, instead of learning the samples independently. Accordingly, in
As shown in
The whole network 920 structure is shown in
Accordingly, in an example, the ResNet+RNN model 920 includes a sequence length S=10. The number of time samples of 1 training sample is T=200. The number of frequency components under 100 Hz is set as F=101. DΘ is the ResNet-18 network without the output layer. A sequence of output features of ResNet-18 is the input of the LSTM-RNN. The LSTM initial state (h0, c0) is trained as parameters, and the last state hs of the LSTM is the prediction of the network.
In some embodiments, training the deep learning networks of
The loss function is defined as the root mean squared error (RMSE) between the predicted output variables and the respective ground truth. The RMSE loss of the prediction:
γi is the ground truth, γip is the network prediction. N is the number of samples in the batch. The training as well as testing performance in terms of RMSE for both models after 12 training epochs is provided in Table I. During training, various combinations of inputs were used, with and without DTS involved. Based on the pre-trained models, one could also retrain the network with different training/testing datasets.
In some embodiments, there are several different workflow combinations using the deep learning models of
In some embodiments, the present techniques enable constrained machine learning prediction. Note that the output quantities the machine learning models are trained to predict are physical properties of the pumping, injecting or production flows, e.g. the pressure and flow rates, and they are governed by the dynamics of the flow regimes and therefore are expected to be constrained rather than completely independent from each other. This has motivated the constrained machine learning prediction where the constraints are learned among the output variables as an additional training step. The learned constraints are then enforced onto the predicted outputs so that they satisfy the learned physics constraints.
In this case the loss of the prediction network is LRMSE. The RMSE loss of the constraint network:
γic is the output from the constraint network. The final loss function becomes:
l=l
RMSE
+αL
RMSE
Ľ
α is adjusted to control the importance of the constraint network. The loss function reverse back to LRMSE when α=0.
In some embodiments, the deep learning neural network is tested and generalized.
In some embodiments, the present techniques enable event localization. For example, for various applications, including hydraulic fracturing profiling, injection or production monitoring, it is important to be able to localize the events triggering or sustaining the measurement signals. For example, the fracking cluster locations, the production inflow distribution along the laterals, and the like. The present techniques integrate DAS, DTS and microseismic data if available, as well as the measured pressure and flow rates data, to identify and localize the events. Localizing the events determines a location, identified my coordinates, distance, or the like, of the event.
In some examples, DAS data, DTS data, and microseismic data measured during the actual hydraulic fracturing pump phases is obtained, and the pump data (e.g., pressure and slurry rate) directly predicted from the DAS/DTS and microseismic measurements over the monitored stages. In some embodiments, microseismic data is optional.
Specifically, the DAS/DTS/microseismic measurements and pumping data are collected from a large number of hydraulic fracture stages together, before randomly partitioning the samples into the training, the validation and the testing subsets. The training and validation sets are used to fit the deep learning models before they are applied to the testing set for performance evaluation.
The DAS and microseismic data are converted into spectral domain segmented over time. As shown in
At block 1602, distributed acoustic sensing (DAS) data, distributed temperature sensing (DTS) data, and microseismic data is captured over monitored stages. At block 1604, operation states and variables are predicted at a respective stage, based on, at least in part, the DAS data, DTS data, or microseismic data. At block 1606, at least one event associated with the predicted operation states and variables at the respective stage is localized. In some embodiments, the predictions are made using machine learning models. Deep learning based models and algorithms are deployed to directly predict pressure and slurry rates during hydraulic fracturing process from measured fiber-optic sensing data sets, including DAS, DTS and optically microseismic data. The deep learning models provide accurate and reliable prediction of these operation variables. When combined with efficient preprocessing of the large volume of fiber optic DAS/DTS data. This will enable and provide the first step towards an automated quantitative framework for intelligence completion and production monitoring, with minimal manual interpretation.
The controller 1700 includes a processor 1710, a memory 1720, a storage device 1730, and an input/output interface 1740 communicatively coupled with input/output devices 1760 (e.g., displays, keyboards, measurement devices, sensors, valves, pumps). Each of the components 1710, 1720, 1730, and 1740 are interconnected using a system bus 1750. The processor 1710 is capable of processing instructions for execution within the controller 1700. The processor may be designed using any of a number of architectures. For example, the processor 1710 may be a CISC (Complex Instruction Set Computers) processor, a RISC (Reduced Instruction Set Computer) processor, or a MISC (Minimal Instruction Set Computer) processor.
In one implementation, the processor 1710 is a single-threaded processor. In another implementation, the processor 1710 is a multi-threaded processor. The processor 1710 is capable of processing instructions stored in the memory 1720 or on the storage device 1730 to display graphical information for a user interface on the input/output interface 1740.
The memory 1720 stores information within the controller 1700. In one implementation, the memory 1720 is a computer-readable medium. In one implementation, the memory 1720 is a volatile memory unit. In another implementation, the memory 1720 is a nonvolatile memory unit.
The storage device 1730 is capable of providing mass storage for the controller 1700. In one implementation, the storage device 1730 is a computer-readable medium. In various different implementations, the storage device 1730 may be a floppy disk device, a hard disk device, an optical disk device, or a tape device.
The input/output interface 1740 provides input/output operations for the controller 1700. In one implementation, the input/output devices 1760 includes a keyboard and/or pointing device. In another implementation, the input/output devices 1760 includes a display unit for displaying graphical user interfaces.
The features described can be implemented in digital electronic circuitry, or in computer hardware, firmware, software, or in combinations of them. The apparatus can be implemented in a computer program product tangibly embodied in an information carrier, for example, in a machine-readable storage device for execution by a programmable processor; and method steps can be performed by a programmable processor executing a program of instructions to perform functions of the described implementations by operating on input data and generating output. The described features can be implemented advantageously in one or more computer programs that are executable on a programmable system including at least one programmable processor coupled to receive data and instructions from, and to transmit data and instructions to, a data storage system, at least one input device, and at least one output device. A computer program is a set of instructions that can be used, directly or indirectly, in a computer to perform a certain activity or bring about a certain result. A computer program can be written in any form of programming language, including compiled or interpreted languages, and it can be deployed in any form, including as a stand-alone program or as a module, component, subroutine, or other unit suitable for use in a computing environment.
Suitable processors for the execution of a program of instructions include, by way of example, both general and special purpose microprocessors, and the sole processor or one of multiple processors of any kind of computer. Generally, a processor will receive instructions and data from a read-only memory or a random access memory or both. The essential elements of a computer are a processor for executing instructions and one or more memories for storing instructions and data. Generally, a computer will also include, or be operatively coupled to communicate with, one or more mass storage devices for storing data files; such devices include magnetic disks, such as internal hard disks and removable disks; magneto-optical disks; and optical disks. Storage devices suitable for tangibly embodying computer program instructions and data include all forms of non-volatile memory, including by way of example semiconductor memory devices, such as EPROM, EEPROM, and flash memory devices; magnetic disks such as internal hard disks and removable disks; magneto-optical disks; and CD-ROM and DVD-ROM disks. The processor and the memory can be supplemented by, or incorporated in, ASICs (application specific integrated circuits).
To provide for interaction with a user, the features can be implemented on a computer having a display device such as a CRT (cathode ray tube) or LCD (liquid crystal display) monitor for displaying information to the user and a keyboard and a pointing device such as a mouse or a trackball by which the user can provide input to the computer. Additionally, such activities can be implemented via touchscreen flat-panel displays and other appropriate mechanisms.
The features can be implemented in a control system that includes a back-end component, such as a data server, or that includes a middleware component, such as an application server or an Internet server, or that includes a front-end component, such as a client computer having a graphical user interface or an Internet browser, or any combination of them. The components of the system can be connected by any form or medium of digital data communication such as a communication network. Examples of communication networks include a local area network (“LAN”), a wide area network (“WAN”), peer-to-peer networks (having ad-hoc or static members), grid computing infrastructures, and the Internet.
While this specification contains many specific implementation details, these should not be construed as limitations on the scope of any inventions or of what may be claimed, but rather as descriptions of features specific to particular implementations of particular inventions. Certain features that are described in this specification in the context of separate implementations can also be implemented in combination in a single implementation. Conversely, various features that are described in the context of a single implementation can also be implemented in multiple implementations separately or in any suitable subcombination. Moreover, although features may be described above as acting in certain combinations and even initially claimed as such, one or more features from a claimed combination can in some cases be excised from the combination, and the claimed combination may be directed to a subcombination or variation of a subcombination.
Similarly, while operations are depicted in the drawings in a particular order, this should not be understood as requiring that such operations be performed in the particular order shown or in sequential order, or that all illustrated operations be performed, to achieve desirable results. In certain circumstances, multitasking and parallel processing may be advantageous. Moreover, the separation of various system components in the implementations described above should not be understood as requiring such separation in all implementations, and it should be understood that the described program components and systems can generally be integrated together in a single software product or packaged into multiple software products.
A number of implementations have been described. Nevertheless, it will be understood that various modifications may be made without departing from the spirit and scope of the disclosure. For example, example operations, methods, or processes described herein may include more steps or fewer steps than those described. Further, the steps in such example operations, methods, or processes may be performed in different successions than that described or illustrated in the figures. Accordingly, other implementations are within the scope of the following claims.
Other implementations are also within the scope of the following claims.
This application claims priority to U.S. Provisional Patent Application No. 63/239,014, filed Aug. 31, 2021, the entire contents of which are incorporated herein by reference.
Number | Date | Country | |
---|---|---|---|
63239014 | Aug 2021 | US |