The present invention relates generally to oil exploration and production. More particularly, the present invention relates to using pattern recognition in combination with geological, geophysical and engineering data processing, analysis and interpretation for hydrocarbon exploration, development, or reservoir management on digital computers.
Many disciplines can benefit from pattern recognition. Disciplines where the benefit is greatest share characteristics and needs. Some common characteristics include large volumes of data, anomalous zones of interest that are mixed together with a large number of similar non-anomalous zones, timeframes too short to allow rigorous manual examination, and anomalies that manifest themselves in many ways, no two of which are exactly the same. Analysis of the data is usually done by highly trained professionals working on tight time schedules. Examples of these disciplines include, but are not limited to, hydrocarbon exploration and medical testing.
Exploring for hydrocarbon reservoirs is a very competitive process. Decisions affecting large amounts of capital investment are made in a time-constrained environment based on massive amounts of technical data. The process begins with physical measurements that indicate the configuration and selected properties of subsurface strata in an area of interest. A variety of mathematical manipulations of the data are performed by computer to form displays that are used by an interpreter, who interprets the data in view of facts and theories about the subsurface. The interpretations may lead to decisions for bidding on leases or drilling of wells.
A commonly used measurement for studying the subsurface of the earth under large geographical areas is seismic signals (acoustic waves) that are introduced into the subsurface and reflected back to measurement stations on or near the surface of the earth. Processing of seismic data has progressed hand-in-hand with the increased availability and capabilities of computer hardware. Calculations performed per mile of seismic data collected have increased many-fold in the past few years. Display hardware for observation by a human interpreter has become much more versatile.
When an interpreter makes decisions from the seismic, and other, data it is used with some knowledge of geology of the area being investigated. The decisions involve identification, analysis, and evaluation of the geological components of an oilfield, which include the presence of a reservoir rock, presence of hydrocarbons, and the presence of a container or trap. The rationale for the decisions that were made was based on both the geologic information and the data. That rationale is not generally documented in detail for seismic data analysis due to the large amount of data and information being analyzed. Therefore, it is difficult to review the history of exploration decisions and repeat the decision process using conventional procedures. The relative importance attached to the many characteristics shown in the seismic data and known from the geology is a subjective value that does not become a part of the record of the exploration process.
It is recognized that seismic data can also be used to obtain detailed information regarding producing oil or gas reservoirs and to monitor changes in the reservoir caused by fluid movement. Description of neural network modeling for seismic pattern recognition or seismic facies analysis in an oil reservoir is described, for example, in “Seismic-Pattern Recognition Applied to an Ultra Deep-Water Oilfield,” Journal of Petroleum Technology August, 2001, page 41). Time-lapse seismic measurements for monitoring fluid movement in a reservoir are well known. The fluid displacement may be caused by natural influx of reservoir fluid, such as displacement of oil by water or gas, or may be caused by injection of water, steam or other fluids. Pressure depletion of a reservoir may also cause changes in seismic wave propagation that can be detected. From these data, decisions on where to drill wells, production rates of different wells and other operational decisions may be made. The neural network technique usually assumes that all significant combinations of rock type are known before analysis is started so that they can be used as a training set. This assumption is usually acceptable when analyzing fully developed fields but breaks down when only a few or no wells have been drilled. Common implementations of the neural network technique usually assume selection of the location of the geology of interest is an input that is determined prior to the analysis and often selects it using an analysis gate of fixed thickness. As the geology of interest is not always well known, the geology of interest should be a product of the analysis, not an input. Moreover, geology of interest rarely has a fixed thickness. The thickness varies significantly as the depositional process varies from place to place, sometimes by an amount that is sufficient to significantly degrade the result of the neural network analysis. This form of analysis includes information extraction and information classification in a single step that has little of no user control.
What is needed is a way to perform unsupervised pattern analysis that does not require a learning set, does not require texture matching, does not classify attributes of a single spatial size, and does not require a-priori knowledge of the location of the geology of interest. Unsupervised pattern analysis requires feature, pattern, and texture extraction from seismic data where the features, patterns, and texture measurements are well chosen for optimal classification and can be interpreted in terms of oilfield components. Optimal means that they:
There is further a need in the art to have a process of creating features, patterns and textures, from data plus a data hierarchy recognizing the relative levels of abstraction along with a pattern database containing all of the information.
From a production standpoint, there is a need in the art to visually classify this information to analyze the interior of a hydrocarbon reservoir more effectively. Direct hydrocarbon indicators should be visually identifiable. Seismic stratigraphy should be performed in a way that includes visual classification of all the seismic stratigraphic information available in the data. In addition the knowledge inherent in the visual classification needs to be captured in a template, stored in a template library, and reused later in an automatic process.
While 3D seismic produces images of structures and features of the subsurface of the earth over very large geographical areas, it does not interpret those images. A trained geoscientist or specialist performs the interpretation. Unfortunately, reliance upon a relatively few qualified individuals increases the cost of the interpretation process and limits the number of interpretations that can be made within a given period. This makes current seismic interpretation techniques impractical for the analysis of the very large volumes of seismic data that are currently available. As a result of the large and growing amount of available data, there is a need in the art for a knowledge capture technique where the information in the 3D seismic data that the specialist looks at is captured by a pattern recognition process. Ideally, the pattern recognition process would be repeated for large amounts of data in a screening process, with the results displayed in an intuitive manner so that the specialist can quickly perform quality control on the results, and correct noise induced errors, if any.
There is further a need in the art for a way to auto-track textures, patterns, and features in order to isolate and measure rock bodies or objects of interest. Preferably, an object should be auto-tracked so that its location is determined both by the properties of its interface with surrounding objects and by the difference between the features, patterns, and textures in the objects interior when compared to those outside the object. This tracks the object directly rather than tracking the object solely based on the varying properties of the interface which, by itself, is unlikely to be as descriptive of the object. Interface tracking tracks the object indirectly, as would be accomplished with boundary representations. An example of automatically detecting objects based on their interior and interface characteristics would be in colorectal cancer screening where the target anomaly (a colorectal polyp) has both distinctive interface and interior characteristics.
Moreover, a data analysis specialist should not be required to rely on analysis of non-visual measures of object characteristics. The information describing the visual characteristics of seismic data should be stored in a way that allows the data specialist to interact with the information to infer and extract geological information and to make a record of the exploration process. Finally, a way should be provided to analyze geologic information with varying levels of abstraction.
The above-identified needs are shared across many disciplines yet the specific nature and the characteristics of the anomalies vary across disciplines and sometimes within a single problem. Thus there is a need for a common method of analysis that is capable of being applied to a wide variety of data types and problems, yet it is capable of being adapted to the specific data and problem being solved in situations where required.
The present invention solves many of the shortcomings of the prior art by providing an apparatus, system, and method for synthesizing known (raw) data into hyperdimensional templates, storing the templates of the known data in a pattern database (“PDB”). The subject data to be analyzed (the target data) is similarly synthesized, and the two sets of templates can be compared to detect desirable characteristics in the subject body. The comparison process is enhanced by the use of specially adapted visualization applications that enable the operator to select particular templates and sets of templates for comparison between known templates and target data. The visualization technique facilitates the visual recognition of desired patterns or indicia indicating the presence of a desired or undesired feature within the target data. The present invention is applicable to a variety of applications where large amounts of information are generated. These applications include many forms of geophysical and geological data analysis including but not limited to 3D seismic.
The processing technique of the present invention generates a result through a series of reduction steps employing a cutting phase, an attribute phase, and a statistics phase. These three phases can be conducted at least once or numerous times over a series of layers as the data is further reduced. Normally, there is an input data layer upon which a cut, an attribute and a statistics processes are imposed to form a feature layer. From a feature layer, the same cut/attribute/statistics process is implemented to form other layers such as a pattern layer and a texture layer. These series of steps, each of which employ the cut/attribute/statistics processes form a hyper-dimensional template akin to a genetic sequence for the particular properties of the localized region within the input data. The input data for each cut/attribute/statistics phase may be taken from another layer (above and/or below) or it may be taken directly from the raw data, depending upon the problem being solved.
The hyper-dimensional templates of the known data are stored essentially to a set of patterns that are stored in a database hence the term “pattern database.” When an operator desires to analyze a set of data, she selects the analysis sequences that they feel would provide the best indication of finding the desired characteristics within the target set of data. The operator would then perform the series of sequences on the target data to obtain a target set of templates. The operator then thereafter would make a comparison of the target set of templates to a set of known hyper-dimensional templates stored within the pattern database. The operator can then employ what is called a binding strength to the various templates and allow the patterns of the known data to seek out and essentially adhere to the similar patterns in the target data. Once the similar patterns are identified, i.e., the desired patterns from the known data are surmised through affinity to the patterns of the target data is a simple matter to back out the physical location of those desired characteristics in a target data using, for example, the a specially developed visualization application.
In general, the present invention performs several steps related to pattern analysis. The process extracts geological information and places it in an information hierarchy called a pattern pyramid. The process includes the extraction of features, using a specific methodology, and then using another methodology computes patterns and textures from that feature base. The patterns are a transformation, which classifies features based on their spatial organization. The same transformation, when applied to patterns, can form the texture of the data to further facilitate recognition by the operator. The process also performs classification of the information in the pattern hierarchy and segmentation through auto-tracking creating objects. The objects are placed in a collection called a scene. The decision surfaces used for the classification are captured in a template that is stored for later reuse. Finally, the process allows the operator to interactively choose classification parameters and dissect objects through visual classification.
The present invention can accept large amounts of information and convert the data into features, patterns, and textures (that are usually stored and displayed as voxels). Classifying the feature, pattern, and texture information requires using a collection of information called a hyperdimensional fragment to classify multiple measurements for the same spatial location (voxel).
The fragments may be further characterized by string length of the fragment between, for example, the zero crossing point of a reference axis. The string length may be further characterized by determining certain variations on the string length, such as a reference string length, or a string length ratio. All of the string length-related values can be compared to other string length-related values, or compared to other information as explained below. Such comparisons can be very useful in determining the presence (or absence) of desired substances such as hydrocarbons in underground formations.
However, it is not simply what is accomplished, but of equal importance is how it is accomplished and how the intermediate and final products are organized and stored. Specifically, it is the order in which the tools of the method of the present invention are used that provides the great benefits of the present invention. In general, the method of the present invention first assembles the data. Thereafter an abstraction hierarchy of, features, patterns, and textures are generated in the given order. All of the levels classify results from previously computed levels in three steps of one dimensional fragment cutting, attribute (feature, pattern, or texture) calculation (usually a form of cluster analysis) and a statistical analysis of the attributes. The pattern database is evaluated by a process of visual classification or comparison to a template containing previously captured interpretation knowledge creating objects which are placed in a collection called a scene. Then an interpreter (typically a human) reviews the objects to determine if desirable (or undesirable) objects are present. Once the objects of interest have been identified they are stored along with the pattern information in a manner that allows the information to be effectively accessed by spatial location and visualized.
The present invention makes extensive use of templates for knowledge capture. Templates are feature, pattern, and texture decision surfaces that are used by the associated classifiers to find like structures. Known patterns found in templates can then be compared, in an automated fashion, to new data to detect similar patterns and hence find the desired features in the new data. The templates also contain all of the processing and display parameters required to start with an initial data set and create a final product in a batch data computer run without human intervention.
In this disclosure, string length is used in association with a fragment (the region between two zero-crossings of a seismic trace). Used in this manner, the string length becomes a type of descriptor of that fragment. Calculating string length in this way allows the analyst to do three things. First, the analyst can compute the string length ratio that is an even more powerful descriptor of a fragment. Second, associating string length and string length ratio with a fragment, as a feature, allows the analyst to combine that measurement for a fragment with like measures from surrounding fragments to form patterns involving string length and string length ratio. Third, the analyst can combine the string length and/or string length ratio features and/or patterns with other features and patterns (derived independently, perhaps with alternate methods) within a system to help identify target geology and to produce geobodies from the seismic data.
Referring now to the drawings, the details of the preferred embodiments of the present disclosure are schematically illustrated.
a is a diagram of the pattern pyramid and associated levels of abstraction according to the teachings of the present disclosure.
b is a diagram of an example of a pattern pyramid for data with three spatial dimensions according to the teachings of the present disclosure.
c is a diagram of the pattern pyramid, an example of the components within each level, plus an example of a hyperdimensional fragment according to the teachings of the present disclosure.
d is a diagram of an example of feature level fragment cuts for a band-limited acoustical impedance trace according to the teachings of the present disclosure.
e is a diagram of an example of pattern level fragment cuts for a band-limited acoustical impedance trace according to the teachings of the present disclosure.
a is a flowchart illustrating an embodiment of a method of 3D seismic first pass lead identification.
b is a flowchart illustrating an embodiment of a method of building a pattern database for geophysical and geological data.
a is a flowchart illustrating an embodiment of a method of building a pattern database for 3D band-limited acoustical impedance.
b is a flowchart illustrating an embodiment of a method of preparing seismic for pattern analysis.
a and 5b is a flowchart illustrating an embodiment of a method of constructing a pattern database for 3D band-limited acoustical impedance.
a, 6b, 6c and 6d are flowcharts illustrating an embodiment of a method of fragment cutting and feature attribute and statistic computation.
a, 7b and 7c are flowcharts illustrating an embodiment of a method of pattern attribute and statistic calculation.
a is a plot of band-limited acoustical impedance as a function of time or distance.
b is a representative plot of broadband acoustical impedance as a function of time or distance, according to the present disclosure.
c is a mathematical expression for computing the RMS amplitude feature for a fragment, according to the present disclosure.
d is a mathematical expression for computing the shape feature for a fragment, according to the present disclosure.
a is a mathematical expression for computing the Horizontal Complexity feature statistic, according to the present disclosure.
b is the definition of a coordinate neighborhood for horizontal complexity and feature and feature function anisotropy feature statistics, according to the present disclosure.
a defines the values, M and □, of feature and feature function anisotropy, according to the present disclosure.
b is an example of feature and feature function anisotropy, according to the present disclosure.
c is an example of no feature and feature function anisotropy, according to the present disclosure.
d is a mathematical expression for computing M and □ for feature and feature function anisotropy, according to the present disclosure.
a is a diagram of pattern space, according to the present disclosure.
b is a diagram showing example fragment lengths of 3 and 3, according to the present disclosure.
c is a diagram showing pattern space for a pattern computed using a fragment length of 3, according to the present disclosure.
d is a diagram showing a multi-feature pattern space, according to the present disclosure.
a is a diagram of a two-dimensional pattern space with pattern locations computed as M and □, according to the present disclosure.
b is mathematical expression for computing M and □, according to the present disclosure.
c is a diagram of a three-dimensional pattern space with pattern locations computed as M, □, □, according to the present disclosure, according to the present disclosure.
d is a mathematical expression for computing M, □, □, according to the present disclosure.
a and 17b are diagrams illustrating two waveforms according to the present disclosure.
The present invention may be susceptible to various modifications and alternative forms. Specific embodiments of the present invention are shown by way of example in the drawings and are described herein in detail. It should be understood, however, that the description set forth herein of specific embodiments is not intended to limit the present invention to the particular forms disclosed. Rather, all modifications, alternatives, and equivalents falling within the spirit and scope of the invention as defined by the appended claims are intended to be covered.
The present disclosure includes a system for and method of extracting, organizing, and classifying features, patterns, and textures from a data set. The data and the information extracted there from, is organized as a pattern hierarchy and stored in a pattern database. The present disclosure also provides a system for the segmentation and the analysis of geological objects, for example, by identifying, extracting, and dissecting the best estimate of hydrocarbon filled reservoir rocks from band-limited acoustical impedance (“RAI”) data computed from 3D seismic data or, if available, broadband acoustical impedance (“AI”) computed from 3D seismic data, stacking velocities, well logs, and user supplied subsurface structural models. In addition the present disclosure includes a system for capturing the knowledge of the geoscientists operating the present disclosure in templates and reusing the templates for automated mining of large volumes of data for additional geological objects.
Pattern Recognition of Geoscience and Geological Data
The first step of the pattern recognition method of the present disclosure is feature extraction. Feature extraction comes in many forms, and tends to be specific to the type of problem encountered. For example, in seismic data analysis, geological features are extracted. Most traditional methods of feature extraction for seismic data involve mathematical algorithms that focus on the measurements of the sound rather than on the visual appearance of the displayed data. Most geophysicists, however, think of geology in a visual way, which makes analysis and interpretation of traditionally extracted seismic signal features difficult. Many other examples and uses for the feature extraction and imaging technique of the present disclosure will be apparent upon examination of this specification.
In general, a mathematical representation of features describes the local state of a system. The features are then represented as a vector in an abstract vector space or tangent space called the feature state space. The axes of the state space are the degrees of freedom of the system, in this case the features of the image. To minimize the amount of information required to represent the state of the image it is preferred that the features, axes of the state space, be linearly independent. The features have the capacity to “span the signal,” or to describe all seismic attributes such that, for example, a geophysicist could accurately re-create the underlying geology.
Using seismic data as an example, geological features are extracted for performing pattern recognition on a seismic data set. Feature descriptors of seismic data tend to be one-dimensional, measuring only one aspect of the image, such as measuring only properties of the signal at specific locations in the signal. These feature descriptors taken singly do not yield enough information to adequately track geology. The relationship these measurements have with their local neighbors contains information about depositional sequences that is also very important geological information. Thus, the relationship features have with their neighbors and the total data set also needed to be analyzed.
The present disclosure utilizes a hierarchical data structure called a pattern pyramid that is stored in a pattern database (“PDB”). The pattern database employs a process that is based on DNA-like pseudo sequencing to process data and places the information into a pattern database. This database contains the data plus features and their relationship with the data and, in addition, information on how the features relate with their neighbors and the entire data set in the form of pattern, textures, and related statistics.
Intuitively the basic concept of the pattern pyramid is that complex systems can be created from simple small building blocks that are combined with a simple set of rules. The building blocks and rules exhibit polymorphism in that their specific nature varies depending on their location or situation, in this case the data being analyzed and the objective of the analysis. The basic building block used by the present disclosure is a fragment sequence built from a one-dimensional string of data samples. A pattern pyramid is built using fragment sequences (simple building blocks) and an abstraction process (simple rules). The specific definition of the building blocks, cutting criteria, exhibits polymorphism in that the algorithm varies depending on the data being analyzed and the goal of the analysis. Similarly, the abstraction process exhibits polymorphism in that the algorithm depends on the data being analyzed and the goal of the analysis.
A pattern database is built for known data, which functions as a reference center for estimating the locations in the target data that are potential hydrocarbon deposits. The estimation is accomplished by building a pattern database for the target data using the same computations as for the known data and comparing the pattern databases. The pattern pyramids have several levels of abstraction that may include features, patterns, and textures. The pattern pyramids are built using an abstraction process. The step of comparing the pattern databases is performed by defining a hyperdimensional fragment that associates the appropriate pattern information in the pattern database to the specific data samples from which they were computed. Classification of the target data into portions that match the known data and portions that do not match is accomplished by searching through the hyperdimensional fragments of the target data and comparing them to the hyperdimensional fragments for the known data (the classifier) to identify matches. Intuitively this means that for the target data to match the known data at any location not only do the data values need to agree but the data values must also be a part of local features, patterns and textures that also agree adequately. Thus, the present disclosure not only performs pattern recognition, but also is capable of performing feature recognition, texture recognition, and data comparison all at the same time as required for solving the problem.
To allow for natural variation or noise in the data, exact matches do not have to be required. This is accomplished by defining a binding strength or an affinity that allows hyperdimensional fragments that are reasonably similar but not exactly the same to be classified as matched. The hyperdimensional fragment selected by the geoscientist operating the present disclosure captures the operators' knowledge of what is a desirable outcome, or in other words what a hydrocarbon filled reservoir looks like.
The hyperdimensional fragments and associate abstraction process parameters can be saved as a template into a template database. One or more templates can be checked out from the library and applied to large volumes of target data to identify targets. Targets that have been segmented out of the data set are stored as objects in a collection of objects called a scene. The objects, along with additional data the geoscientist adds to them, become a list of drilling opportunities.
Oil Exploration & Production Uses
This disclosure is capable of being used for geological, geophysical and engineering data processing, analysis and interpretation for hydrocarbon exploration, development, or reservoir management. It supports application for a variety of types of geophysical data. The present disclosure is flexible and extensible allowing adaptation to provide solutions to many geoscientific problems.
For example, the present disclosure is capable of being used to analyze 3D seismic target data set with the goal of identifying the drilling target locations that represent potential hydrocarbon bearing reservoir rock. An ideal path to reaching this goal is to directly locate and analyze the hydrocarbons in reservoirs. Experience has shown that geology is diverse and complex and geophysical tools (other than drilling) do not directly measure the existence of hydrocarbons. Thus, oil finders build a set of corroborating evidence to decrease risk and increase the probability of drilling success, where success is defined as locating profitable hydrocarbon accumulations. Accomplishing this involves using several forms of geological and geophysical analysis, the goal of which is to identify sufficient evidence of the three basic components of an oil field, which are a reservoir, a charge, and a trap. Identifying a reservoir involves collecting evidence of the existence of a rock having the property that it is capable of holding sufficient hydrocarbons (adequate porosity) and the property that allows the hydrocarbons to be removed from the earth (adequate permeability). Identifying a charge involves collecting evidence that a hydrocarbon is present in the reservoir rock (bright spots, fluid contacts, and others). Another way is to identify a source rock that is or was expelling hydrocarbons and a hydrocarbon migration path to the trap. Identifying a trap involves collecting evidence that the earth structure and/or stratigraphy forms a container in which hydrocarbons collect forming structural traps, stratigraphic traps, or a combination of the two. When the identification of a reservoir, charge, and trap are complete the result is called a lead. After a full analysis of the reservoir, charge, and trap plus risk analysis, economic analysis, and drilling location selection the lead becomes a prospect that is ready to be drilled. The probability of success is highest when there is strong evidence that a reservoir, charge, and trap all exist, that they exist in the same drillable location, and that they can be profitable exploited. Our objective is to construct a pattern recognition process and associated tools that identify a location with all of the constituent parts of a lead and to quantify them to covert a lead into a prospect.
When it is applied to 3D seismic, the present disclosure identifies a potential reservoir through feature analysis, identifies hydrocarbon indications through pattern and texture analysis, and identifies the presence of a depositional process that deposits reservoir rock though texture analysis. It is also capable of identifying the presence of a trap by determining the presence of stratigraphic sequences that create stratigraphic traps through texture analysis and the determining the presence of structural trapping components through fault identification by edge identification. In addition it is capable of identifying the presence of a charge by locating stratigraphic sequences capable of expelling hydrocarbons through feature, pattern, and texture analysis plus determining the presence of faults in the neighborhood through fault identification. The final step of associating and validating the three components of an oil field is usually accomplished by a geoscientist.
After a lead has been identified the pattern database, along with appropriate visualization, could be used to perform reservoir dissection. This is a study of the internal characteristics of the reservoir to estimate the economics and convert the lead into a prospect.
After an oil field has been discovered the present disclosure is capable of being used to improve reservoir characterization, which is the estimation of rock properties (rock type, porosity, permeability, etc.), and fluid properties (fluid type, fluid saturations, etc.). Rock types and properties are a function of the geologic process that deposited them. In addition to information about the rock's acoustical impedance, the local features, patterns and textures contain information about depositional processes. Thus, the rock type and property estimations can be improved by including the feature, pattern, and texture information while estimating them.
In addition to the above seismic analysis methods, the present disclosure could be used for portions of data processing. Examples include but are not limited to, automatic stacking velocity picking, automatic migration velocity picking, noise identification, and noise muting.
The present disclosure is also capable of performing data set registration and comparison by successively aligning textures, patterns, and features. When applied to seismic it includes registering shear data to compressional data, registering 4D seismic data, registering angle stacks for AVA analysis, and others.
3D Seismic First Pass Lead Identification
This example performs first pass lead identification through simultaneous identification of a potential reservoir through feature analysis, identification of hydrocarbon indications through pattern and texture analysis, and identification of the presence of a depositional process, that deposits reservoir rock, though texture analysis. One way to do this is to use a known data set, which represents a successful lead or example lead, and compare the target data to known data. For this example, the goal is to identify reservoirs that occur in all forms of traps. Thus, it is preferable to disassociate the structural aspects of the earth from the stratigraphic, rock property, and hydrocarbon indication aspects. During this analysis, the structural aspects are not used. After the potential lead is identified using this example, the existence of a trap and charge will be determined.
For 3D seismic lead identification, the overall process starts by building a pattern database with successive levels of abstraction (features, patterns, and textures) for the known data. After the pattern database building process has been applied to a set of known data, and the minimum set of attributes that characterize the known data has been identified, the pattern database is applied to a set of data to be analyzed (the “target data”). The data of each set are subjected to the same series of steps within the abstraction process.
Before or during the comparison, an affinity or binding strength is selected by the operator that determines how closely the known data has to match the target data to result in a target being identified. The binding strength helps to identify features, patterns, and textures in the target data that adequately match, but do not exactly match, the desired features, patterns, and textures in the known data.
Next the pattern database for the known data is compared to that of the target data. The comparison is performed by identifying a hyperdimensional fragment from the known data pattern database that adequately and reasonably uniquely characterizes the known data. This hyperdimensional fragment relates the data at the location where the hydrocarbons were found, or were expected to be found, to the set of features, patterns, and textures that were derived from it. The hyperdimensional fragment and associated abstraction process parameters can be combined into a template. Templates can be used immediately or stored in a template database on one or more mass storage devices, and then retrieved when needed.
When templates are applied to target data sets the resulting targets are identified. These targets are stored as objects which represent leads. The leads objects are the locations in the target data sets which have a potential reservoir identified through feature analysis, potential hydrocarbon indications identified through pattern and texture analysis, and the potential presence of a depositional process that deposits reservoir rock identified though texture analysis. A collection of objects are stored in a scene. The scene represents the physical locations of the leads identified by the present disclosure in this example. Geological and other required properties of the leads can be stored with them.
Because the nature of the reservoirs, and possibly the hydrocarbons trapped in them, varies across each data set due to natural geological variations, it is often necessary to create more than one template to identify all of the leads any given area offers. A collection of templates can be created and stored in a template database. These may be sequentially applied to one or many target data sets in a process called data mining. When multiple templates are applied to the same target data set, the results are several scenes each containing lead objects. The scenes and their associated objects, one scene from each template, can be combined by performing Boolean operations on the scenes containing the objects creating one composite scene.
3D Seismic Pattern Pyramid
A 3D seismic data set exemplary embodiment of the layers of abstraction associated with the method of the present disclosure is illustrated in
In the exemplary embodiment, the pattern pyramid 100 contains three layers of abstraction above the data level 108 (see
The pattern pyramid shown in
The pattern database building process identifies the minimum set of attributes (features, patterns, and textures) of one or several examples of known data so that, when the known data is compared to the target data, only the desired characteristics need to be considered. The results of each step are represented in the pattern pyramid 130 as shown in
Geophysical & Geological Data
The data in the foundation of the pattern database can by any type of a variety of geophysical and geological data types. The data types include many forms of indirect and direct measurements. Direct measurements involve obtaining physical samples of the earth by mining or drilling. Indirect measurements include active and passive data gathering techniques. Passive techniques involve studying naturally occurring signals or phenomena in the earth such as magnetic field variations, gravitational field variations, electrical field variations, sound (such as naturally occurring micro-seismicity or earthquakes), and others. Active measurements involve introducing signals or fields into the earth and measuring the returns including magneto-telluric, seismic, and others. Active and passive measurements are acquired on the surface of the earth and in wells. These include but are not limited to seismic, electrical, magnetic, and optical data. It is capable of being applied to data sets with any number of spatial dimensions, usually one, two, or three dimensions. It also works on higher dimension data. Examples include, but not limited to, 4D pre-stack seismic cubes containing offset data, 3D pre-stack cubes containing all of the offsets for a 3D seismic line, 4D seismic cubes containing multiple angle stacks, 4D seismic taken at different calendar dates, combinations of these, or others.
When applied to seismic data, the wave propagation types include but are not limited to compressional, shear, combinations and other types. The seismic can be in the form of pre-stack and post-stack data or both. It can be as acquired (raw) or processed. It can also include modified seismic data including but not limited to acoustical impedance computed by a seismic inversion. If the goal is to study AVO or AVA effects reflection coefficient data of elastic impedance data may be used.
Each data sample has at least, but is not limited to, one data value. An example of a single data value at a sample includes, but is not limited to, the band-limited acoustic impedance information obtained from seismic data. An example of a sample with multiple data values includes, but is not limited to, multi-component seismic.
When the goal of the analysis is seismic interpretation of 3D seismic data, the properties of the geological layers need to be studied instead of the properties of their boundaries where reflections occur. The preferred, but not only, way to accomplish this is by analyzing an acoustical impedance cube with the broadest possible bandwidth that can be reliably created by seismic inversion. The analysis can be band-limited acoustical impedance computed from reflection data. The analysis can also be broadband acoustical impedance computed from seismic data plus well log data, and/or seismic stacking velocities, and/or seismic migration velocities, and/or operator constructed models. For the lead identification example, the technique is applied to 3D seismic data that has been inverted creating a band-limited acoustical impedance 3D voxel cube.
PDB Abstraction—Cutting
The first step of the abstraction process, for each pattern pyramid level, is to cut fragments. Each fragment is a one-dimensional interval that has a physical length and physical location. It corresponds to an associated fragment sequence that is a sequence of data, attribute, or statistics values from a lower layer in the pattern pyramid.
In the most computationally efficient embodiment of the present disclosure, pre-defined or operator-supplied cutting criteria are applied to the data to generate the fragments. The specific cutting criteria that are applied for cutting can be a function of the problem, of the data being analyzed, or both. The cutting criteria can include, for example, fixed spatial length cuts, cuts derived from lower level pattern pyramid information, or cuts determined from a user supplied example.
Some forms of geophysical and geological data are amenable to using fixed-length fragments, and the present disclosure can easily accommodate fixed-length fragments. Fixed length fragments associate a fragment with a fixed number of data samples.
For band-limited acoustical impedance the most common cutting criteria are to use cuts derived from the information in any lower level of the pattern pyramid. For example, feature cutting criteria is a function of the data values. Pattern cutting criteria can be a function of the feature level cuts, feature level attributes, feature level statistics, or data values. In this case the cutting criteria remains constant for the level while the underlying data typically varies, with the results that fragment sequences are often variable in spatial length. Variable length fragments, that track geology, are preferred.
For some problems cutting criteria need be selected interactively. Here the operator paints an example of data on one side of the cut and paints a second example of data on the other side of the cut. The application then performs a statistical analysis of all or some of the information in lower levels of the pattern pyramid to identify the information that classifies the two examples provided by the operator as different then uses that classification to determine the cut. This is the computationally most inefficient method.
While the cutting criteria for a step of cutting typically remain constant, the specific criteria can vary from layer to layer in the pattern database. As higher levels of the pattern database are computed the associated fragments created during the cutting process become larger.
Because geological unconformities occur in band-limited acoustical impedance zero crossings, it is necessary, when the present disclosure is used for seismic interpretation, to constrain the fragment cuts for all of the levels of the pattern database above the feature level to occur at the same spatial location as the fragment cuts for the feature level. The imposition of the constraint is accomplished by restricting the cutting criteria to be a function of the information one level below it. Other problems may not have need of the constraint.
It should be noted that the choice of the grid coordinate system, on which the data is sampled, typically has no relationship to the spatial distribution of the geology being studied and the associated data measurements. When the spatial dimensions of the data are higher than one, a fragment orientation needs to be selected. For geophysical data, the natural fragment orientation is to align it with geology. This is accomplished by computing a geology aligned coordinate system, which is an earth manifold, and using it to align the fragments and fragment sequences with geology. To simplify the implementation the fragments can be aligned with the seismic traces recognizing that, as geological dip becomes large, the approximation quality decreases.
When the coordinate system on which the underlying data is sampled is not aligned with the geology, edge noise can occur during cutting, attribute calculations, and statistic calculations. For optimum performance, the edge noise should be eliminated or attenuated by using a continuous representation (local or global spline fit) of the data when performing computations. The best, but computationally most inefficient, solution is a manifold with continuous local coordinate charts.
PDB Abstraction—Attributes
In the second step of the abstraction process, the attributes at each fragment are computed and are stored at the attribute location for the appropriate level in the pattern database. The specific attribute computations can be the same or can vary from level to level. The attributes may be stored in a pattern database, as software objects (parameters or methods) stored in RAM, as objects stored in an object database, as objects or data stored in an appropriately mapped relational or object-relational database, or stored via some other storage technique or mechanism.
PDB Abstraction—Statistics
The third step of the process is the statistical analysis of the previously calculated attributes. The statistical analysis gives the probability of the attribute occurring in its local neighborhood (local statistic) and in the total data set (global statistic). Some statistics may represent estimates or properties (sometimes called features) of the attributes for the next level up in the pattern pyramid. An example is attribute complexity or local attribute anisotropy.
In practice, other types of information may be stored along with statistics in the present disclosure including correction parameters. An example of a correction value occurs when the data is provided in a Euclidean format. However, geological measurements are best expressed in a geology-aligned fashion. To align the analysis with geology it needs to be aligned with the earth manifold. The corresponding earth manifold definition and/or local coordinate chart dip and azimuth values can be computed and saved within the database in the statistics level.
Additional properties, which are needed but are not captured by the attributes, may also be stored as statistics. These include properties of the earth manifold, measured on the local topology of the earth, such as local curvature.
Hyperdimensional Fragment and Binding Strength
c illustrates how a particular point of space in the input data 140 and 143, represented by the point 156, has corresponding points 154 and 153 in the feature layer, 150 and 148 in the pattern layer, plus 146 and 144 in the texture layer. The ordered set of points 156, 154, 153, 150, 148, 146, and 144 forms a trajectory called a hyperdimensional fragment of the data point 156 in question. The pattern pyramid has a set of hyperdimensional fragments that associate each data sample to the features, patterns, and textures to which it contributed. Because the type of abstraction analysis is problem specific, so too is the resultant hyperdimensional fragment.
When comparing the known data hyperdimensional fragment to the collection of target data hyperdimensional fragments the amount of similarity required to consider them matched is determined by the binding strength or affinity. This disclosure implements the concept of a binding strength by setting a range of acceptable feature, pattern, and texture values at each pattern pyramid level that the hyperdimensional fragment passes through. The result is that exact matches are no longer required but similar matches are allowed.
When the above-described process is completed, the hyperdimensional fragment and associated threshold becomes a template that is used for object identification. Making a comparison between the known data and the target data is accomplished by applying the template to the target data. The comparison is accomplished by searching through all of the hyperdimensional fragments in the target data set and determining if the feature, pattern, and texture values though which they pass are the same within the binding strength as the values in the known data hyperdimensional fragment. Templates can be stored in a template database and retrieved for later use on any target data set.
Scenes and Objects
The result of applying a template to a target data set pattern database is a scene that contains null values where matches did not occur and a value representing matched where matches did occur. The next step is to identify all data connected points where matches occurred and assign them to an object. The identification is accomplished by stepping through all of the points that are marked as matched and performing an auto-track that assigns all connected points that are marked as matched to an object. This is repeated until all points that are marked as matched have been assigned to connected objects. The result is a scene containing connected objects that represent potential hydrocarbon deposits. These objects represent a simultaneous analysis of how well they represent a potential reservoir through feature analysis, represent hydrocarbon indications through pattern and texture analysis, and include the presence of a depositional process that deposits reservoir rock though texture analysis.
Objects can have associated properties. For example, a 3D manifold (also referred to as a shrink-wrap) can be placed on the boundary (outside edge) of an object forming an object space. Topological properties of the object surface, such as local curvature, can be measured and stored as an object property.
Next, the scene, the collection of objects, is then analyzed in a quality control step to determine if the system is correctly creating the desired objects. If the system creates the expected objects, but the objects are incomplete or obscured due to seismic noise, the binding strength is modified and the data mining is repeated. If the expected objects are not created or too many objects that are false positives are created the amount of information in the PDB or associated parameters are modified, a new template is created and the data mining is repeated.
Finally the collection of objects, in the scene(s), is viewed to manually identify and remove any remaining false positives. The goal is to minimize the work in this step by a good choice of PDB construction.
Data Mining
Templates can be pre-computed from known data sets, stored in a template database, and used the pattern databases for one or many target data sets creating resultant scenes containing objects that satisfy the templates. This process is often referred to as data mining. The collection of objects becomes a lead inventory.
Feature Level Cutting Criteria, Attributes, and Statistics
For the 3D seismic first pass lead identification example the data being analyzed is band-limited acoustical impedance. The objective is to identify hydrocarbon filled reservoir rocks. In order to identify the hydrocarbons, it is preferable to gather information about the band-limited acoustical impedance values, depositional process, and the presence of hydrocarbon indicators (bright spots, dim spots, flat spots, etc.) but exclude the geological structure so that we can find opportunities for all possible trap structures. For this example the cutting criteria for features is cutting at each zero crossing of band-limited acoustical impedance as shown in
The feature attributes for this example are chosen to form a visual feature set. This set describes the band-limited acoustical impedance using the same descriptions as used by seismic stratigraphers when communicating their work. This choice ensures that the features are interpretable, or understood by geoscientists. Because the features are based on naturally occurring, geological visual properties, and because seismic stratigraphers have had considerable success using them, they are known classifiable. These interpretable features include the length of the fragment (also called thickness), the absolute value of the maximum acoustical impedance of the data within the fragment (also called max amp), the shape of the data values in the fragment, and the sign of the data values (+ or −). There are many ways to measure shape. One way to measure shape is to measure all of the statistical moments of the data in the fragment. This set of measurements represents all of the degrees of freedom of the problem. In practice, not all of the statistical moments are required to solve the problem. Often, only the first moment is used.
The statistics, for this example, consist of a global statistic. It is the probability of the given feature occurring in the entire data cube. Two local statistics is also computed. One is the data complexity in a local coordinate patch. Data complexity is the normalized sum of the data value variances. The second is local feature anisotropy. It computes the direction and magnitude of the feature variances in the local coordinate neighborhood. Both can be considered local texture estimates (also called texture features or texture statistics).
For seismic data the computationally most efficient method is to measure fragments for features aligned with seismic traces and is the way that seismic stratigraphers typically perform the task. Variations in structural dip may cause variations in the feature values that are not associated with rock or fluid variations. If the effects of these variations become too large, the fragments on which the features are measured must be aligned with the earth manifold. Since inline and xline fragments will carry primarily information about the earth's structure they are not used for this example. However, when the goal of the analysis is to identify structure similarity, inline and xline fragments should be used.
Pattern Level Cutting Criteria, Attributes, and Statistics
For the 3D seismic first pass lead identification example the pattern level cutting criteria is to cut the patterns so that the top and the bottom of the pattern fragments occurs at band-limited acoustical impedance zero crossings. The easiest way to accomplish this is by cutting the pattern level fragments from a combination of feature level fragments.
A shorter pattern fragment length can be computed by dropping one feature length off the top or one feature length off the bottom when performing the calculation. This is often referred to as a pattern fragment length of 3 feature lengths and is an example of what is referred to as an even feature length pattern fragment.
Longer pattern fragments can be constructed by extending either the odd or the even feature length pattern fragment described above. This is accomplished by adding one feature length to each end. Extending on both ends can be repeated as many times as required.
The pattern level attributes can be computed by performing a transformation of the feature attribute values associated with the pattern fragments into pattern space. After the transformation each location in pattern space contains the population density of pattern fragments that transform into it. Peaks in the population density can be identified and the space can be broken into clusters by placing decision surfaces between the clusters or peaks. The regions between decision surfaces for each cluster are assigned pattern attribute values. The pattern attribute values can then be transformed back to physical space and assigned to the pattern intervals as pattern attributes. This is the most computationally intensive technique and is too costly to be used for production processing and data mining.
A second method of computing pattern attributes is performed by breaking the pattern space up into user-defined bins. To do this the binding strength needs to be selected at this point of the analysis. The bin size is determined from the binding strength. For each pattern location the bin into which the feature attributes associated with the given pattern fragment transforms is easily computed and stored as the pattern attribute value at the pattern location. This association is, computationally, the most efficient method. However, the association method has the drawback that the binding strength must to be set at this point of the analysis rather than be selected dynamically or interactively later, when the known data and target data pattern databases are compared. If the binding strength is not known it will be difficult to use this method. Sometimes it is determined by trial end error where the user-repeats the analysis with different binding strengths and chooses the one that gives the best results. This method is often refereed to as fixed bin clustering or quick clustering.
A third method is to compute the coordinates of the pattern space location into which the feature attributes associated with the pattern fragment transforms and storing the coordinates as the pattern attribute values at the pattern location. The coordinates can be expressed in spherical coordinates, Cartesian coordinates, or any useful projection. In this method the pattern attributes have several values. The maximum number of values is equal to the number of feature fragments that are combined to create the pattern fragment. This is the computationally less efficient than the second method but much faster than the first method and can be used for data mining. It has the drawback that each pattern attribute has multiple associated values thus uses a lot of space on disk and in RAM. It is possible to decrease the storage requirements by discarding of combining values. It has the benefit that the binding strength selection can be accomplished during pattern database comparison, which makes it the most flexible method.
Any or all of the above methods of computing pattern attributes can be included as one or several levels in the pattern level of the pattern pyramid. Other methods of unsupervised classification, usually clustering methods, can also be used. The specific choices depend on how well and how uniquely the algorithm isolates out (classifies) the targets of interest from the rest of the target data.
Statistics can include the same algorithms used at the feature level of the pattern pyramid but applied to the pattern attribute values.
For seismic data the computationally most efficient method is to measure pattern fragments that are aligned with seismic traces. This is the way seismic stratigraphers typically perform the task. Variations in structural dip may cause variations in the feature attribute values that are not associated with rock or fluid variations. If the effects of these variations become too large, the fragments on which the feature attributes are measured must be aligned with the earth manifold. Since inline and xline fragments will carry primarily information about the earth's structure, they are not used for this example. When the goal of the analysis is to identify similar structures, the inline and xline fragments should be used. Fragment orientations that are aligned with the earth manifold or along local dip and strike will capture information about stratigraphic variations in the rocks and fluid variations related to the transition from hydrocarbon filled reservoir rock to brine filled reservoir rock. For the 3D seismic first pass lead identification example it might be useful to use a 3D pattern pyramid and populate the strike and dip sides of the pattern pyramid with strike and dip oriented pattern attributes and statistics computed from feature attributes from the vertical level of the pattern pyramid. This is computationally intensive, thus it might be faster to estimate them by computing them in the inline and xline directions but limiting the calculation to local coordinate patches with a common feature sign.
Texture Level Cutting Criteria, Attributes, and Statistics
For the 3D seismic first pass lead identification example the cutting criteria, attribute calculations, and statistics calculations are the same as for the pattern level with the following exceptions. First, the cutting criteria are computed as multiples of the pattern fragments rather than feature fragments. Second, the texture level attributes are stored at texture locations and are calculated from the pattern level attributes rather than the feature level attributes. The input to the transformation is texture fragments and the transformation is to texture space rather than pattern space. Third, the statistics only include the global statistics.
PDB Comparison, Objects, and Scenes
For the 3D seismic first pass lead identification example, the PDB comparison is performed by comparing hyperdimensional fragments. The binding strength is specified for each level of the pattern pyramid where it was not already specified during pattern database construction usually by using the quick clustering technique above. When this step is performed for the first time it is often performed interactively during visualization of a target data set and the related pattern database. When the optimal binding strength has been chosen, the template is applied to the target data set. This step is often referred to as applying a scene construction tool. After this is accomplished the spatially connected objects are computed using another tool that is also referred to as a scene tool.
Data Mining and Lead Inventory
For the 3D seismic first pass lead identification example the template computed above is saved in the template database. The appropriate templates are checked out and applied to all of the data in the geographical region being analyzed. The resulting scenes and associated templates are combined using Boolean operations that are usually referred to as Boolean scene tools. The final product is a lead inventory that is associated with a scene containing a list of multiple leads (objects) and lead parameters. The lead parameters include lead names, locations, spatial sizes, global statistics, local statistics, and other useful information as required by the operator.
Implementation
The present disclosure is preferably implemented as a set of one or more software processes on a digital computer system. However, the present disclosure may also be implemented purely in hardware, or may be virtually any combination of hardware and software.
The present disclosure may be modeled on a digital computer with the aid of various software objects that encapsulate data in the form of properties, and computations as methods. Moreover, these various object may have one or more methods through which selected functionality is performed. Each of these objects has a class definition and is interconnected according to the following descriptions and referenced drawings.
The Apparatus of the Present Disclosure
The RAM 230 can be high speed memory to accelerate processing. High-speed graphics card 250 is preferably an ultrahigh-speed graphics card like the Intense 3D Wildcat (manufactured by 3DLabs of Huntsville, Ala.). High-resolution display 270 is the highest resolution display currently available in order to support the applications, which are intensely graphic in nature, and is electrically connected to main unit 210 by VGA cable 275. Also electrically connected to main unit 210 are: CD-ROM drive 280, 8 mm tape drive 290, mouse 292, and keyboard 294. Other peripheral devices may also be connected to the PC 200.
In operation, seismic data enters the enhanced PC 200 via, for example, the 8 mm tape drive 290, the CD-ROM drive 280 and/or the network card 240. This seismic data is stored in, for example, SEG-Y format in database 262 and is processed by CPU 220 using applications 266, with mouse 292, and keyboard 294 as input devices and high-speed memory 230 to facilitate processing. The processed seismic data is then stored in a PDB 264 format.
The information handling system 200 can be equipped information, such as one or more instructions, that implement the methods disclosed herein. The instructions can be sequential, event driven, or a mix thereof. Moreover, the instructions can be written in a variety of programming languages, including but not limited to: C, C++, Python, FORTRAN, Java, and the like. The instructions utilized by the information handling system 200 can be stored in persistent memory, such as a hard drive, temporarily in RAM, and/or stored permanently on a computer-readable medium, such as a compact disk (“CD”) that can be read by CD-ROM drive 280. The computer-readable medium would contain, for example, a file system and/or data structure that would enable the information handling system to read instructions from the medium and load the instructions into RAM 230 and/or the CPU 220.
After pattern analysis, a PDB contains a collection of data volumes. The collection includes a 3D seismic data volume, multiple associated pattern, feature, texture volumes, and multiple scene volumes. The data values are stored so that they can be addressed as spatially vertical columns or horizontal slabs with the columns and slabs made up of subsets called bricks. A stack of bricks that extend from the top of the cube to the bottom is a column. A mosaic of bricks that extends horizontally across the volume is a slab. The brick size is chosen to optimize data access, for example, 64 by 64 samples in size. The samples are 8-bit integer, 32-bit floating point, or any other desired format. Each volume contains metadata including:
The PDB collection, and associated metadata, can be stored as files on a file system, as information in a database, or as a combination of the two.
After modification, a seismic template is created for each geoscientist, and this template is stored in template library 268. During processing, the seismic data is viewed on the high-resolution display 270. After further processing, the seismic data is stored in template library 268, and output to 8 mm tape drive 290 or CD-ROM 280, or transmitted via the network card 240.
The methods illustrated in
Method of 3D Seismic First Pass Lead Identification
The present disclosure employs the above-identified apparatus for various purposes. The method of the present disclosure will now be illustrated via the 3D seismic first pass lead identification example method of the present disclosure is illustrated in
Method of Building a Pattern Data Base for Geophysical and Geological Data
An additional embodiment of the present disclosure is a system for and method of building a pattern database. This method will now be illustrated via a building a pattern database for geophysical and geological data example method that is illustrated in
Method of Building a Pattern Data Base for 3D Band-Limited Acoustical Impedance
An additional embodiment of the present disclosure is a system for and method of building a pattern database. This method will now be illustrated via a building a pattern database for 3D band-limited acoustical impedance example that is illustrated in
Method of Preparing Seismic Data for Pattern Analysis
An additional embodiment of the present disclosure is a system for and method of preparing seismic data for pattern analysis.
Method of Constructing a Pattern Data Base for 3D band-limited Acoustical Impedance
An additional embodiment of the present disclosure is a system for and method of constructing a pattern database. This method will now be illustrated via a preparing 3D seismic for pattern analysis example method that is illustrated in
Note that this alternate embodiment of the present disclosure is practiced after completing the method described in
Fragment Cutting and Feature Attribute and Statistic Computation
An additional embodiment of the present disclosure is a system for, and method of, selecting fragments and extracting features from the prepared data.
Features are measurements on the input data. As a collection, features describe the data as fully as necessary to perform the desired data classification. Mathematically, features are represented as a vector state space where each vector represents a state of the image. The axes of the state space represent the degrees of freedom of the problem; in this case the image features. To represent the data as fully as possible and as efficiently as possible, the state space axes, and thus the features, should span the space and be linearly independent.
For this application, the features have an added requirement. Because features will be visualized along with the data and interpreted as a visual image, they need to represent simple visual properties of the seismic data that are familiar to geoscientists. In step 602, the batch application reads the feature parameters from the job queue. In step 604, the batch application initializes the column pointer so that, when incremented, the column pointer points to the location of the first column on the hard drive 260. In step 606, the batch application increments the column pointer to the location of the next column on the hard drive 260, reads the input column from disk, and places the input column in RAM. In step 607, the batch application identifies the location of the PDB on the hard drive 220 and reads the PDB in one of two ways:
Processing the entire data cube in memory all at once allows the system operator to visualize the data cube and modify parameters during processing. Modification of parameters is accomplished to select and tune the parameters on a relatively small subset of the entire data set. Streaming enables brick-by-brick processing of large sets of data that exceed the size of memory. The available high-speed memory 230, and the size of the data cube, dictate which storage method is used.
The following description describes the batch operation with data streaming. However, the same pattern computations can be used for the case where the data is all in memory. All of the steps in method 400 (illustrated in
Pattern Attribute and Statistic Calculation
An additional embodiment of the present disclosure is a system for and method of generating pattern attributes from feature attributes in a pattern abstraction database. An example of a method for accomplishing this is to do it by computing for each voxel in physical space the pattern space location into which the data associated with the voxel would transform if it were transformed into pattern space. This is equivalent to a fiber view. The pattern space is assigned pattern attribute, usually cluster or bin numbers, which are assigned to the appropriate voxel in physical space.
Pattern space is represented mathematically as a vector state space representing the patterns formed by the features associated with neighboring fragments measured from the data set. The vector space has one axis for each feature being analyzed. The features may be multiple features at the same spatial location, the same feature from neighboring locations, or a combination of both. The pattern is labeled by its location in the pattern space that is given by the values of the associated features that make up the pattern.
The entire pattern space is then visualized in order to facilitate the analysis of the underlying geology from which the geological features were extracted, and thereby determine the location of, for example, hydrocarbon deposits.
a, 7b, and 7c illustrate a method of generating patterns from features in a pattern abstraction database as described herein. Note that this alternate embodiment of the present disclosure is practiced after completing the method described in
There is a concept called a “fiber view” that is created to go with pattern space 1300. The entire data set is transformed into the pattern space and put into bins, as represented by bin 1310. The bins are numbered and then the bin numbers are transformed back to physical space and placed in a data cube in their physical space locations. The numbers that have been transformed back to physical space is what topologists call a fiber view of transform space.
In step 725, the batch application initializes the pointer for a column, which is a vertical column of voxels, so that when incremented it points to the location of the first column on the hard drive 260. Step 725 is the beginning of the process that will assign pattern values (also referred to as bin values) to every fragment within the data cube. In step 730, the batch application increments the column pointer to the location of the next column on the hard drive 260 and reads the input column from disk, placing it in RAM. In step 731, the batch application reads the input feature column(s) from the hard drive 260, placing it in RAM. In step 732, the batch application initializes the pointer for a trace so that, when incremented, it points to the location of the first trace in the current column on the hard drive 260. In step 734, the batch application increments the trace pointer to the location of the next trace in the current column on the hard drive 260. In step 736, the batch application initializes the pointer for a fragment so that, when incremented, it points to the location of the first fragment in the current trace on the hard drive 260. In step 738, the next fragment in the first column is identified by the batch application. In step 740, the pattern space location, i.e., the bin, is computed for every fragment in every column by the batch application. In this step, the pattern space location from step 740 is stored as a pattern value by the batch application. The pattern value corresponds to the bin number, wherein bins 1322 to 1326 have pattern values of 0 to 8 (see
Data Mining Using a Template
An additional embodiment of the present disclosure is a system for and method of performing data mining using a previously created template.
Quality Control Analysis of Feature Attributes
An additional embodiment of the present disclosure is a system for and method of performing quality control of features computed by method 500.
Quality Control Analysis of Pattern Attributes
An additional embodiment of the present disclosure is a system for and method of performing quality control of patterns computed by method 700.
Method of Adding Cutting, Attribute, or Statistic Algorithms to the Pattern Data Base Building Application
An additional embodiment of the present disclosure is a system for and method of adding additional functionality such as additional cutting algorithms, feature attributes, feature statistics, pattern attributes, pattern statistics, texture attributes, and texture statistics to the pattern database building software. An example method is shown in
Note that this alternate embodiment of the present disclosure is practiced as a part of method 1100 described in
Similarly, new cutting algorithms, feature attributes, feature statistics, pattern attributes, pattern statistics, texture attributes, and texture statistics might be needed to solve the geological problem. The present disclosure facilitates satisfaction of the need to add features and functions. In step 1115, definitions are developed for the custom features or other functionality needed according to geo-scientist specifications from step 1105. These definitions are used to modify the source code in the pattern abstraction software program that embodies and implements the methods of the present disclosure. These modifications are implemented using standard practices and commercially available object-oriented analysis and design language, such as those from the Object Management Group (“OMG”), which uses the Unified Modeling Language (“UML”) diagramming methodology. The modification is constructed as source code for an application plug-in. The plug-in can be compiled to create a static or dynamically linked library (“DLL”) that is placed in the software directory containing the application. When executed, the application recognizes and executes the plug-in according to standard software techniques. It should be noted that while it is preferred to utilize object-oriented programming techniques, non-object oriented implementations may also be used to generate the software and/or hardware needed to implement the methods of the present disclosure. Moreover, virtually any programming language may be used to implement the methods the present disclosure.
Although the method described in
a illustrates a plot of band limited acoustical impedance 1200 as a function of time or depth that is typical of seismic data.
b illustrates a plot of broadband acoustical impedance 1250 as a function of time or depth that is typical of seismic data on the left and its first derivative 1260 on the right.
Fragments are a function of acoustical impedance related to time or depth. The seismic data is frequently measured as a function of the two way travel time of the sound (down to the reflector and back up), although the data can be converted to a measurement of depth, which is the preferred measurement for seismic data analysis. Either time or depth measurement can be used in all of the seismic data processing, but because it is costly to convert the seismic data to depth, data is frequently acquired and processed in terms of time.
The length of each fragment is measured in a different manner for band limited AI and broadband AI. For band limited AI it is the distance between zero crossings, as shown by top of fragment 1205 and bottom of fragment 1220 in AI fragment 1210 that is a portion of the band limited AI function 1200. Thickness 1215 is measured along X-axis 1230, and is the distance between top of fragment 1205 and bottom of fragment 1220, in time or depth. Top of fragment 1205 is the zero crossing at the minimum depth or initial time. Bottom of fragment 1220 is the zero crossing at the maximum depth or maximum time. Max amplitude 1225 is the top point of the function of band limited AI 1200 within fragment 1210 as measured on Y-axis 1235. The band limited AI fragment 1210 represents a portion of one trace in one column of information in a data cube. Fragments may be determined by zero crossings, non-zero crossings, or a mix of zero and non-zero crossings (meaning one end of the fragment is at a zero crossing and the opposite end is at a non-zero crossing).
For broadband AI 1250 in
c provides the mathematical expression for calculating RMA amplitude and definitions of the terms in the expression. This equation is implemented as a method of one or more than one software objects in the computer software portion of the implementation of the present disclosure.
d provides the mathematical expression for calculating one example of a shape feature that is the first statistical moment.
a provides the mathematical expression for calculating horizontal complexity.
b shows the coordinate neighborhood for observation #51305. The coordinate neighborhood has an Xline axes 1315 and an Inline axes 1320. The neighborhood has a diameter of 2 samples 1310. It contains 9 observations #1 to #9 shown as dots 1301 to 1309. Larger diameters may be used thereby increasing the number of samples in the coordinate neighborhood.
a, 14b, 14c, and 14d define feature and feature function anisotropy.
a and 16b define the M and a pattern space measures.
c and 16d extend the concept to three dimensions giving M, α, and β pattern space measures.
a illustrates a diagram of pattern space 1500 for the shape feature for a fragment length of two, including pattern space length 1510, central bin length 1515, bins 1520 to 1528, fragment #2 axis 1525. The fragment #2 axis 1525 extends from a top loaded shape 1530 through a symmetric shape 1535 to a bottom loaded shape 1540. The fragment #1 axis 1560 extends from top loaded shape 1555 through symmetric shape 1550 to bottom loaded shape 1545.
Pattern space 1500 shows eight separate bins, represented by bins 1520 to 1528, that are used to organize feature fragment sequences (for example, shape) as shown in
Fragment sequences are considered as they are analyzed, in this example, with respect to their shape. Since the fragment length is two, two fragment sequences must be categorized. If both the first and second fragments are bottom-loaded, the fragment sequence falls into bin 1528. Similarly, each of the other bins is filled with a particular fragment sequence. The bin tolerance is defined as the ratio of the central bin length 1515 to the pattern space length 1510. The tolerance determines how the bin sizes vary. The less common values are on the outside of
In a similar fashion,
In another embodiment, an additional feature can be added, namely “string length.” The string length is another mechanism for providing additional discrimination of seismic waveform character in addition to (or in lieu of) other shape features.
As mentioned above, string length is used in association with a fragment (the region between two zero-crossings of a seismic trace). Used in this manner, the string length becomes a type of descriptor of that fragment. Calculating string length in this way allows the analyst to do three things. First, the analyst can compute the string length ratio that is an even more powerful descriptor of a fragment. Second, associating string length and string length ratio with a fragment, as a feature, allows the analyst to combine that measurement for a fragment with like measures from surrounding fragments to form patterns involving string length and string length ratio. Third, the analyst can combine the string length and/or string length ratio features and/or patterns with other features and patterns (derived independently, perhaps with alternate methods) within a system to help identify target geology and to produce geobodies from the seismic data.
In one embodiment of the disclosure, the shape of the waveform is used indicate whether a fragment is symmetric, bottom-loaded or top-loaded. String length is typically the arc length, i.e., the length of the seismic trace between two points over a fragment, with the length of the fragment being defined by crossing the reference axis at two inflection points. Referring to
Thickness=padstart+padend+n
The length of the padstart 1804 along the reference axis 1702 is defined as:
padstart=AFS/(AFS−AFS−1)
The length of the padend 1810 along the reference axis 1702 is defined as:
padend=AFE/(AFE−AFE+1)
When determining the string length (“SL”), the string length for the two partial segments nearest the zero crossings (padstart 1804 and padend 1810) are used along with the sample interval (“DIGI”) which is typically measured in a unit of time, such as seconds. The sample index of the first sample in the fragment (“FS”) has amplitude AFS 1806, and the sample index of the last sample in the fragment (“FE”) has amplitude AFE 1812. Both amplitudes AFS 1806 and AFE 1812 are used to calculate the string length. Specifically:
SL=((AFS2+(padstart*DIGI)2)0.5)+((AFE2+(padend*DIGI)2)0.5)
For each fragment, one loops through the sample values for each fragment. In pseudocode, we could write:
The absolute string length is equivalent to SL above. In some embodiments, it may be useful to remove the contribution of thickness in the string length to generate the value SLL, which may be calculated by the following equation:
SLL=SL−Thickness
In other embodiments, it may be useful to use a signed string length SLS, which is calculated by multiplying the string length by the fragment's sign (positive or negative).
SLS=Sgn(A0)*SL
Because fragments are defined by zero-crossings, each value A0 within a fragment has the same sign, so any portion A0 of a fragment may be used to determine the fragment's sign.
In yet another embodiment, the string length ratio is used. To calculate the string length, a reference string length (SLref). For data types with zero crossings, the reference string length for each fragment is determined by:
SLref=2*{Amax2+DIGI2)0.5}
where Amax is the maximum amplitude for the fragment in question. Incidentally, because the maximum amplitude is squared, it may be the absolute Amax, or the signed Amax. In one embodiment, SLref will always be positive. However, alternate embodiments may have SLref always be negative.
Two other values may be useful to the exploration of hydrocarbons, namely absolute string length ratio (SLAB) and signed string length ratio (SLS). Where:
SLAB=SL/Sref
and
SLS=Sgn(A0)*SLAB
While the string length may be viewed as a straightforward calculation, string length is particularly meaningful when compared to something else. For example, comparing string length with other measurements referenced above, particularly in conjunction with vertical patterns and horizontal measures. Moreover, useful information can be obtained by comparing the string length result to a reference string length for a given fragment to generate a string length ration. The reference quantity is calculated differently for input data that has zero crossings (such as RFC or RAI data) and input data that have no zero crossings, such as broadband inversions. For each of the characterizations, a “plug-in” may be created for use with software that is used for the PDB. For example, three different plug-ins may be implemented for string length, string length ratio for zero crossing data, and string length ration for non-zero crossing data. Alternatively, a single plug-in could be used to calculate two or more of the above-identified string characteristics for output as a pattern database (PDB). A user could be given the option of selecting one or more of the string length characterizations for his/her input data type.
In another embodiment, string length ratio can be calculated in other ways. for example, an assumption can be made that DIGI2 (see paragraph [0222]) is small and can be ignored. Thus string length would simply be the square root of the sum of the squared differences, or:
SL=((AFS2+padstart2)0.5)+((AFE2+padend2)0.5)
which, by eliminating the squaring and square root operations, can simply express the string length ratio by summing the absolute values of the differences, or:
SLR=SUM{ABS(Ai−Ai−1)}/(2*ABS(MaxAmp)
where the sum is the number of samples in a fragment. In other words, the absolute value of the sample-to-sample differences are added up for the numerator.
|1−0|+|2−1|+|3−2|+|2−3|+|1−2|+|0−1|=1+1+1+1+1+1=6
and the denominator would be
2*|AmplitudeMax|=2*|3|=6
and thus the string length ration would be equal to 6/6=1 for the fragment 1900. Fragments are most commonly half-sinusoids, and the same analysis would apply, meaning that their string length ratio would be close to one.
Fragments, however, sometimes have a more complex shape, such as a doublet, i.e., a fragment having two peaks, such as the fragment 2006 of
|1−0|+|3−1|+|1−3|+|13−1|+|1−3|+|0−1|=1+2+2+2+2+1=10
and the denominator would be
2*|AmplitudeMax|=2*3=6
and thus the string length ration would be 10/6=1.667.
The functionality for string length, string length ratio, zero crossing data, and/or string length ratio for non-zero crossing data may be implemented directly into a software application, or as a plug-in, or directly in hardware, such as the system illustrated in
The present invention, therefore, is well adapted to carry out the objects and to attain the ends and advantages mentioned, as well as others inherent therein. While the present invention has been depicted, described, and is defined by reference to particular preferred embodiments of the present invention, such references do not imply a limitation on the present invention, and no such limitation is to be inferred. The present invention is capable of considerable modification, alteration, and equivalents in form and function, as will occur to those of ordinary skill in the art. The depicted and described preferred embodiments of the present invention are exemplary only, and are not exhaustive of the scope of the present invention. Consequently, the present invention is intended to be limited only by the spirit and scope of the appended claims, giving full cognizance to equivalents in all respects.
This application is a continuation-in-part of U.S. patent application Ser. No. 10/308,933, entitled “PATTERN RECOGNITION APPLIED TO OIL EXPLORATION AND PRODUCTION” which was filed by inventors Robert Wentland, Peter Whitehead, Fredric S. Young, Jawad Mokhtar, Bradley C. Wallet and Dennis Johnson on Dec. 3, 2002, and which is a conversion of U.S. Provisional Application Nos. 60/395,960 and 60/395,959 both of which were filed on Jul. 12, 2002 and all are hereby incorporated by reference herein for all purposes. This application is also a continuation-in-part of U.S. patent application Ser. No. 10/308,938, entitled “METHOD, SYSTEM AND APPARATUS FOR COLOR REPRESENTATION OF SEISMIC DATA AND ASSOCIATED MEASUREMENTS” which was filed by inventors Robert Wentland and Jawad Mokhtar on Dec. 3, 2002, and which is a conversion of U.S. Provisional Application Nos. 60/395,960 and 60/395,959 both of which were filed on Jul. 12, 2002, and all are hereby incorporated by reference herein for all purposes. This application is related to U.S. patent application Ser. No. ______ entitled “METHOD AND SYSTEM FOR HORIZONTAL PATTERN STATISTICAL CALCULATION IN SEISMIC ANALYSIS” by Robert Wentland, which was filed on ______, which is assigned to the same entity as the present application.
Number | Date | Country | |
---|---|---|---|
60395960 | Jul 2002 | US | |
60395959 | Jul 2002 | US | |
60395960 | Jul 2002 | US | |
60395959 | Jul 2002 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 10308933 | Dec 2002 | US |
Child | 11147643 | Jun 2005 | US |
Parent | 10308938 | Dec 2002 | US |
Child | 11147643 | Jun 2005 | US |