Aspects of the present inventive concept generally relate to recovery of relatively weak signals that are contaminated by noise and learn otherwise non recoverable using conventional processing methods.
Various types of data may be collected via use of one or a network of various types of sensors such as audio sensors, cameras, etc. Examples of such data can be seismic data collected in a field that can be analyzed to understand surface/subsurface seismic activity, surface temperature data, weather data, traffic data, etc. Such data may be contaminated by noise during the collection process. Currently, various conventional noise attenuation methods are utilized to recover signals representing the collected data. These conventional noise attenuation methods often use different characteristics in frequency, wave number or other transform domains to separate the signals representing the collected data from the noise. Examples of such noise attenuating methods include, but are not limited to, f-x projection filtering for random noise attenuation, a prediction error filtering to estimate coherent signal in the f-x domain, using low-rank structure in the t-f domain to attenuate ice-break noise, etc.
A shortcoming of all such noise attenuation methods is that they all involve a trade-off between preservation of the signals and the amount of filtered noise such that as noise attenuation increases, the target signals (e.g., seismic signals) are adversely affected. Moreover, some signals are several orders of magnitude lower than the contaminating noise. Therefore, these signals are lost/eliminated together with the noise during the filtering process, which adversely affect an accurate and complete analysis of the seismic data for underlying applications.
The present inventive concept provides a system and method to extract signals that represent underlying collected data from contaminating noise signals, and to prevent the loss of portions of such signals that are several orders of magnitude smaller than the noise signals, which may be referred to as weak signals or relatively weak signals. The system and method of the present inventive concept are operable to apply machine learning to preserve weak signals during a de-noising process for extracting the signals.
In one aspect, a computer-implemented method of noise contaminated signal recovery includes receiving, at a server, a first signal including a first portion and a second portion, the first portion indicative of data collected by a plurality of sensors, the second portion representing noise; performing a first denoising process on the first signal to remove the noise to yield a first denoised signal; applying a machine learning model to determine a residual signal indicative of a difference between the first signal and the first denoised signal; and determining a second signal by adding the residual signal to the first denoised signal, the second signal comprising (i) signals of the first portion with higher magnitudes than the noise in the second portion, and (ii) signals of the first portion having lower magnitudes than the noise in the second portion.
In another aspect, the data corresponds to seismic data collected by the plurality of sensors.
In another aspect, determining the second signal is based on a single dictionary representing the noise trained using the machine learning model.
In another aspect, determining the second signal is based on a dual dictionary representing the noise and the data, the dual dictionary being trained using the machine learning model.
In another aspect, the method further includes training the machine learning model with the first denoised signal every time the server receives a different instance of the first signal and determines a corresponding first denoised signal.
In another aspect, training the machine learning model includes determining a solution to an optimization problem using a k-singular value decomposition algorithm.
In another aspect, training the machine learning model is unsupervised.
In another aspect, determining the residual signal includes determining a solution to a dual domain sparse inversion problem using a sparse representation of the residual signal and the noise.
In another aspect, solving the dual domain sparse inversion problem includes using a deterministic algorithm, the deterministic algorithm corresponding to one of a nonmonotone alternating direction method, or stochastic algorithm such as matching pursuit.
In one aspect, a system for noise contaminated signal recovery includes memory having computer-readable instruction stored therein, and one or more processors. The one or more processors are configured to execute the computer-readable instructions to receive a signal contaminated by noise, the signal including data collected by a plurality of sensors, the second portion representing noise; and process the signal using a machine learning model to yield a processed signal such that (1) the noise is removed from the processed signal and (2) portions of the data with magnitudes lower than magnitudes of the noise are preserved in the processed signal after removal of the noise.
In another aspect, the one or more processors are configured to execute the computer-readable instructions to process the signal by performing a denoising process on the signal to filter out the noise to yield a denoised signal; applying a machine learning model to determine a residual signal indicative of a difference between the signal and the denoised signal; and determining the processed signal by adding the recovered residual signal to the denoised signal.
In another aspect, the one or more processors are configured to control the plurality of sensors to determine the residual signal by determining a solution to a dual domain sparse inversion problem using a sparse representation of the residual signal and the noise.
In another aspect, solving the dual domain sparse inversion problem includes using a deterministic algorithm, the deterministic algorithm corresponding to one of a nonmonotone alternating direction method, or stochastic algorithm such as matching pursuit.
In another aspect, the one or more processors are configured to execute the computer-readable instructions to train the machine learning model with the denoised signal every time a different instance of the signal is received and a corresponding denoised signal is determined.
In another aspect, the one or more processors are configured to execute the computer-readable instructions to train the machine learning model by determining a solution to an optimization problem using a k-singular value decomposition algorithm.
In another aspect, the one or more processors are configured to control the plurality of sensors to collect the data over a specified period of time.
In one aspect, one or more non-transitory computer-readable medium have computer-readable instructions, which when executed by one or more processors of a system for noise contaminated signal recovery, cause the one or more processors to receive a signal contaminated by noise, the signal including data collected by a plurality of sensors, the second portion representing noise; and process the signal using a machine learning model to yield a processed signal such that (1) the noise is removed from the processed signal and (2) portions of the data with magnitudes lower than magnitudes of the noise are preserved in the processed signal after removal of the noise.
In another aspect, the execution of the computer-readable instructions by the one or more processors, cause the one or more processors to process the signal by performing a denoising process on the signal to filter out the noise to yield a denoised signal; applying a machine learning model to determine a residual signal indicative of a difference between the signal and the denoised signal; and determining the processed signal by adding the recovered residual signal to the denoised signal.
In another aspect, the execution of the computer-readable instructions by the one or more processors, cause the one or more processors to control the plurality of sensors to determine the residual signal by determining a solution to a dual domain sparse inversion problem using a sparse representation of the residual signal and the noise.
In another aspect, solving the dual domain sparse inversion problem includes using a deterministic algorithm, the deterministic algorithm corresponding to one of a nonmonotone alternating direction method, or stochastic algorithm such as matching pursuit.
In another aspect, the execution of the computer-readable instructions by the one or more processors, cause the one or more processors to form a dataset using the signal and perform the denoising process on the data set.
In another aspect, the execution of the computer-readable instructions by the one or more processors, cause the one or more processors to train the machine learning model with the denoised signal every time a different instance of the signal is received and a corresponding denoised signal is determined, by determining a solution to an optimization problem using a k-singular value decomposition algorithm.
In another aspect, the data corresponds to seismic data collected by the plurality of sensors over a specified period of time.
In order to describe the manner in which the above-recited and other advantages and features of the present inventive concept can be obtained, a more particular description of the principles briefly described above will be rendered by reference to specific example embodiments thereof which are illustrated in the appended drawings. Understanding that these drawings depict only exemplary embodiments of the present inventive concept and are not therefore to be considered to be limiting of its scope, the principles herein are described and explained with additional specificity and detail through the use of the accompanying drawings in which:
Various example embodiments of the present inventive concept are discussed in detail herein. While specific implementations are discussed, it should be understood that this is done for illustration purposes only. A person skilled in the relevant art will recognize that other components and configurations may be used without parting from the spirit and scope of the present inventive concept. Thus, the following description and drawings are illustrative and are not to be construed as limiting. Numerous specific details are described to provide a thorough understanding of the present inventive concept. However, in certain instances, well-known or conventional details are not described in order to avoid obscuring the description. References to one or an example embodiment of the present inventive concept can be references to the same example embodiment or any example embodiment; and, such references mean at least one of the example embodiments.
Reference to “one embodiment” or “an embodiment” means that a particular feature, structure, or characteristic described in connection with the example embodiment is included in at least one example embodiment of the present inventive concept. The appearances of the phrase “in one embodiment” in various places in the specification are not necessarily all referring to the same example embodiment, nor are separate or alternative example embodiments mutually exclusive of other example embodiments. Moreover, various features are described which may be exhibited by some example embodiments and not by others.
The terms used in this specification generally have their ordinary meanings in the art, within the context of the present inventive concept, and in the specific context where each term is used. Alternative language and synonyms may be used for any one or more of the terms discussed herein, and no special significance should be placed upon whether or not a term is elaborated or discussed herein. In some cases, synonyms for certain terms are provided. A recital of one or more synonyms does not exclude the use of other synonyms. The use of examples anywhere in this specification including examples of any terms discussed herein is illustrative only, and is not intended to further limit the scope and meaning of the present inventive concept or of any example term. Likewise, the present inventive concept is not limited to various example embodiments given in this specification.
Without intent to limit the scope of the present inventive concept, examples of instruments, apparatus, methods and their related results according to the example embodiments of the present inventive concept are given below. Note that titles or subtitles may be used in the examples for convenience of a reader, which in no way should limit the scope of the present inventive concept. Unless otherwise defined, technical and scientific terms used herein have the meaning as commonly understood by one of ordinary skill in the art to which the present inventive concept pertains. In the case of conflict, the present document, including definitions will control.
Additional features and advantages of the present inventive concept will be set forth in the description which follows, and in part will be obvious from the description, or can be learned by practice of the herein disclosed principles. The features and advantages of the present inventive concept can be realized and obtained by means of the instruments and combinations particularly pointed out in the appended claims. These and other features of the present inventive concept will become more fully apparent from the following description and appended claims, or can be learned by the practice of the principles set forth herein.
Turning to
Sensors 104 may be any type of known or to be developed device capable of collecting the underlying data including, but not limited to, conventional cameras, infrared cameras, audio/video sensors, etc. In one example, sensors 104 may be equipped with known or to be developed components for communicating with other sensors and/or other network components such as access points 106 and/or processing center 110 (which may also be referred to as receiver 110 or server 110). These components can be wireless connection interfaces, ports for wired connection to other devices, etc.
Setting 100 further includes one or more access points 106, which may include any type of known or to be developed access point such as a wireless access point, a base station, a 4G/5G node B, etc.
Sensors 104 and access points 106 may communicate via any known or to be developed wired and/or wireless communication schemes and protocols.
Access points 106 may be communicatively coupled to processing center 110 via internet 108. Processing center 110 may be a single unit or comprised of multiple units remotely placed from one another but communicatively coupled with one another. Processing center 110 may function to provide a network operator with the capability to access, monitor, manage and maintain access points 106 and/or sensors 104, as will be described below.
At S200, processing center 110 may receive collected data from sensors 104 via one or more of access points 106, as described above with reference to
In one example, processing center 110 may configure sensors 104 to collect such data continuously or for a specific period of time (e.g., for 24 hours, 48 hours, a week, a month, etc.). The collected data may be sent to processing center 110 by sensors 104 as a live stream as the data is being collected or may be sent all at once after collection (e.g., after the data is collected for the specified period of time). In another example, a command send by processing center 110 to sensors 104 may trigger transmission of the collected data from sensors 104 to processing center 110.
At S202, processing center 110 may group the collected data or portions thereof together to create a dataset to be analyzed. This grouping of the collected data or portions thereof may be based on timestamps associated with the received data. For example, when processing center 110 receives the collected data continuously, at S202, processing center 110 may select a portion of the data received over a specific period of time (last 24 hours, 48 hours, week, month, etc.) into a dataset to be analyzed.
In another example, sensors 104 may not be operating continuously. For example, processing center 110 may be able to turn them on and off. As such, processing center 110, prior to S200, may turn sensors 104 on, receive the collected data for a specific period of time, turn sensor 104 off and then at S202 group the received data into a dataset. This dataset may be represented by d. d may be composed of a signal portion (desired signal representing the collected data), represented by u, and a noise portion representing either coherent noise or incoherent noise depending on the noise type that contaminated the signal portion, represented by n. Accordingly, d may be given by formula (1) below:
d=u+n (1)
At S204, processing center 110 may perform an initial filtering (denoising) process for filtering out (removing) the noise portion to the extent possible using any known or to be developed filtering/denoising process. The signal portion after the initial filtering process may be denoted as û (may be referred to as the initial filtering result).
S204 results in a “clean” mode for training, therefore, a certain degree of signal leakage is permissible. In one example, therefore, the initial filter process may apply a simple filtering method due to their efficiency such as an f-x deconvolution process or f-k filter with moveout correction to filter the noise portion or any other known or to be developed filtering methods or algorithms.
At S205, processing center 110, may rearrange (partition/divide) u into overlapping patches and vectorizes each path. By exploiting the redundancy of seismic data, û may be expressed by a multiplication of a dictionary D and sparse coefficient matrix x, per formula (2) shown below:
û=Dx (2)
As noted above, methods described herein to recover relatively weak signals is based on applying machine learning. According to an example of such machine learning process, which may also be referred to as an unsupervised machine learning, a dictionary D may be trained using û. Accordingly, at S206, processing center 110 applies the initial filtering result (û) to train dictionary D. In one example, training dictionary D is based on formula (2) below:
minD,x∥û−Dx∥F2s.t.∥x∥0≤T. (3)
In formula (3), T is a sparse threshold, T represents the sparsity threshold, ∥⋅∥F denotes the Frobenius norm, and ∥⋅∥0, referred to as the l0 norm counts the number of nonzero entries. Formula (3) is an example of an optimization problem can be solved using a K-Singular Value Decomposition (K-SVD) algorithm. Adopting the idea of block coordinate descent, the K-SVD algorithm can reduce the l0 minimization problem into two subproblems—sparse coding and dictionary updating. The sparse coding aspect includes calculating the sparse coefficients through orthogonal matching pursuit (OMP) or a variant of OMP for a given dictionary D, while the dictionary updating step updates atoms by solving a series of rank-1 approximations to decrease the data misfit. Dictionary updating involves performing singular value decomposition repeatedly. To reduce the complexity of dictionary learning, exact singular value decomposition may be replaced with randomized singular value decomposition, or the solution thereof approximated by taking a single iteration of gradient descent.
At S208, processing center 110 updates dictionary D with the solution to the optimization problem given by formula (3). Accordingly, as more and more data is collected, dictionary D is trained, which will then be used to extract relatively weak signals that may otherwise be filtered with noise using conventional noise filtering/attenuation processes. The accuracy of the residual signal recovery using dictionary D increases as more and more data is used to train dictionary D.
At S210, processing center 110 stores the updated dictionary Din a memory or an associated database of processing center 110.
At S212, processing center 110 applies the stored dictionary D to derive a residual signal, which is a difference between d and û, given by formula (4) below:
{circumflex over (d)}=d−û (4)
which denotes the residual signal with noise. For random Gaussian noise, we can employ a standard sparse inversion procedure for recovery. The sparse representation of the residual signal, x1, can be obtained by solving:
minx
where ϵ is the error threshold dictated by the noise variance, and ∥⋅∥1 denotes the l1 norm used to impose sparsity with robustness. However, the assumption of Gaussian noise is rarely held for field data and, therefore, solving the standard sparse inversion problem may provide suboptimal results.
This residual signal, is a signal that contains the relatively weak signals (e.g., weak seismic signal relative to the magnitude of the contaminating noise signal) and therefore by applying dictionary D thereto, the relatively weak signals can be recovered and added to û to obtain a final denoised signal. This final denoised signal has more signals, including the relatively weak signals, retained therein, when compared to the final denoised signals obtained by apply conventional filtering methods to d, as previously discussed.
In one example and in applying dictionary D, an assumption is made that the initial filtering result (i.e., û) and the residual signal (i.e., {circumflex over (d)}) share the same set of basis functions.
In one example, to better separate the residual signal from the noise, processing center 110 may invert both (the residual signal and the noise) simultaneously using a dual-domain sparse inversion processing given by formula (6) below:
minx
In formula (6), x1 and x2 are sparse representations of the residual signal and the noise signal (i.e., e), S is the chosen sparsity basis for the noise, ε is a configuration error tolerance, which may be determined according to experiments and/or empirical studies indicative of the collected data (e.g., seismic data) and ∥x1∥0/∥x2∥0 may be referred to as the l0-norm. In one example, sparsity of the noise and residual signals may be imposed using l1-norm instead of l0-norm.
With x1 derived per formula (6), at S212, processing center 110 applies D to x1 to derive (determine) the residual signal (e.g., Dx1). Thereafter, at S214, processing center 110 may determine a final filtered signal, which can be obtained by addition of Dx1 to û. This final filtered signal may be referred to as uout given by formula (7) below:
uout=û+Dx1 (7)
The final filtered (denoised) signal derived per process of
The above process described with reference to
The proposed method of
For this dual dictionary training based method, assume that a noise estimate can be attained by applying the conventional method, denoted by g; i.e.,
{circumflex over (n)}=g(d). (8)
Next, the noise model 11 may be partitioned/rearranged in a same way as the signal model û described above with reference to S205 of
The dual dictionary learning may require additional estimate of the noise model and one more training step prior to recovery, which unavoidably increases the computation cost. However, the method provides an effective alternative for attenuating some of the most complex noise.
The proposed method with single or dual dictionary learning differs from many residual signal recovery methods since it requires no assumptions on local similarities between the initial estimate and the residual signal. A common set of bases functions in a global context suffices to assure the recovery.
One advantage of determining the final filtered signal per process of
The collected data is first sorted into offset-vector tiles (OVTs) shown in graph 400. According to process of
Results of graph 402 is then fed into the above described dictionary to derive the residual signal shown in graph 406 (S206 to S212 of
Graph 410 illustrates the difference between the final denoised result of graph 408 and the initial collected data shown in graph 400.
Comparing graphs 404 and 406, it can be observed that relatively weak signals indicative of seismic activity are extracted with only a fraction of noise remaining.
For initial noise attenuation, envelop soft mute followed by windowed Fourier domain filtering (WFT) is employed to preserve flat or near-flat events after static and moveout corrections. Graph 502 shows the initial denoised result which was served as the signal estimate for training. To generate the initial noise estimate, envelop soft mute was applied again on the differences to extract the high-amplitude portion of the scattered and ground roll noise, as shown in graph 504. Both models are next input for dictionary learning to adaptively form D and S, described above. Graphs 506 and 508 displays the subsets of learned dictionaries—atoms with distinct characters were trained which enables further separation of signal and noise. Graph 510 shows the recovered residual signal by incorporating dual dictionaries in inversion, and graph 512 shows the final denoised result with signal recovery.
We next stacked the data for further quality control (QC), the result of which is shown in
Signal recovery using single and dual dictionary learning are described above. However, the present disclosure is not limited thereto. Any number of N dictionaries may be used where N is a positive integer greater than 2 conditions on input signal being partitioned into N distinct signals/noises, where sparse invention can be applied to such N dictionaries to recover underlying data signal.
Examples described above provide an unsupervised machine learning and sparse inversion based method for recovery of weak signals indicative of seismic activity (weak relative to existing noise signals) that would otherwise be lost during noise filtering/attenuation process utilized by current signal recovery methods. These examples are equally applicable to both coherent and incoherent noise on pre-stack and/or stacked images.
Having described examples of machine learning based signal recovery with application to signals indicative of seismic activity and turning to
To enable user interaction with system 700, an input device 745 can represent any number of input mechanisms, such as a microphone for speech, a touch-sensitive screen for gesture or graphical input, keyboard, mouse, motion input, speech and so forth. An output device 735 can also be one or more of a number of output mechanisms known to those of skill in the art. In some instances, multimodal systems can enable a user to provide multiple types of input to communicate with system 700. Communications interface 740 can generally govern and manage the user input and system output. There is no restriction on operating on any particular hardware arrangement and therefore the basic features here may easily be substituted for improved hardware or firmware arrangements as they are developed.
Storage device 730 is a non-volatile memory and can be a hard disk or other types of computer readable media which can store data that are accessible by a computer, such as magnetic cassettes, flash memory cards, solid state memory devices, digital versatile disks, cartridges, random access memories (RAMs) 725, read only memory (ROM) 720, and hybrids thereof.
Storage device 730 can include service 1732, service 2734 and/or service 3736 for execution by processor 710 to cause processor 710 to carryout functionalities described above with reference to
For clarity of explanation, in some instances the present technology may be presented as including individual functional blocks including functional blocks comprising devices, device components, steps or routines in a method embodied in software, or combinations of hardware and software.
In some embodiments the computer-readable storage devices, mediums, and memories can include a cable or wireless signal containing a bit stream and the like. However, when mentioned, non-transitory computer-readable storage media expressly exclude media such as energy, carrier signals, electromagnetic waves, and signals per se.
Methods according to the above-described examples can be implemented using computer-executable instructions that are stored or otherwise available from computer readable media. Such instructions can comprise, for example, instructions and data which cause or otherwise configure a general purpose computer, special purpose computer, or special purpose processing device to perform a certain function or group of functions. Portions of computer resources used can be accessible over a network. The computer executable instructions may be, for example, binaries, intermediate format instructions such as assembly language, firmware, or source code. Examples of computer-readable media that may be used to store instructions, information used, and/or information created during methods according to described examples include magnetic or optical disks, flash memory, USB devices provided with non-volatile memory, networked storage devices, and so on.
Devices implementing methods according to the present inventive concept can comprise hardware, firmware and/or software, and can take any of a variety of form factors. Typical examples of such form factors include laptops, smart phones, small form factor personal computers, personal digital assistants, rackmount devices, standalone devices, and so on. Functionality described herein also can be embodied in peripherals or add-in cards. Such functionality can also be implemented on a circuit board among different chips or different processes executing in a single device, by way of further example.
The instructions, media for conveying such instructions, computing resources for executing them, and other structures for supporting such computing resources are means for providing the functions described herein with respect to the present inventive concept.
Although a variety of examples and other information was used to explain aspects within the scope of the appended claims, no limitation of the claims should be implied based on particular features or arrangements in such examples, as one of ordinary skill would be able to use these examples to derive a wide variety of implementations. Further and although some subject matter may have been described in language specific to examples of structural features and/or method steps, it is to be understood that the subject matter defined in the appended claims is not necessarily limited to these described features or acts. For example, such functionality can be distributed differently or performed in components other than those identified herein. Rather, the described features and steps are disclosed as examples of components of systems and methods within the scope of the appended claims.
The previous description is provided to enable any person skilled in the art to practice the various aspects described herein. Various modifications to these aspects will be readily apparent to those skilled in the art, and the generic principles defined herein may be applied to other aspects. Thus, the claims are not intended to be limited to the aspects shown herein, but is to be accorded the full scope consistent with the language of the claims, wherein reference to an element in the singular is not intended to mean “one and only one” unless specifically so stated, but rather “one or more.” Unless specifically stated otherwise, the term “some” refers to one or more. A phrase referring to “at least one of” a list of items in the claims and/or specification refers to any combination of those items, including single members or multiple members. As an example, “at least one of a, b, and c” is intended to cover a; b; c; a and b; a and c; b and c; or a, b and c.
This application claims priority to U.S. Provisional Application No. 62/739,260 filed on Sep. 30, 2018, the entire content of which is incorporated herein by reference.
Number | Name | Date | Kind |
---|---|---|---|
2906363 | Clay, Jr. et al. | Sep 1959 | A |
3747055 | Greene, Jr. | Jul 1973 | A |
3747056 | Treybig et al. | Jul 1973 | A |
3840845 | Brown | Oct 1974 | A |
3877033 | Unz | Apr 1975 | A |
4330873 | Peterson | May 1982 | A |
4404684 | Takada | Sep 1983 | A |
4509151 | Anderson | Apr 1985 | A |
4553221 | Hyatt | Nov 1985 | A |
4559605 | Norsworthy | Dec 1985 | A |
4596005 | Frasier | Jun 1986 | A |
4597066 | Frasier | Jun 1986 | A |
4721180 | Haughland et al. | Jan 1988 | A |
4852004 | Manin | Jul 1989 | A |
4958331 | Wardle | Sep 1990 | A |
4967400 | Woods | Oct 1990 | A |
4992990 | Langeland et al. | Feb 1991 | A |
5079703 | Mosher et al. | Jan 1992 | A |
5092423 | Petermann | Mar 1992 | A |
5148406 | Brink et al. | Sep 1992 | A |
5168472 | Lockwood | Dec 1992 | A |
5353223 | Norton et al. | Oct 1994 | A |
5469404 | Barber et al. | Nov 1995 | A |
5487052 | Cordsen | Jan 1996 | A |
5517463 | Hornbostel et al. | May 1996 | A |
5724306 | Barr | Mar 1998 | A |
5774417 | Corrigan et al. | Jun 1998 | A |
5787051 | Goodway et al. | Jul 1998 | A |
5835450 | Russell | Nov 1998 | A |
5973995 | Walker et al. | Oct 1999 | A |
6009042 | Workman et al. | Dec 1999 | A |
6493636 | DeKok | Dec 2002 | B1 |
6509871 | Bevington | Jan 2003 | B2 |
6590831 | Bennett et al. | Jul 2003 | B1 |
6691038 | Zajac | Feb 2004 | B2 |
6876599 | Combee | Apr 2005 | B1 |
7167412 | Tenghamn | Jan 2007 | B2 |
7234407 | Levine et al. | Jun 2007 | B1 |
7359283 | Vaage et al. | Apr 2008 | B2 |
7408836 | Muyzert et al. | Aug 2008 | B2 |
7451717 | Levine et al. | Nov 2008 | B1 |
7499374 | Ferber | Mar 2009 | B2 |
7499737 | Mizuta et al. | Mar 2009 | B2 |
7515505 | Krohn et al. | Apr 2009 | B2 |
7545703 | Lunde et al. | Jun 2009 | B2 |
7646671 | Pan et al. | Jan 2010 | B2 |
7993164 | Chatterjee et al. | Aug 2011 | B2 |
8509027 | Strobbia et al. | Aug 2013 | B2 |
8559270 | Abma | Oct 2013 | B2 |
8619497 | Sallas et al. | Dec 2013 | B1 |
8681581 | Moldoveanu et al. | Mar 2014 | B2 |
8711654 | Moldoveanu et al. | Apr 2014 | B2 |
8737184 | Yamazaki | May 2014 | B2 |
8897094 | Eick et al. | Nov 2014 | B2 |
9110177 | Opfer | Aug 2015 | B1 |
9291728 | Eick et al. | Mar 2016 | B2 |
9529102 | Eick et al. | Dec 2016 | B2 |
9632193 | Li et al. | Apr 2017 | B2 |
9690003 | Sallas | Jun 2017 | B2 |
9823372 | Eick et al. | Nov 2017 | B2 |
9846248 | Eick et al. | Dec 2017 | B2 |
10267939 | Eick et al. | Apr 2019 | B2 |
10267940 | Nguyen | Apr 2019 | B2 |
10514474 | Eick et al. | Dec 2019 | B2 |
10605941 | Li et al. | Mar 2020 | B2 |
10809402 | Li et al. | Oct 2020 | B2 |
10823867 | Eick et al. | Nov 2020 | B2 |
10989826 | Eick et al. | Apr 2021 | B2 |
11035968 | Li et al. | Jun 2021 | B2 |
20040172199 | Chavarria et al. | Sep 2004 | A1 |
20050088914 | Ren et al. | Apr 2005 | A1 |
20060164916 | Krohn et al. | Jul 2006 | A1 |
20060190181 | Deffenbaugh | Aug 2006 | A1 |
20060268682 | Vasseur | Nov 2006 | A1 |
20070013546 | McConnell et al. | Jan 2007 | A1 |
20070025182 | Robertsson | Feb 2007 | A1 |
20070027656 | Baraniuk et al. | Feb 2007 | A1 |
20070027660 | Swift | Feb 2007 | A1 |
20070276660 | Pinto | Nov 2007 | A1 |
20080008037 | Welker | Jan 2008 | A1 |
20080049551 | Muyzert et al. | Feb 2008 | A1 |
20080080309 | Elkington et al. | Apr 2008 | A1 |
20080089174 | Sollner et al. | Apr 2008 | A1 |
20080144434 | Hegna et al. | Jun 2008 | A1 |
20080151688 | Goujon | Jun 2008 | A1 |
20080205193 | Krohn et al. | Aug 2008 | A1 |
20080225642 | Moore et al. | Sep 2008 | A1 |
20080240203 | Baqai | Oct 2008 | A1 |
20080285380 | Rouquette | Nov 2008 | A1 |
20090006053 | Carazzone et al. | Jan 2009 | A1 |
20090010101 | Lunde et al. | Jan 2009 | A1 |
20090067285 | Robertsson et al. | Mar 2009 | A1 |
20090073805 | Tulett et al. | Mar 2009 | A1 |
20090092006 | Teigen et al. | Apr 2009 | A1 |
20090122641 | Hillesund et al. | May 2009 | A1 |
20090141587 | Welker et al. | Jun 2009 | A1 |
20090213693 | Du et al. | Aug 2009 | A1 |
20090231956 | Schonewille | Sep 2009 | A1 |
20090251992 | Van Borselen et al. | Oct 2009 | A1 |
20090262601 | Hillesund et al. | Oct 2009 | A1 |
20090279384 | Pavel | Nov 2009 | A1 |
20090279386 | Monk | Nov 2009 | A1 |
20090323472 | Howe | Dec 2009 | A1 |
20100002536 | Brewer et al. | Jan 2010 | A1 |
20100103772 | Eick et al. | Apr 2010 | A1 |
20100128563 | Strobbia et al. | May 2010 | A1 |
20100195434 | Menger et al. | Aug 2010 | A1 |
20100208554 | Chiu et al. | Aug 2010 | A1 |
20100211321 | Ozdemir et al. | Aug 2010 | A1 |
20100265799 | Cevher et al. | Oct 2010 | A1 |
20100299070 | Abma | Nov 2010 | A1 |
20110019502 | Eick et al. | Jan 2011 | A1 |
20110038227 | Kostov et al. | Feb 2011 | A1 |
20110128818 | Eick et al. | Jun 2011 | A1 |
20110156494 | Mashinsky | Jun 2011 | A1 |
20110170796 | Qian | Jul 2011 | A1 |
20110218737 | Gulati | Sep 2011 | A1 |
20110286302 | Welker et al. | Nov 2011 | A1 |
20110305106 | Eick et al. | Dec 2011 | A1 |
20110305107 | Eick et al. | Dec 2011 | A1 |
20110305113 | Eick et al. | Dec 2011 | A1 |
20110307438 | Fernández Martínez | Dec 2011 | A1 |
20110317517 | Borresen et al. | Dec 2011 | A1 |
20120002503 | Janiszewski et al. | Jan 2012 | A1 |
20120014212 | Eick et al. | Jan 2012 | A1 |
20120051181 | Eick et al. | Mar 2012 | A1 |
20120082004 | Boufounos | Apr 2012 | A1 |
20120113745 | Eick et al. | May 2012 | A1 |
20120143604 | Singh | Jun 2012 | A1 |
20120281499 | Eick et al. | Nov 2012 | A1 |
20120294116 | Kamata | Nov 2012 | A1 |
20120300585 | Cao et al. | Nov 2012 | A1 |
20130135966 | Rommel et al. | May 2013 | A1 |
20130250720 | Monk et al. | Sep 2013 | A1 |
20130294194 | Pritchard | Nov 2013 | A1 |
20140133271 | Sallas | May 2014 | A1 |
20140146638 | Renaud | May 2014 | A1 |
20140211590 | Sallas | Jul 2014 | A1 |
20140278289 | Etgen | Sep 2014 | A1 |
20140303898 | Poole | Oct 2014 | A1 |
20140334260 | Chang | Nov 2014 | A1 |
20140362663 | Jones et al. | Dec 2014 | A1 |
20150016218 | Welker et al. | Jan 2015 | A1 |
20150078128 | Eick et al. | Mar 2015 | A1 |
20150124560 | Li et al. | May 2015 | A1 |
20150272506 | Childs et al. | Oct 2015 | A1 |
20150348568 | Li et al. | Dec 2015 | A1 |
20160018547 | Eick et al. | Jan 2016 | A1 |
20160341839 | Kazinnik et al. | Nov 2016 | A1 |
20170031045 | Poole et al. | Feb 2017 | A1 |
20170059732 | Björnemo | Mar 2017 | A1 |
20170068004 | Takanashi | Mar 2017 | A1 |
20170082761 | Li et al. | Mar 2017 | A1 |
20170090053 | Eick et al. | Mar 2017 | A1 |
20170090054 | Willis | Mar 2017 | A1 |
20170108604 | Turquais | Apr 2017 | A1 |
20180003846 | Wago | Jan 2018 | A1 |
20180067221 | Eick et al. | Mar 2018 | A1 |
20180196155 | Nguyen | Jul 2018 | A1 |
20180259664 | Li | Sep 2018 | A1 |
20180335536 | Li et al. | Nov 2018 | A1 |
20190015052 | Scalzo | Jan 2019 | A1 |
20190129050 | Li et al. | May 2019 | A1 |
20190293813 | Li et al. | Sep 2019 | A1 |
20190310387 | Eick et al. | Oct 2019 | A1 |
20190385094 | Alex | Dec 2019 | A1 |
20200225377 | Li et al. | Jul 2020 | A1 |
20210033741 | Li et al. | Feb 2021 | A1 |
Number | Date | Country |
---|---|---|
103954993 | Jul 2014 | CN |
2103959 | Sep 2009 | EP |
2592439 | May 2013 | EP |
WO-2005019865 | Mar 2005 | WO |
WO-2008073178 | Jun 2008 | WO |
WO-2009092025 | Jul 2009 | WO |
WO-2010149589 | Dec 2010 | WO |
WO-2011156491 | Dec 2011 | WO |
WO-2011156494 | Dec 2011 | WO |
WO-2012166737 | Dec 2012 | WO |
WO-2013105075 | Jul 2013 | WO |
WO-2014057440 | Apr 2014 | WO |
WO-2014195508 | Dec 2014 | WO |
WO-2014195508 | Mar 2015 | WO |
WO-2015066481 | May 2015 | WO |
WO-2016009270 | Jan 2016 | WO |
WO-2018085567 | May 2018 | WO |
Entry |
---|
Ala'i R., “Shallow Water Multiple Prediction and Attenuation, case study on data from the Arabian Gulf,” SEG International Exposition and 72nd Annual Meeting, Salt Lake City, Utah, Oct. 6-11, 2002, 4 pages. |
Almendros J., et al., “Mapping the Sources of the Seismic Wave Field at Kilauea Volcano, Hawaii, Using Data Recorded on Multiple Seismic Antennas,” Bulletin of the Seismological Society of America, vol. 92(6), Aug. 2002, pp. 2333-2351. |
Amir V., et al., “Structural Evolution Of The Northern Bonaparte Basin, Northwest Shelf Australia,” Proceedings, Indonesian Petroleum Association, Thirty-Fourth Annual Convention & Exhibition, May 2010, 17 Pages. |
Baraniuk R.G., “Compressive Sensing,” IEEE Signal Processing Magazine, Jul. 2007, vol. 24(4), 9 pages. |
Barzilai J., et al., “Two Point Step Size Gradient Methods,” IMA Journal of Numerical Analysis, 1988, vol. 8, pp. 141-148. |
Buia M., et al., “Shooting Seismic Surveys in Circles,” Oilfield Review, 2008, pp. 18-31. |
Candes E., et al., “Sparsity and Incoherence in Compressive Sampling,” Applied and Computational Mathematics, Caltech, Pasadena, CA 91125 and Electrical and Computer Engineering, Georgia Tech, Atlanta, GA 90332, Nov. 2006, 20 pages. |
Carlson D., et al., “Increased Resolution and Penetration from a Towed Dual-Sensor Streamer”, First Break, Dec. 2007, vol. 25, pp. 71-77. |
Cordsen A., et al., “Planning Land 3D Seismic Surveys,” Geophysical Developments Series No. 9, Society of Exploration Geophysicists (SEG), Jan. 2000, 16 pages. |
Dragoset B., et al., “A Perspective on 3D Surface-Related Multiple Elimination”, Geophysics, Sep.-Oct. 2010, vol. 75, No. 5, pp. 75A245-75A261. |
Foster D.J., et al., “Suppression of Multiple Reflections Using the Radon Transform”, Mar. 1992, Geophysics, vol. 57, No. 3, pp. 386-395. |
Hennenfent G., et al., “Application of Stable Signal Recovery to Seismic Data Interpolation,” Gilles Hennenfent and Felix J. Herrmann Earth & Ocean Sciences Dept., University of British Columbia 2006, 4 pages. |
Hennenfent G., et al., “Simply Denoise: Wavefield Reconstruction via Jittered undersampling,” Geophysics, May-Jun. 2008, vol. 73(3), pp. V19-V28. |
Herrmann F.J., “Randomized Sampling and Sparsity: Getting More Information from Fewer Samples,” Geophysics, vol. 75(6), Nov.-Dec. 2010, pp. WB173-WB187. |
Hindriks K., et al., “Reconstruction of 3D Seismic Signals Irregularly Sampled Along Two Spatial Coordinates,” Geophysics, Jan.-Feb. 2000, vol. 65(1), pp. 253-263. |
Huang H., et al., “Joint SRME and Model-Based Water-Layer Demultiple for Ocean Bottom Node”, 2016 SEG International Exposition and Annual Meeting, Retrieved from Internet: URL: https://www.cgg.com/sites/default/files/2020-11/cggv_0000026243.pdf, pp. 4508-4512. |
International Search Report and Written Opinion for Application No. PCT/US11/039640, dated Oct. 26, 2011, 8 Pages. |
International Search Report for Application No. PCT/US2016/053750, dated Dec. 27, 2016, 2 Pages. |
International Search Report for Application No. PCT/US2017/59760, dated Apr. 13, 2018, 2 pages. |
Jin H., et al., “MWD for Shallow Water Demultiple: A Hibernia Case Study,” Geo Convention 2012: Vision, 5 Pages. |
Kumar R., et al., “Source Separation for Simultaneous Ttowed-Streamer Marine Acquisition—A Compressed Sensing Approach,” Geophysics, vol. 80(6), Nov.-Dec. 2015, pp. WD73-WD88. |
Li C., et al., “A Multi-Stage Inversion Method for Simultaneous Source Deblending of Field Data,” SEG Annual Meeting 2014, Denver, Colorado, USA, Oct. 26, 2014, pp. 3610-3615. |
Li C., et al., “Aspects of Implementing Marine Blended Source Acquisition in the Field,” SEG International Exposition and 87th Annual Meeting, 2017, pp. 42-46. |
Li C., et al., “Improving Streamer Data Sampling and Resolution via Non-Uniform Optimal Design and Reconstruction,” SEG International Exposition and 87th Annual Meeting, 2017, pp. 4241-4245. |
Li C., et al., “Interpolated Compressive Sensing for Seismic Data Reconstruction,” SEG Las Vegas 2012 Annual Meeting, 2012, 6 pages. |
Li C., et al., “Joint Source Deblending and Reconstruction for Seismic Data,” SEG Houston 2013 Annual Meeting, 2013, pp. 82-87. |
Li C., et al., “Marine Towed Streamer Data Reconstruction Based on Compressive Sensing,” SEG Houston 2013 Annual Meeting, 2013, pp. 3597-3602. |
Lin D., et al., “3D SRME Prediction and Subtraction Practice for Better Imaging”, 2005, SEG Houston Annual Meeting, 5 pgs. |
Liu B., et al., “Minimum Weighted Norm Interpolation of Seismic Records,” Geophysics, Nov.-Dec. 2004, vol. 69(6), pp. 1560-1568. |
Mahdad A., et al., “Separation of Blended Data by Iterative Estimation and Subtraction of Blending Interference Noise,” Geophysics, vol. 76(3), May-Jun. 2011, pp. Q9-Q17. |
Martin J., et al., “Acquisition of Marine Point Receiver Seismic Data With a Towed Streamer,” SEG Technical Program Expanded Abstracts 2000, 4 pages. |
Maurer H., et al., “Recent advances in optimized geophysical survey design,” Seismic Data Acquisition, Geophysics, Sep.-Oct. 2010, vol. 75(5), SEG Press Book, pp. 75A177-75A194. |
Memorandum Opinion and Order, ConocoPhillips Company v. In-Depth Compressive Seismic, Inc., et al., Civil Action No. H-18-0803, entered Apr. 26, 2019, 49 pgs. |
Milton A., et al., “Reducing Acquisition Costs with Random Sampling and Multidimensional Interpolation,” SEG San Antonio 2011 Annual Meeting, 2011, pp. 52-56. |
Moldoveanu N., “Random Sampling: A New Strategy for Marine Acquisition,” SEG Expanded Abstracts, Denver, CO, 2010 Annual Meeting, 2010, pp. 51-55. |
Mosher C., et al., “Increasing the Efficiency of Seismic Data Acquisition Via Compressive Sensing,” Offshore Technology conference, Asia, Kuala Lumpur, Malaysia, Mar. 25-28, 2014, 4 pages. |
Mosher C.C., et al., “An In-situ Analysis of 3-D Seismic Lateral Resolution,” Borehole Geophysics, BHG 6.3, 1985, pp. 109-111. |
Mosher C.C., et al., “Compressive Seismic Imaging: Moving from research to production,” SEG International Exposition and 87th Annual Meeting, 2017, pp. 74-78. |
Mosher C.C., et al., “Compressive Seismic Imaging,” SEG Las Vegas 2012 Annual Meeting, 2012, DOI http://dx.doi.org/10.1190/segam2012-1460.1, 5 pages. |
Mosher C.C., et al., “Non-Uniform Optimal Sampling for Seismic Survey Design,” 74th EAGE Conference and Exhibition, Extended Abstracts, X034, Copenhagen, Denmark, Jun. 4-7, 2012, 5 pages. |
Mosher C.C., et al., “Non-Uniform Optimal Sampling for Simultaneous Source Survey Design,” SEG Annual Meeting, 2014, pp. 105-109. |
Mosher C.C., “Generalized Windowed Transforms for Seismic Processing and Imaging,” 2012 Annual SEG Meeting Las Vegas, Nevada, One Petro, SEG-2012-1196, Published by Society of Exploration Geophysicists, 4 pages. |
Musser J.A., et al., “Streamer Positioning and Spread Stabilization for 4D Seismic,” SEG 2006 Annual Meeting, New Orleans, 2006, 4 pages. |
Sacchi M.D., “A Tour of High Resolution Transforms,” Frontiers & Innovation, CSPG, CSEG, CWLS Convention, Calgary, Alberta, Canada, Expanded Abstracts, 2009, pp. 665-668. |
Shapiro H.S., et al., “Alias-Free Sampling of Random Noise,” SIAM Journal on Applied Mathematics, 1960, vol. 8(2), pp. 225-248. |
Stolt R.H., “Seismic Data Mapping and Reconstruction,” Geophysics, May-Jun. 2002, vol. 67(3), pp. 890-908. |
Thomsen L., “Weak Elastic Anisotropy”, Geophysics, Oct. 1986, vol. 51, No. 10, Society of Exploration Geophysicists, pp. 1954-1966. |
Trad D., “Interpolation and Multiple Attenuation with Migration Operators,” Geophysics, vol. 68(6), Nov.-Dec. 2003, pp. 2043-2054. |
Wang L., et al., “Distributed Reconstruction via Alternating Direction Method,” Hindawi Publishing Corporation, Computational and Mathematical Methods in Medicine, 2013, vol. 2013, Article ID 418747, pp. 1-7. |
Wang P., et al., “Model-Based Water-Layer Demultiple”, 2011, SEG San Antonio Annual Meeting, pp. 3551-3555. |
Wang Y., et al., “Recovery of Seismic Wavefields based on Compressive Sensing by an I1-norm Constrained Trust Region Method and the Piecewise Random Subsampling,” Geophysical Journal International, 2011(187), pp. 199-213. |
Zhang H., et al., “A Nonmonotone Line Search Technique and its Application to Unconstrained Optimization,” Society of Industrial and Applied Mathematics, 2004, vol. 14(4), pp. 1043-1056. |
Zwartjes P.M., et al., “Fourier Reconstruction of Non-uniformly Sampled, Aliased Data,” SEG Int'l Exposition and 74th Annual Meeting, Denver, Colorado, Oct. 10-15, 2004, 4 pages. |
Zwartjes P.M., et al., “Fourier Reconstruction of Nonuniformly Sampled, Aliased Seismic Data,” Geophysics, Jan.-Feb. 2007, vol. 72(1), pp. V21-V32. |
International Preliminary Report on Patentability for for International application No. PCT/US2019/053174, dated Apr. 8, 2021, 7 pages. |
International Search report and Written opinion for International application No. PCT/US2019/053174, dated Dec. 10, 2019, 8 pages. |
Lotter T., et al., “Noise Reduction by Maximum a Posteriori Spectral Amplitude Estimation With Supergaussian Speech Modeling,” International Workshop on Acoustic Echo and Noise Control (IWAENC2003), Kyoto, Japan, retrieved from URL: https://pdfs.semanticscholar.org/06e2/ad185cc5a809bb7493f8aea8afdad13105fb.pdf, on Nov. 16, 2019, Sep. 2003, pp. 83-86. |
Extended European Search Report for Application No. 19866442.7 dated Aug. 18, 2022 (8 pages). |
Lingchen Zhu et al., “Seismic data denoising through multiscale and sparsity-promoting dictionary learning”, Geophysics, vol. 80, No. 6, Nov. 1, 2015 (Nov. 1, 2015), pp. WD45-WD57, XP057734071, US ISSN:0016-8033, DOI:10.1190/geo2015-0047.1. |
Number | Date | Country | |
---|---|---|---|
20200104745 A1 | Apr 2020 | US |
Number | Date | Country | |
---|---|---|---|
62739260 | Sep 2018 | US |