Generally, the present disclosure relates to the field of data processing. The present disclosure pertains to the domain of Partial Differential Equation (PDE)-constrained optimization, with a special focus on wave phenomena optimization. More specifically, the present disclosure relates to methods, systems, apparatuses, and devices for facilitating waveform inversion using a Sifr optimizer.
Waveform imaging techniques hold a pivotal role in a variety of sectors, including but not limited to healthcare, construction, and resource exploration. These techniques are designed to construct high-resolution models of investigated mediums, such as geological formations or human tissues, based on the principles of wave propagation. The core of this process is essentially an optimization problem. The primary objective is to minimize the discrepancy between the observed data, and the synthetic data, generated through computational models. While first-order optimization methods like gradient descent are prevalent, second-order methods offer the advantage of faster convergence rates. However, these second-order methods come with their own set of computational challenges, particularly the complexities associated with handling the Hessian matrix.
One of the most significant challenges in waveform imaging is the phenomenon known as cycle skipping. This is caused by the oscillatory nature of waves and can lead to minimal mismatches between observed and synthetic data, even when the waveforms are out of phase. Also, an inversion that offers second-order optimization in waveform imaging and employs a functional which may be a zero-valued functional that allows for the efficient incorporation of second-order derivative information into the optimization process, effectively bypassing the computational burden associated with the Hessian matrix grapples with the issue of cycle skipping, especially when low-frequency components are not adequately present.
At the heart of waveform imaging, there are optimization techniques. While gradient descent methods are commonly used for model refinement, second-order methods offer the promise of faster convergence. However, these methods are computationally demanding due to the sheer size and complexity of the Hessian matrix. Furthermore, existing methods like BFGS and Hessian-free approaches for optimization have huge computation drawbacks.
Therefore, there is a need for improved methods, systems, apparatuses, and devices for facilitating waveform inversion using a Sifr optimizer that may overcome one or more of the above-mentioned problems and/or limitations.
This summary is provided to introduce a selection of concepts in a simplified form, that are further described below in the Detailed Description. This summary is not intended to identify key features or essential features of the claimed subject matter. Nor is this summary intended to be used to limit the claimed subject matter's scope.
Disclosed herein is a method for facilitating waveform inversion using a Sifr optimizer, in accordance with some embodiments. Accordingly, the method may include a step of receiving, using a communication device, at least one observed data associated with at least one object from at least one device. Further, the method may include a step of obtaining, using a processing device, at least one initial model associated with the at least one object. Further, the method may include a step of obtaining, using the processing device, at least one synthetic data from the at least one initial model. Further, the method may include a step of determining, using the processing device, a discrepancy between the at least one observed data and the at least one synthetic data. Further, the method may include a step of creating, using the processing device, a cost function based on the determining of the discrepancy. Further, the method may include a step of updating, using the processing device, the at least one initial model iteratively using the Sifr optimizer until at least one condition may be met. Further, the at least one condition may include at least one of a minimization of the cost function below a threshold, an elapsing of a given number of iterations, and a meeting of a convergence criterion. Further, the Sifr optimizer provides an update for the updating of the at least one initial model iteratively based on a regression comprising a least squares resolution of at least one Sifr equation. Further, the method may include a step of generating, using the processing device, at least one final model based on the updating. Further, the method may include a step of storing, using a storage device, the at least one final model.
Further, disclosed herein is a system for facilitating waveform inversion using a Sifr optimizer, in accordance with some embodiments. Accordingly, the system may include a communication device configured for receiving at least one observed data associated with at least one object from at least one device. Further, the system may include a processing device communicatively coupled with the communication device. Further, the processing device may be configured for obtaining at least one initial model associated with the at least one object. Further, the processing device may be configured for obtaining at least one synthetic data from the at least one initial model. Further, the processing device may be configured for determining a discrepancy between the at least one observed data and the at least one synthetic data. Further, the processing device may be configured for creating a cost function based on the determining of the discrepancy. Further, the processing device may be configured for updating the at least one initial model iteratively using the Sifr optimizer until at least one condition may be met. Further, the at least one condition may include a minimization of the cost function below a threshold, an elapsing of a given number of iterations, and a meeting of a convergence criterion. Further, the Sifr optimizer provides an update for the updating of the at least one initial model iteratively based on a regression comprising a least squares resolution of at least one Sifr equation. Further, the processing device may be configured for generating at least one final model based on the updating. Further, the system may include a storage device communicatively coupled with the processing device. Further, the storage device may be configured for storing the at least one final model.
Both the foregoing summary and the following detailed description provide examples and are explanatory only. Accordingly, the foregoing summary and the following detailed description should not be considered to be restrictive. Further, features or variations may be provided in addition to those set forth herein. For example, embodiments may be directed to various feature combinations and sub-combinations described in the detailed description.
The patent or application file contains at least one drawing executed in color. Copies of this patent or patent application publication with color drawing(s) will be provided by the Office upon request and payment of the necessary fee.
The accompanying drawings, which are incorporated in and constitute a part of this disclosure, illustrate various embodiments of the present disclosure. The drawings contain representations of various trademarks and copyrights owned by the Applicants. In addition, the drawings may contain other marks owned by third parties and are being used for illustrative purposes only. All rights to various trademarks and copyrights represented herein, except those belonging to their respective owners, are vested in and the property of the applicants. The applicants retain and reserve all rights in their trademarks and copyrights included herein, and grant permission to reproduce the material only in connection with reproduction of the granted patent and for no other purpose.
Furthermore, the drawings may contain text or captions that may explain certain embodiments of the present disclosure. This text is included for illustrative, non-limiting, explanatory purposes of certain embodiments detailed in the present disclosure.
As a preliminary matter, it will readily be understood by one having ordinary skill in the relevant art that the present disclosure has broad utility and application. As should be understood, any embodiment may incorporate only one or a plurality of the above-disclosed aspects of the disclosure and may further incorporate only one or a plurality of the above-disclosed features. Furthermore, any embodiment discussed and identified as being “preferred” is considered to be part of a best mode contemplated for carrying out the embodiments of the present disclosure. Other embodiments also may be discussed for additional illustrative purposes in providing a full and enabling disclosure. Moreover, many embodiments, such as adaptations, variations, modifications, and equivalent arrangements, will be implicitly disclosed by the embodiments described herein and fall within the scope of the present disclosure.
Accordingly, while embodiments are described herein in detail in relation to one or more embodiments, it is to be understood that this disclosure is illustrative and exemplary of the present disclosure, and are made merely for the purposes of providing a full and enabling disclosure. The detailed disclosure herein of one or more embodiments is not intended, nor is to be construed, to limit the scope of patent protection afforded in any claim of a patent issuing here from, which scope is to be defined by the claims and the equivalents thereof. It is not intended that the scope of patent protection be defined by reading into any claim limitation found herein and/or issuing here from that does not explicitly appear in the claim itself.
Thus, for example, any sequence(s) and/or temporal order of steps of various processes or methods that are described herein are illustrative and not restrictive. Accordingly, it should be understood that, although steps of various processes or methods may be shown and described as being in a sequence or temporal order, the steps of any such processes or methods are not limited to being carried out in any particular sequence or order, absent an indication otherwise. Indeed, the steps in such processes or methods generally may be carried out in various different sequences and orders while still falling within the scope of the present disclosure. Accordingly, it is intended that the scope of patent protection is to be defined by the issued claim(s) rather than the description set forth herein.
Additionally, it is important to note that each term used herein refers to that which an ordinary artisan would understand such term to mean based on the contextual use of such term herein. To the extent that the meaning of a term used herein—as understood by the ordinary artisan based on the contextual use of such term—differs in any way from any particular dictionary definition of such term, it is intended that the meaning of the term as understood by the ordinary artisan should prevail.
Furthermore, it is important to note that, as used herein, “a” and “an” each generally denotes “at least one,” but does not exclude a plurality unless the contextual use dictates otherwise. When used herein to join a list of items, “or” denotes “at least one of the items,” but does not exclude a plurality of items of the list. Finally, when used herein to join a list of items, “and” denotes “all of the items of the list.” The following detailed description refers to the accompanying drawings. Wherever possible, the same reference numbers are used in the drawings and the following description to refer to the same or similar elements. While many embodiments of the disclosure may be described, modifications, adaptations, and other implementations are possible. For example, substitutions, additions, or modifications may be made to the elements illustrated in the drawings, and the methods described herein may be modified by substituting, reordering, or adding stages to the disclosed methods. Accordingly, the following detailed description does not limit the disclosure. Instead, the proper scope of the disclosure is defined by the claims found herein and/or issuing here from. The present disclosure contains headers. It should be understood that these headers are used as references and are not to be construed as limiting upon the subjected matter disclosed under the header.
The present disclosure includes many aspects and features. Moreover, while many aspects and features relate to, and are described in the context of systems, methods, apparatuses, and devices for facilitating waveform inversion using a Sifr optimizer, embodiments of the present disclosure are not limited to use only in this context.
In general, the method disclosed herein may be performed by one or more computing devices. For example, in some embodiments, the method may be performed by a server computer in communication with one or more client devices over a communication network such as, for example, the Internet. In some other embodiments, the method may be performed by one or more of at least one server computer, at least one client device, at least one network device, at least one sensor, and at least one actuator. Examples of the one or more client devices and/or the server computer may include, a desktop computer, a laptop computer, a tablet computer, a personal digital assistant, a portable electronic device, a wearable computer, a smart phone, an Internet of Things (IoT) device, a smart electrical appliance, a video game console, a rack server, a super-computer, a mainframe computer, mini-computer, micro-computer, a storage server, an application server (e.g., a mail server, a web server, a real-time communication server, an FTP server, a virtual server, a proxy server, a DNS server, etc.), a quantum computer, and so on. Further, one or more client devices and/or the server computer may be configured for executing a software application such as, for example, but not limited to, an operating system (e.g., Windows, Mac OS, Unix, Linux, Android, etc.) in order to provide a user interface (e.g., GUI, touch-screen based interface, voice based interface, gesture based interface, etc.) for use by the one or more users and/or a network interface for communicating with other devices over a communication network. Accordingly, the server computer may include a processing device configured for performing data processing tasks such as, for example, but not limited to, analyzing, identifying, determining, generating, transforming, calculating, computing, compressing, decompressing, encrypting, decrypting, scrambling, splitting, merging, interpolating, extrapolating, redacting, anonymizing, encoding and decoding. Further, the server computer may include a communication device configured for communicating with one or more external devices. The one or more external devices may include, for example, but are not limited to, a client device, a third party database, a public database, a private database, and so on. Further, the communication device may be configured for communicating with the one or more external devices over one or more communication channels. Further, the one or more communication channels may include a wireless communication channel and/or a wired communication channel. Accordingly, the communication device may be configured for performing one or more of transmitting and receiving of information in electronic form. Further, the server computer may include a storage device configured for performing data storage and/or data retrieval operations. In general, the storage device may be configured for providing reliable storage of digital information. Accordingly, in some embodiments, the storage device may be based on technologies such as, but not limited to, data compression, data backup, data redundancy, deduplication, error correction, data finger-printing, role based access control, and so on.
Further, one or more steps of the method disclosed herein may be initiated, maintained, controlled, and/or terminated based on a control input received from one or more devices operated by one or more users such as, for example, but not limited to, an end user, an admin, a service provider, a service consumer, an agent, a broker and a representative thereof. Further, the user as defined herein may refer to a human, an animal or an artificially intelligent being in any state of existence, unless stated otherwise, elsewhere in the present disclosure. Further, in some embodiments, the one or more users may be required to successfully perform authentication in order for the control input to be effective. In general, a user of the one or more users may perform authentication based on the possession of a secret human readable secret data (e.g., username, password, passphrase, PIN, secret question, secret answer, etc.) and/or possession of a machine readable secret data (e.g., encryption key, decryption key, bar codes, etc.) and/or possession of one or more embodied characteristics unique to the user (e.g., biometric variables such as, but not limited to, fingerprint, palm-print, voice characteristics, behavioral characteristics, facial features, iris pattern, heart rate variability, evoked potentials, brain waves, and so on) and/or possession of a unique device (e.g., a device with a unique physical and/or chemical and/or biological characteristic, a hardware device with a unique serial number, a network device with a unique IP/MAC address, a telephone with a unique phone number, a smartcard with an authentication token stored thereupon, etc.). Accordingly, the one or more steps of the method may include communicating (e.g., transmitting and/or receiving) with one or more sensor devices and/or one or more actuators in order to perform authentication. For example, the one or more steps may include receiving, using the communication device, the secret human readable data from an input device such as, for example, a keyboard, a keypad, a touch-screen, a microphone, a camera, and so on. Likewise, the one or more steps may include receiving, using the communication device, the one or more embodied characteristics from one or more biometric sensors.
Further, one or more steps of the method may be automatically initiated, maintained, and/or terminated based on one or more predefined conditions. In an instance, the one or more predefined conditions may be based on one or more contextual variables. In general, the one or more contextual variables may represent a condition relevant to the performance of the one or more steps of the method. The one or more contextual variables may include, for example, but are not limited to, location, time, identity of a user associated with a device (e.g., the server computer, a client device, etc.) corresponding to the performance of the one or more steps, environmental variables (e.g., temperature, humidity, pressure, wind speed, lighting, sound, etc.) associated with a device corresponding to the performance of the one or more steps, physical state and/or physiological state and/or psychological state of the user, physical state (e.g., motion, direction of motion, orientation, speed, velocity, acceleration, trajectory, etc.) of the device corresponding to the performance of the one or more steps and/or semantic content of data associated with the one or more users. Accordingly, the one or more steps may include communicating with one or more sensors and/or one or more actuators associated with the one or more contextual variables. For example, the one or more sensors may include, but are not limited to, a timing device (e.g., a real-time clock), a location sensor (e.g., a GPS receiver, a GLONASS receiver, an indoor location sensor, etc.), a biometric sensor (e.g., a fingerprint sensor), an environmental variable sensor (e.g., temperature sensor, humidity sensor, pressure sensor, etc.) and a device state sensor (e.g., a power sensor, a voltage/current sensor, a switch-state sensor, a usage sensor, etc. associated with the device corresponding to performance of the or more steps).
Further, the one or more steps of the method may be performed one or more number of times. Additionally, the one or more steps may be performed in any order other than as exemplarily disclosed herein, unless explicitly stated otherwise, elsewhere in the present disclosure. Further, two or more steps of the one or more steps may, in some embodiments, be simultaneously performed, at least in part. Further, in some embodiments, there may be one or more time gaps between performance of any two steps of the one or more steps.
Further, in some embodiments, the one or more predefined conditions may be specified by the one or more users. Accordingly, the one or more steps may include receiving, using the communication device, the one or more predefined conditions from one or more devices operated by the one or more users. Further, the one or more predefined conditions may be stored in the storage device. Alternatively, and/or additionally, in some embodiments, the one or more predefined conditions may be automatically determined, using the processing device, based on historical data corresponding to performance of the one or more steps. For example, the historical data may be collected, using the storage device, from a plurality of instances of performance of the method. Such historical data may include performance actions (e.g., initiating, maintaining, interrupting, terminating, etc.) of the one or more steps and/or the one or more contextual variables associated therewith. Further, machine learning may be performed on the historical data in order to determine the one or more predefined conditions. For instance, machine learning on the historical data may determine a correlation between one or more contextual variables and performance of the one or more steps of the method. Accordingly, the one or more predefined conditions may be generated, using the processing device, based on the correlation.
Further, one or more steps of the method may be performed at one or more spatial locations. For instance, the method may be performed by a plurality of devices interconnected through a communication network. Accordingly, in an example, one or more steps of the method may be performed by a server computer. Similarly, one or more steps of the method may be performed by a client computer. Likewise, one or more steps of the method may be performed by an intermediate entity such as, for example, a proxy server. For instance, one or more steps of the method may be performed in a distributed fashion across the plurality of devices in order to meet one or more objectives. For example, one objective may be to provide load balancing between two or more devices. Another objective may be to restrict a location of one or more of an input data, an output data, and any intermediate data therebetween corresponding to one or more steps of the method. For example, in a client-server environment, sensitive data corresponding to a user may not be allowed to be transmitted to the server computer. Accordingly, one or more steps of the method operating on the sensitive data and/or a derivative thereof may be performed at the client device.
The present disclosure describes systems, methods, apparatuses, and devices for facilitating wave-based inversion using Time-Domain Sifrian Inversion (TDSI).
Further, the present disclosure describes the Time-Domain Sifrian Inversion for Optimized Second-Order Imaging and Model Building with Cycle-Skipping Mitigation.
Further, the Time-Domain Sifrian Inversion (TDSI) extends the foundational principles of Sifrian inversion, traditionally applied in the frequency domain, to the time domain. The time domain offers several advantages, including enhanced temporal resolution and the capability for parallel processing. The time domain is particularly well suited for addressing challenges like cycle skipping, as it allows access to the signal envelope and enables the use of sophisticated distance metrics derived from Optimal Transport theory, such as Wasserstein distances. The Time-Domain Sifrian Inversion (TDSI) represents a groundbreaking adaptation of the Sifrian inversion to the time domain. It not only provides a time-domain formulation of the Sifrian inversion but also addresses the cycle-skipping issue through the use of suitable cost functions. The focus of TDSI is on enhancing imaging clarity, precision, and computational efficiency.
Introduced by Mehouachi et al. in 2023, the Sifrian inversion offers an innovative approach to second-order optimization in waveform imaging. It employs a unique mathematical construct known as the Sifrian functional, which intriguingly remains at zero value. The name “Sifrian” is derived from the Arabic word “Sifr,” which means zero. This zero-valued functional allows for the efficient incorporation of second-order derivative information into the optimization process, effectively bypassing the computational burden associated with the Hessian matrix.
Further, the extension of the foundational principles of Sifrian inversion allows TDSI to be compatible with existing cost functions that are robust to cycle skipping, a frequent challenge in waveform inversion when low frequencies are absent. TDSI aims to enhance both temporal resolution and imaging accuracy. In terms of computational efficiency, TDSI has the potential to reduce the parameter search space, with the number of time steps serving as an upper limit. By balancing mathematical rigor with computational practicality, TDSI is well suited for a variety of applications, including but not limited to geophysics, seismology, and medical imaging. TDSI aims to establish new benchmarks that evaluate the robustness of imaging techniques in the presence of cycle skipping.
Further, the present disclosure describing the Time-Domain Sifrian Inversion pertains to the domain of Partial Differential Equation (PDE)-constrained optimization, with a special focus on wave phenomena optimization. The Time-Domain Sifrian Inversion (TDSI) is a computational approach that extends the foundational principles of the Sifrian inversion applied exclusively in the frequency domain. TDSI represents a significant advancement in the field, offering an accurate and efficient second-order optimization method that is robust to cycle skipping which is a challenge often encountered in waveform inversion, especially when low frequencies are missing. This novel approach is particularly well suited for applications in geophysics, medical imaging, and other scientific disciplines requiring high-quality imaging. By leveraging the Sifrian functional which is a Lagrangian-like function successful in the frequency domain, the TDSI introduces its time-domain counterparts to address the limitations of cycle-skipping resolution more effectively. In doing so, TDSI establishes new benchmarks in imaging techniques, harmonizing mathematical rigor with computational practicality.
Further, the present disclosure describes a method for Time-Domain Waveform Imaging. Further, the method for the Time-Domain Waveform Imaging is a method for waveform imaging in the time domain. Further, the method may be associated with a computing system that receives observed data from a surveyed medium via one or preferably multiple sensors. This system then generates a model or an image of the surveyed medium. The generation process involves obtaining an initial model, producing synthetic data using this model, and comparing the synthetic data to the observed data. If the observed and synthetic data do not correspond within a predefined matching threshold, the model undergoes iterative updates using the Time-Domain Sifrian Inversion (TDSI) until a match is achieved. This iterative updating is derived from the augmentation of a wavefield equation, an adjoint wavefield, and a gradient definition of the model by introducing specific augmentation variables. Notably, a cost function, which quantifies discrepancies between the synthetic and observed data, is strictly excluded, during the process. The TDSI assembles the augmented components to create a unique functional in the time-domain, followed by a variational perturbation to characterize the update. The formulated updated is better expressed in a sequence of tensors multiplication, addition, and inversion. The final model or image of the surveyed medium is then generated based on the iterative outputs from the TDSI method.
Further, the present disclosure describes a device for Time-Domain Waveform Imaging. Further, the device for the Time-Domain Waveform Imaging is a device designed for waveform imaging and model construction in the time domain. This device comprises an interface and a data processing unit. The interface accommodates both an initial model and observed data from a surveyed medium. The data processing unit, connected to the interface, is designed to produce synthetic data, compare it with observed data, and determine if they match within a set threshold. If not, the unit employs the TDSI to iteratively refine the initial model until a match is achieved. This iterative process leverages the unique properties of the TDSI to ensure accurate and efficient model updates. The data processing unit then outputs the refined model or image of the surveyed medium.
Further, the present disclosure describes a Time-Domain Waveform Imaging Procedure. Further, the Time-Domain Waveform Imaging Procedure is a comprehensive method for waveform imaging and model development in the time domain. This method starts by collecting observed data from a surveyed medium using multiple sensors. An initial model is selected, either from expert guesses, previous modeling, or other methods like tomography that provide long-wavelength information. This initial model is injected into the training loop, and synthetic data derived from this model is produced. The core of the method involves the iterative refinement of the model using TDSI until the synthetic data aligns with the observed data within a set threshold. One of the key advantages of TDSI over its frequency-domain counterpart is its compatibility with cost functions that are robust to cycle skipping. TDSI's unique approach to augmenting and assembling various components of the model ensures efficient leverage of the curvature in the time domain. The method is particularly effective for high-resolution imaging and offers fast convergence, potentially aiding in the identification of objects or areas suitable for further exploration or investigation.
Further, the present disclosure describes the mathematical and computational aspects integral to this disclosure. Further, the present disclosure describes the state-of-the-art methods for gradient computation in waveform inversion specifically an adjoint method which is a technique that is pivotal for efficient optimization and is analogous to backpropagation in machine learning. Further, the present disclosure describes the novelty and utility of the Time-Domain Sifrian Inversion (TDSI) which is an innovative approach to second-order optimization in waveform imaging and model building in the time domain.
Further, the present disclosure describes the Differentials Convention used in this disclosure. Further, the tensor convention is adopted for partial derivatives and represented as
Here, J is the function undergoing differentiation, while m denotes the variable with respect to which differentiation occurs. If J is scalar, the resultant
will have the same dimensions of m, be it scalar, vector, or a higher-order tensor. If J is tensorial, the resultant tensor
will have a rank equal to the sum of the tensor ranks of both J and m, with m contributing its dimensions first. This notation ensures consistency and clarity across all mathematical formulations.
Further, the present disclosure describes Inner Product in Time-Space Functions. In the mathematical framework of waveform inversion, the inner product plays a pivotal role. To better illustrate the ideas, the following canonical inner product is considered as an example: for two finite-energy 2 functions ƒ(t,x) and g(t,x) defined over a spatial domain Ω and time interval [0, T], the
2 inner product is given by:
This inner product captures the “similarity” between two functions across both space and time, providing a mechanism to quantify their alignment or misalignment. The functional space could be generalized with various types of functions and inner products. The inner product could be defined only on the space domain, without time integration, in such case it is denoted as Ω.
Further, the present disclosure describes The Concept of Adjoint Operators. At the heart of many formulations in waveform inversion is the concept of the adjoint operator. Given an operator L, its adjoint L* is defined such that for any two suitable functions ƒ and g:
In essence, the adjoint of an operator captures the “dual” action of the original operator, facilitating the transfer of operations from one function to another within the inner product.
Further, the present disclosure describes Gradient Computation Using the Adjoint Method. The gradient of the cost function, pivotal for optimization, usually requires the computation of a Jacobian matrix which can be computationally expensive. Fortunately, there is an efficient way (adjoint method) to compute the gradient which bypasses the computation of the Jacobian and is derived using both the forward and adjoint equations. Here is a structured overview of the gradient computation process:
1. Forward Modeling (in Time): Given a model parameter set m and a source term q(t,x), the forward problem computes the state u(t,x), over a specified time interval [0, T]. Mathematically, this is represented as:
2. Misfit Quantification: The misfit between the predicted state u(t,x) and the observed data d(t,x) is encapsulated in the objective function J(m). A prevalent choice for this function is the least-squares formulation:
3. Lagrangian: To compute the gradient ∇mJ, an adjoint state λ=λ(t,x) is introduced within the following Lagrangian:
4. Lagrangian Differentiation and Gradient Computation: Differentiation of the Lagrangian with respect to the model parameters m yields:
The adjoint λ(t,x) need to be picked such that
in order to avoid computing the Jacobian
These considerations yield the following adjoint equation and gradient expression:
The gradient of the objective function, crucial for optimization, is determined using both the forward state u(t,x) and the adjoint state λ(t,x). This gradient steers optimization algorithms, facilitating iterative model refinements.
Further, the present disclosure describes Spatial Adjoint Operators. To utilize adjoint operators effectively in inversion tasks, it is crucial to understand their properties, particularly in the context of the inner product, which serves as a cornerstone for these operations. In wave propagation scenarios, the Sommerfeld radiation condition is often invoked. This condition stipulates that radiated wavefields diminish to zero at great distances (ideally, at infinity) from the source. In the realm of wave equations, the primary spatial operators of interest are the nabla operator (∇) and the Laplacian (Δ). When the Sommerfeld radiation condition is applied, integration by parts reveals the following adjoint relationships for these operators:
Further, the present disclosure describes Time Adjoint Operators. For wave equations, the time operators are essentially first and second order derivatives. Similar to the space operator, applying integration by part in the inner product is straightforward, however, it creates boundary values at t=0 and t=T. Given a time axis, and a force/excitation/source applied at time t=0, the nullity of the forward wavefields and their derivatives are assumed at t=0, i.e., u(t=0, x)=0, ∂tu(t=0, x)=0. For any adjoint wavefield and to express easily the adjoints it needs to be explicitly assumed that at t=T, λ(t=T, x)=0, ∂tλ(t=T, x)=0. With this assumption, the following adjoints are obtained:
Having a null boundary condition at t=T, induces time reversal as the resolution of any partial differential equation for an adjoint wavefield is performed with a change of variable τ=T−t.
Further, the present disclosure describes an Example Wave Equation which is an illustrative example of the following basic wave equation:
The adjoint wavefield verifies the same equation but with boundary conditions on the final time:
The gradient is given by:
Further, the present disclosure describes Efficiency and Practical Considerations. By introducing the concept of the inner product and defining the adjoint within the framework of the adjoint method, the underpinnings of the gradient computation process are better understood. The inner product provides a mechanism to quantify the “similarity” between functions, while the adjoint facilitates efficient gradient computation.
The appeal of the gradient-based optimization approach lies in its computational efficiency. It computes the gradient at a cost comparable to a single forward and backward propagation, irrespective of model parameter count. This scalability renders it apt for large-scale inverse problems. However, the direct application of this approach to second-order optimization presents challenges. These challenges have been addressed in the frequency domain through the Sifrian inversion method. Further, the present disclosure describes the development of the Sifrian inversion to the time domain and offers enhanced compatibility with methods for mitigating cycle-skipping issues.
Further, the present disclosure describes encompasses systems, methods, and computer-readable media related to Time-Domain Sifrian Inversion (TDSI) for wave-based imaging, inversion, and optimization. TDSI adapts the Sifrian functional which is originally conceived for frequency domain applications, to address challenges such as cycle skipping in the time domain. This adaptation allows for the use of time-oriented cost functions that are robust to the challenges of cycle skipping. The method uniquely characterizes the Newton direction for optimization without the computational burden of directly involving the Hessian matrix. This innovation enables more efficient and robust model updates, addressing the slow convergence often associated with traditional optimization techniques.
Further, the present disclosure describes a method for wave-based inversion in the time domain. This method involves collecting wave-based data like seismic or electromagnetic signals, initializing model parameters (e.g., velocity or resistivity models), and iteratively refining these models. The TDSI framework guides this iterative process, incorporating robust cost functions that are resilient to cycle skipping and leveraging second-order updates for enhanced efficiency and robustness.
Further, the present disclosure describes a computing system designed for wave-based inversion. This system receives wave-based data and initiates the inversion process, which includes parameter initialization to prevent null updates. Utilizing the Sifrian functional, the system effectively characterizes the second-order Newton direction, thereby optimizing the inversion process. While the use of a cycle-skipping robust cost function is optional, it is highly recommended. The TDSI framework offers a streamlined and efficient solution that capitalizes on second-order curvature information and provides a time-based remedy for cycle skipping. The system iteratively refines the model until the predicted data aligns with the desired output within a predetermined tolerance level.
Further, the present disclosure describes a wave-based data inversion apparatus consisting of an interface and a data processing unit. The interface is designed to receive, store, and manage datasets essential for the inversion process. The data processing unit, which is interconnected with the interface, supervises the model updating routine and handles necessary synchronization. It generates predictive data by propagating an input sample through the inversion model and employs a predetermined loss function that is robust to cycle skipping to evaluate the difference between the predicted and desired outputs. If this difference exceeds a predefined threshold, the data processing unit triggers further model updates. Utilizing the TDSI framework, the unit iteratively refines the model until the predicted data closely matches the desired output. The data processing unit is also equipped with a specialized solver for executing TDSI updates, thereby optimizing the model to minimize discrepancies between predicted and actual data.
Further, the Sifrian functional which is designed for frequency-domain applications, has been adapted to the time domain. The frequency-domain version operates on individual frequencies, limiting its efficacy in managing cycle skipping. To mitigate cycle skipping, techniques like signal envelopes or Wasserstein distances often require a time-domain formulation. To this end, the time-domain Sifrian functional is introduced as follows:
This functional serves as a Lagrangian-like entity, becoming null when the system reaches equilibrium. It integrates both forward and backward equations in the time domain, as well as the gradient definition. Unlike traditional Lagrangian, the Sifrian functional and its derivatives become null when equilibrium conditions for forward, backward, and gradient definitions are met. The term “Sifrian” originates from the Arabic word for zero, “Sifr.”
The time-domain Sifrian introduces three new adjoint variables, denoted as (γ, ζ, N). These variables are pivotal for characterizing the Newton direction. Specifically, (γ) and (ζ) are second-order multipliers associated with the forward and backward wavefield equations, while (N) is linked to the gradient definition. Further, the present disclosure provides a demonstration that (N) serves as the Newton direction, and (ζ) and (γ) are perturbations in the forward and backward wavefields induced by the model perturbation through the Newton update (N).
Upon making the choices:
It has been found that
which corresponds to the Newton direction.
Differentiating the Sifrian functional in the time domain yields the following system of equations:
Further, differentiating the time-domain Sifrian functional yields a system of equations that is inherently non-convex and challenging to solve. While Gauss-Newton truncation is one approach, the analysis indicates that the Newton method, derived from a Taylor expansion of the gradient, lacks proper truncation. To address this, additional quadratic terms are introduced for formulating an enhanced Sifrian system of equations which are better suited for resolution.
Further differentiation reveals that certain terms are absent from the original formulation. These missing terms are essential for achieving a more accurate and efficient inversion process. The Newton method aims to find an update that nullifies the gradient, as expressed below:
Upon analyzing the Sifrian equations, it has been found that the new adjoint variables, ζ and γ are essentially perturbations in the wavefields introduced by the Newton update N.
Taking these aspects into account, the full Sifrian system, now incorporating the quadratic terms, is given by:
The enhanced Sifrian system is convex quadratic and better suited for resolution. The Gauss-Newton system is the linear approximation of the full Sifrian system, when (λ+γ) is assumed to be small. Instead of linearizing (Gauss-Newton), solving the full Sifrian system is opted for.
One approach to ensure that the perturbed gradient is null involves selecting the Newton update such that (λ+γ) becomes null. Nullifying (u+ζ), implies the absence of a forward wavefield which is not desirable. However, nullifying the adjoint is suitable, as it indicates that the mismatches have been eliminated—ultimately the objective of the entire optimization process.
With the perturbed adjoint nullified, two equations remain to be solved:
The first equation presents a challenge: the derivatives of the cost function involve the receiver's sampling operator, localizing the information about ζ at the sensors. This is not ideal for gaining insights into the entire subsurface. However, these equations are computationally efficient and can be interpreted as identifying a suitable perturbation that leads to a second-order update at the receivers. The reduced Sifrian equations are notably streamlined.
To solve these equations, the properties of the adjoint are employed once more. The Lagrange multiplier λ(t,x) and a Dirac function δt are introduced to achieve time localization. Formally:
is introduced which serves as an instantaneous gradient.
By leveraging both the gradient and the Dirac function, and using also the adjoint equation, the following equation is achieved:
The differential
is localized so it could be inverted for ζ inside the space-only inner product Ω. This leads to the Sifr equation:
The Sifr equation is in essence a regularization of the Sifrian system combined with a dimensionality reduction which aims essentially to provide the scaling expected from second-order optimization method.
Solving the Sifr equation can be achieved through discretization in both time and space, thereby converting the system into a rectangular one.
Further, the resolution of the Sifr equation can be approached in multiple ways, but the underlying principle involves discretization in both time and space. Any forward solver that provides the losses and the instantaneous gradient can be employed. The system can be expressed in matrix format as MGN=l, where MG is a matrix with its first dimension corresponding to nt the number of time steps, and its second dimension reflecting the size of the model nx. At this juncture, it is worth noting that it is advantageous to consider the Gram matrix MGTMG. This significantly reduces the size of the problem, as nt is typically much smaller than nx. and the inversion would involve a nt×nt matrix. The inversion of this matrix yields the Sifr update.
For extremely large-scale problems where even the computation of the Gram matrix is computationally prohibitive, an ad-hoc conjugate gradient search can be implemented. In this scenario, N can be expressed as N=MGTv, where v is a vector of size nt×1. Damping is an essential component in the inversion process, as is well-known in practice. If the conjugate gradient method is used, the computation of N=MGTv can be efficiently performed by modifying the adjoint source terms
through multiplication along the time axis with the coefficient v.
One of the most important advantages of Time-Domain Sifrian Inversion (TDSI) lies in its robustness against cycle skipping, especially when dealing with absent low frequencies. Unlike its frequency-domain counterpart, TDSI provides a flexible and adaptive framework that accommodates a variety of cycle-skipping-resistant cost functions. These cost functions can be rooted in optimal transport theory, envelope measures, or convolutional measures, thereby offering practitioners a comprehensive toolkit for tackling cycle skipping across diverse applications. This adaptability sets TDSI apart from the original frequency-domain Sifrian method. Furthermore, TDSI can serve as a preliminary step in frequency-domain implementations, effectively mitigating challenges associated with missing frequencies. This dual applicability underscores the innovative nature of TDSI.
In the context of TDSI, it is crucial to note that the second derivative of the cost function with respect to u, denoted as
is required. Therefore, as long as the norm or
distance measure used in the cost function can provide this second derivative, it aligns well with the TDSI framework. This opens the door for incorporating a wide array of distance measures that are robust to cycle skipping, further enhancing the practicality and efficacy of the TDSI approach.
Further, the present disclosure describes a system architecture explicitly designed for implementing the Time-Domain Sifrian Inversion (TDSI). This system comprises a data acquisition unit, a data storage unit, a computational unit, and a user interface. The data acquisition unit is configured to collect oscillatory data from various sensors and instruments, which is then stored in the data storage unit. The computational unit, equipped with high-performance processors and/or GPUs and/or TPU and similar processing specialized devices, is the core of the TDSI implementation. It is programmed to execute the TDSI algorithm, performing iterative model updates based on the Sifrian functional and its resultant Sifr equation, preferably with a one or a combination of robust or non-robust cost function. The computational unit interacts with the data storage unit to fetch the necessary data and to store the inversion results. The user interface optionally allows operators to input initial models, set additional constraints, set optimization parameters, and visualize the inversion results. This integrated system architecture ensures seamless data flow and efficient computation, making TDSI a practical and robust solution for wave-based inversion challenges. The system can be deployed as a standalone unit or integrated into existing data processing pipelines, offering flexibility and scalability to meet various operational needs.
Further, the present disclosure describes systems, methods, apparatuses, and devices for facilitating waveform inversion using a Sifr optimizer. Further, the present disclosure serves as a comprehensive guide to the mathematical and computational principles essential for waveform inversion techniques. Central to this exploration is the adjoint-state method, a concept analogous to backpropagation in machine learning, which is pivotal for achieving efficient optimization. The present disclosure introduces the necessary elements to carry on waveform inversion in general in both the time and space domains before moving to the Sifr optimizer.
Further, the present disclosure describes forward modeling. Given a model parameter m and a source term in time q(t,x) or frequency q(ω, x), the forward problem computes the state u(t,x) or u(ω, x) over a specified time interval [0, T] or for specific frequencies {ω0, ω1, ω2 . . . }. Mathematically, this is represented as:
Further, the forward operator is linear in u and is represented by an operator L(m) to emphasize the linearity. The transition between the time and frequency domain could be achieved through a Fourier transform, which the present disclosure omits here for the sake of clarity. The above equations although continuous in time/frequency and space, could also be discretized using various discretization schemes (finite difference, finite element . . . ) and would yield a similar matrix form for the considered quantities.
Further, the present disclosure describes the Functional Inner Product. Without loss of generalizations, all optimizations are occurring in a Banach space. For the sake of simplicity, the inner products are considered to be either in the time or frequency domain. For two finite-energy 2 functions ƒ(t,x) and g(t,x) defined over a spatial domain Ω and time interval [0, T], the
2 inner product is given in the time domain by:
If the functions ƒ and g are defined in the frequency domain, the canonical inner product involves the complex conjugation:
The inner product could also be defined only in the space domain with a single time or frequency value, i.e., without time or frequency integration, in such case it is denoted as Ω.
The inner product is helpful for quantifying mismatches between the desired and actual outputs of the synthetic modeling and is used to create cost/loss functions.
Beyond time and frequency, the Sifr optimizer is applicable to any domain representation as long as the loss/cost function involves an integral/summation over various (at least one) samples of the domain representation, this concept will be elucidated later by considering a unified representation of time/frequency with a unified “sample” denoted p.
Further, the present disclosure describes Misfit Quantification. In practical applications, the locations of sensors are specific; hence, the present disclosure describes a spatial sampling operator R(x)to indicate the location of the sensors in the space Ω. Such an operator is needed because data is collected using a finite number of sensors, and it is essentially a sum of Dirac functions, which, when discretized, would correspond to a matrix of ones and zeros (one indicates the presence of a sensor). The sampling operator verifies the basic but essential property R(x)*R(x)=R(x). If the sampling operator is discretized into a matrix, the product must be taken in the Hadamard sense (i.e., term by term).
Further, the misfit between the predicted state u and the observed data d is encapsulated in the objective function J(m). An illustrative choice for this function is the least-squares formulation:
In the frequency domain, the time variable is changed to frequencies.
Be it in the time or frequency domain, the loss/cost function is expressed as a sum for various time steps or frequencies or even one single frequency (Dirac measure). As such the following per-time or per-frequency is introduced, such that:
Or in the frequency domain:
To further generalize time and frequency an abstract “samples” space that covers both time and frequency is considered and it would be referred to as p∈D such that:
In essence, p could be time or frequency and D is a set of time or frequency values. Before moving to gradient computations.
Further, the present disclosure describes differential notations. Further, the present disclosure describes the utilization of the tensor convention for differentiation, for illustration the dimension of
is elucidated. J is the function undergoing differentiation, while m denotes the variable with respect to which differentiation occurs. If J is scalar, the resultant
will have the same dimensions as m. If J is tensorial, the resultant tensor
will have a rank equal to the sum of the tensor ranks of both J and m, with m contributing its dimensions first.
Further, the present disclosure describes a concept of adjoint operators. At the heart of many formulations in waveform inversion is the concept of the adjoint operator. Given an operator L, its adjoint L* is defined such that for any two suitable functions ƒ and g:
In essence, the adjoint of an operator captures the “dual” action of the original operator, facilitating the transfer of operations from one function to another within the inner product.
Further, the present disclosure describes Adjoint Operators. In scenarios involving the propagation of waves, the Sommerfeld radiation condition is often used. This condition states that wavefields radiated from a source must decrease to zero at great distances, ideally at infinity. In wave equations, the nabla (∇) and Laplacian (Δ) operators are of primary importance. When the Sommerfeld radiation condition is applied, integration by parts reveals the following adjoint relationships for these operators:
Further, the time operators are essentially first and second-order derivatives. Like the space operator, applying integration by part in the inner product is straightforward. However, it creates boundary values at t=0 and t=T. Given a time axis and a force/excitation/source applied at time t=0, the nullity of the forward wavefields and its derivatives at t=0, i.e., u(t=0, x)=0, ∂tu(t=0, x)=0 is assumed. For any adjoint wavefield λ and to express the adjoints easily, it is explicitly assumed that at t=T, λ(t=T, x)=0, ∂tλ(t=T, x)=0.
With this assumption, the following adjoints are:
Having a null boundary condition at t=T, induces time reversal as the resolution of any partial differential equation for an adjoint wavefield is performed with a change of variable τ=T−t.
Furthermore, in the frequency domain, the time derivative becomes a multiplication:
The adjoint is simply given by the complex conjugate.
Lagrangian: To compute the gradient ∇mJ, an adjoint state λ=λ(t,x) or λ=λ(ω, x) is introduced within the following Lagrangian:
The adjoint λ can have any value, and one that simplifies the gradient computation is selected.
Lagrangian Differentiation and Gradient Computation: Differentiation of the Lagrangian with respect to the model parameters m yields:
If the wavefield u verifies the forward equation, then the first part of the Lagrangian is null and
The adjoint λ need to be picked such that
to avoid computing the Jacobian
These considerations yield the following adjoint equation and gradient expression:
In the time formulation, the following is obtained:
In the frequency formulation, complex differentiation in the Writinger sense should be used but leads essentially to the same results, the following is obtained:
The form of the gradient depends on the selected L operator, for the following basic wave equation in the time domain:
The adjoint wavefield verifies the same equation but with boundary conditions on the final time:
The gradient is given by:
Working in the frequency domain, the following basic wave equation yields:
The adjoint verifies:
Leading to the following gradient:
Further, the present disclosure describes the Sifrian. Further, the Sifrian functional was originally created for machine learning to characterize the Newton direction. It was then adapted for use in waveform inversion, first in the frequency domain and later in the time domain, as described in the present disclosure. The main concept is to build a second-order Lagrangian using forward, backward, and gradient definitions, and then apply different types of differentiation. In the time domain, the Sifrian algorithm is implemented as follows:
Further, this function acts in a similar way to the Lagrangian, becoming zero when the system reaches equilibrium. It combines both forward and backward equations in the time domain, as well as the gradient definition. Unlike the traditional Lagrangian, the Sifrian function and its derivatives become zero when the equilibrium conditions for forward, backward, and gradient definitions are fulfilled. The term “Sifrian” comes from the Arabic word for zero, “Sifr.
Further, three new adjoint variables are introduced, denoted as (γ, ζ, N). Technically, the adjoint variables have specific purposes: (N) serves as the Newton direction, and (ζ) and (γ) are perturbations in the forward and backward wavefields induced by the model perturbation through the Newton update (N). Differentiating with respect to the model parameter m, yields the following characteristic equation:
This leads to the following classic choices:
Further, finding
Which corresponds to the Newton direction. Expressing the above differentials yields the following equations that need resolution to get the second-order update:
Further, the above system is inherently non-convex and challenging to solve. While Gauss-Newton truncation is one approach, our analysis indicates that the Newton method, derived from a Taylor expansion of the gradient, lacks proper truncation. To address this, additional quadratic terms are introduced, formulating an enhanced Sifrian system of equations better suited for resolution.
Further, in the frequency domain, changing the inner product and the definition of the gradient provides nearly the same equations which presents the same type of difficulty to solve. The Sifrian in the frequency domain is:
Which eventually leads to a similar system to solve in the frequency domain:
The above systems are inherently complex to solve because of fundamental limitations related to the way they are crafted. Further, the present disclosure describes an interpretation that leads to the iteration of the Sifr optimizer.
Further, the development of the Sifr optimizer as disclosed in the present disclosure represents a significant leap in waveform inversion. This optimizer emerges from a comprehensive evaluation of the traditional Newton method via the novel Sifrian functional approach which led to the development of the TDSI (Time domain Sifrain Inversion). Further, the present disclosure illustrates the critical limitations intrinsic to Newton's method, particularly its initial truncation of higher-order terms, which complicates the resolution of the Sifrian equations. Further, the present disclosure describes reintroduction these higher-order terms and deviates from the Newton optimization framework to establish a novel paradigm. This paradigm pivots on the efficient resolution of the Sifr equations and the practical implementation of the Sifr optimizer, emphasizing a second-order update primarily at the receiver's locations. The present disclosure describes circumventing a dimensionality reduction.
Further, the present disclosure describes not generating a second-order update at the output layer output layer. Instead, the present disclosure describes a reinterpretation of the utility of the original Sifr equation toward directly minimizing the loss function. The term ‘Sifr’, meaning zero in Arabic, symbolizes the goal of nullifying the loss/cost function, which encapsulates the essence of this optimizer. The revised Sifr optimizer which is a novel variant of the Sifr optimizer, abandons traditional solutions to find the Newton direction in favor of a robust second-order method that not only facilitates large step optimizations but is also substantiated with convergence proofs in this document. Further, the novel variant of the Sifr optimizer requires a least squares resolution which could be carried in any way deemed suitable by the practitioner without loss of generalization.
Further, the present disclosure further clarifies, that the original goal of the Newton method is to find an update that nullifies the gradient, as expressed below:
Upon analyzing the Sifrian equations, the new adjoint variables, ζ and γ are found which are essentially perturbations in the wavefields introduced by the Newton update N.
In the time domain:
Or in the frequency domain as:
Taking these aspects into account, the full Sifrian system, now incorporating the quadratic terms, is given by:
In the frequency domain, the system is:
Further, the enhanced Sifrian system is convex quadratic and better suited for resolution. The Gauss-Newton system is the linear approximation of the full Sifrian system, when(λ+γ) is assumed to be small. Instead of linearizing (Gauss-Newton), the present disclosure describes solving of the full Sifrian system.
Further, the present disclosure describes the Sifr equation. Ideally, one approach to ensure that the perturbed gradient is null involves selecting the Newton update such that (λ+γ) becomes null. Nullifying (u+ζ), implies the absence of a forward wavefield which is not desirable. However, nullifying the adjoint is overall suitable, as it indicates that the mismatches have been eliminated—ultimately the objective of the entire optimization process.
Further, in practice exactly canceling the gradient is not realistic, since the loss landscape is more complex than the quadratic approximation used for second-order methods and hence a more achievable goal is to get the gradient as small as possible. Cauchy Schwarz inequality provides another perspective on the cancellation of the adjoint inducing a null gradient:
Overall steering the optimization such that:
However, another “statistical way” to allow the focus on canceling (λ+γ) or at least its average if statistical independence is provided. The time or frequency integral could be seen as an average operator.
Further, the above property is statistical independence and requires that (λ+γ)=∈ is independent of
The Sifr equation was introduced initially in the time domain and the time integration was considered as averaging (expectation operator), the model was updated such that (λ+γ)=E is a small residual of a regression operation.
Further, within this approach there are two equations to be solved in the least square sense:
In the frequency domain, it is:
Further, at this stage, it is important to remember that {u, ζ} are time or frequency indexed while {J, m, N} are not, using p to denote either the time or frequency the following is obtained:
The Sifr system could be further expressed with the samples p as:
Further, for the sake of simplicity, εp˜(0, σ) is assumed, even though Gaussianity is optional, and only the null average is needed. To avoid any further mix-up the sought-after model update is changed from N to δm. For further clarity, the real case is considered so the adjoint operation would change to transposition, and the inner product, wrutinger derivative should be used if considering the complex case.
Further, the first equation presents a challenge: the derivatives of the cost function involve the receiver's sampling operator, localizing the information about ζ at the sensors. This is not ideal for gaining insights into the entire propagation medium. However, these equations are computationally efficient and can be interpreted as identifying a suitable perturbation that leads to a second-order update at the receivers.
Furthermore, be it in the frequency or time domain and independently of the format the wave equation takes, the following characterization of the adjoint variable ζ is obtained:
Injecting in the first equation yields the Sifr equation:
Further, the above equation indicates that the Sifr update N is the least squares estimate that creates a second-order update at the output level with some random noise εp whose mean should be zero p˜D[εp]=∫p∈D εpdp=0, for an unbiased estimation. Further hypotheses from the regression field such as the Gaussian distribution of the noise could be used or added for further analysis and properties.
Further, the dimensionality reduction that was described earlier, is used to make the resolution manageable. Essentially multiplying by
on the left side under the further independence hypothesis; it was essential for computational tractability.
Further, the present disclosure describes removing this dimensionality reduction without compromising computational efficiency. Such a feat is explained below:
The main intuition is that Newton's method has inherently several limitations. The version of the Sifr equation that is seeking an update N that creates as best a Newton update “only” at the output layer; such an endeavor is limited in a sense because the original objective of the learning process is to nullify the loss function, ideally overall available samples. In a statistical sense, the leftover mismatches should be small with a null mean and with a small variance. The original Newton method seeks to nullify the gradient because that corresponds to critical points which could be the sought-after minima, but also maxima or saddle point, hence the importance of convexity assumption. In waveform inversion, convexity is not guaranteed and practically unrealistic to assume, which is an existential dilemma for second-order optimization.
The present disclosure describes recreation of the Newton method dynamics not to cancel the gradient but to zero out the loss function directly. The main starting point is to notice that the Sifr equation could “almost” be factorized with respect to a derivation:
Such a relation assumes that the derivative
with respect to up is null, which could be reasonable since
is the Jacobian which describes the sensitivity of the forward wavefield to the model's parameters and often with linear operators, it does not depend on the output itself. δm having a null derivative with respect to up is assumed which is not necessarily true, but for the sake of this intuitive explanation it being assumed. Most importantly, the re-apparition of the gradient
is noticed within a very familiar first-order Taylor expansion. The intuition behind this disclosure is instead of using dimensionality reduction and regression, the following regression is directly solved:
Or in a slightly compact notations, a novel variant of the Sifr equation is:
εp is a scalar, assuming that the samples p∈D are i.i.d. (independent identically distributed), the following least squares estimator, the integral is obtained:
This new variant of the Sifr optimizer drops the search for a null gradient and focuses on nullifying the cost function directly (hence the Sifr name) in a quadratic manner. The intuitive reason is that the loss function is positive and once it reaches a null (or quasi-null) value, the gradient (first derivative) should also be null, otherwise, it would be possible to find a small perturbation that breaks the positivity of the loss/cost function. Even in the case of quasi-null loss, the gradient should be also quasi-null, again to avoid breaking the positivity of the loss.
The above formula bears some resemblance with the Natural Gradient method, especially with the covariance matrix. Such a term corresponds to per-sample gradients weighted with their contribution to the loss/cost function. Also, the practitioners would immediately notice that using the Gram matrix of the per-sample gradients to carry on the regression would be more efficient than using the covariance matrix in this case.
At this stage it is important to recognize that the solving for the Newton update is not performed. “The Sifr method”, is a complete novelty, the Sifr name (zero in Arabic) is well deserved because the present disclosure describes zeroing out the loss function directly since with the least squares estimator the following is obtained:
Therefore:
Further, the present disclosure describes Convergence. The new variant of the Sifr optimizer is further referred to as the Sifr optimizer directly. The update created using the Sifr optimizer is novel. Further, a convergence needs to be established for the update. Assuming a sequence of updates 0,1,2 . . . n and the associated Sifr updates δm0, δm1, . . . , δmn. Further, a sequence of weights m0, m1, . . . , mn may also be used. Further, assuming C2 continuity of the loss, at the nth step the following Taylor-Lagrange development is obtained:
Since δmn solves the least squares problems, the following is obtained:
The Gaussian assumption is not necessary (and can be relaxed) but it is introduced to better clarify that the residuals have a zero mean and a certain variance. At this stage, the following is obtained:
Hence:
The resolution of the optimal δmn+1 yields
Denoting the Fisher matrix as Fn+1=[∫∇θp(mn+1)∇θ
p(mn+1)Tdp] and assuming the residuals from the previous step to be independent of the current gradients ∇m
p(mn+1), the following is obtained:
Hence:
Assuming classic uniform boundness hypotheses on the Inverse of the Fisher Matrix and the gradients of the loss, then ∃M>0:
Therefore the model m0, m1, . . . , mn parameters are a special case of Cauchy sequences. By assuming ∃k ∈: such that M∥δmk∥<1, then in any Banach space there is guaranteed convergence of the model parameter sequence.
Further, the present disclosure describes the convergence of the Sifr Optimizer to a Minimum. For establishing the convergence of the Sifr optimizer to a minimum, convergence and hypotheses from the previous section are assumed, hence
which by definition of the Sifr equation yields J(m*)=∫p(m*)dp=0. Around the convergence point, the following is obtained:
Quadratic Convergence Rate of the Sifr Optimizer: Using the following notation
Applying the gradient and taking the integral yields:
Therefore under some extra classic assumption about the Fisher matrix and the gradients norms a quadratic rate of convergence is obtained:
This means that the Sifr optimization would converge quadratically towards a minimum.
Further, the present disclosure describes the important role of damping in the Sifr optimizer. Further, the Sifr optimizer performs an inversion of the Fisher Matrix or the corresponding Gram matrix, however, these matrices may be ill-conditioned and thus require damping with a matrix D to preserve the stability of the training process.
This modification alters the dynamics of convergence, with the following equation representing the ideal undamped update, for the sake of simplicity it is assumed that the Fisher is invertible for conceptual illustration purposes:
In practice, the damped update is computed:
The relation between these two quantities is as follows:
Or more suitably:
Injecting into the Taylor Lagrange development yields:
It becomes evident that the term (∇mp(mn))T Fn−1Dδmn, might jeopardize the quadratic convergence. A null value of D could trivially preserve the quadratic convergence, but will not solve the ill-conditioning of Fn. An arbitrary selection of damping could therefore decelerate convergence or fail to address the ill-conditioning of the Fisher matrix. With these considerations in mind, it could be stipulated that practitioners might select a damping that aims to fulfill D˜∫∥∇θ
p(θn)∥dp∥δmn∥, which should at least preserve the quadratic convergence rate. This heuristic for selecting a damping value is suggested as a general approach within this disclosure. It serves as a guideline rather than a rule, acknowledging that alternative and potentially more fitting damping parameters might be determined by practitioners skilled in the art, and tailored to specific scenarios and requirements.
Further, the present disclosure describes approaches to resolution. The resolution of the Sifr equation can be approached in multiple ways, but the underlying principle involves discretization in both time and space. Any forward solver that provides the losses and the instantaneous gradient can be employed. The system can be expressed in matrix format as MG δm=l, where MG is a matrix with its first dimension corresponding to np the number of time/frequency steps, and its second dimension reflecting the size of the model nx. At this juncture, it is worth noting that it is advantageous to consider the Gram matrix MGT MG. This significantly reduces the size of the problem, as np is typically much smaller than nx and the inversion would involve a np×np matrix. The inversion of this matrix yields the Sifr update.
Further, for extremely large-scale problems where even the computation of the Gram matrix is computationally prohibitive, an ad-hoc conjugate gradient search can be implemented. In this scenario, N can be expressed as N=MGTv, where v is a vector of size np×1. Damping is an essential component in the inversion process, as is well-known in practice. If the conjugate gradient method is used, the computation of N=MGTv can be efficiently performed by modifying the adjoint source terms
through multiplication along the time axis with the coefficient v.
Further, the present disclosure describes the system architecture for Sifr optimizer implementation. Further, in a specialized embodiment, the present disclosure discloses a system architecture explicitly designed for implementing the Sifr optimizer for waveform inversion and optimization This system comprises a data acquisition unit, a data storage unit, a computational unit, and a user interface. The data acquisition unit is configured to collect oscillatory data from various sensors and instruments, which is then stored in the data storage unit. The computational unit, equipped with high-performance processors and/or GPUs and/or TPU and similar processing specialized devices, is the core of the Waveform inversion/optimization that leverages the Sifr optimizer implementation. It is programmed to execute the Sifr optimizer, performing iterative model updates based on the least squares resolution of the Sifr equation, with a cost/loss function. The computational unit interacts with the data storage unit to fetch the necessary data and store the inversion results. The user interface optionally allows operators to input initial models, set additional constraints, set optimization parameters, and visualize the inversion results. This integrated system architecture ensures seamless data flow and efficient computation, making Sifr optimizer a practical and robust solution for wave-based inversion challenges and optimization. The system can be deployed as a standalone unit or integrated into existing data processing pipelines, offering flexibility and scalability to meet various operational needs.
Further, the disclosure describes the Sifr Optimizer for waveform inversion, an approach designed to enhance both the precision and computational efficiency of waveform inversion processes across time-domain and frequency-domain applications. This innovation refines and expands upon the capabilities of previous technologies, specifically the Time-Domain Sifrian Inversion (TDSI), by introducing a conceptual simplification that eliminates the need for dimensionality reduction and extends its core principles into the frequency domain. The optimizer employs sophisticated mathematical formulations to target discrepancies between observed and synthetic data directly, facilitating robust and accurate interpretation and analysis of waveform data, essential in fields such as geophysics, seismology, and medical imaging. The enhanced Sifr optimizer supports a second-order method with quadratic convergence, offering a significant advancement over traditional Newton-type methods and their limitations.
The present disclosure pertains to waveform inversion, specifically emphasizing enhancing the efficacy of the Sifr optimizer across both time-domain and frequency-domain applications. This disclosure introduces sophisticated methodologies, advanced system architectures, and computer-readable media designed to leverage the Sifr optimizer for substantial improvements in the precision and computational efficiency of waveform inversion processes. Such advancements are critical across various applications, including, but not limited to, geophysics, seismology, and medical imaging, where the need for high-resolution imaging and efficient optimization of wave-based data is paramount. The enhanced variant of the Sifr optimizer refines and expands upon technological developments related to the Time-Domain Sifrian Inversion (TDSI). This refinement includes a conceptual simplification that eliminates the need for dimensionality reduction while further extending its core principles into the frequency domain. This innovation marks a significant progression in the field, seamlessly merging the theoretical aspects of waveform physics with practical, scalable optimization strategies that converge quadratically. Consequently, it facilitates more robust and accurate interpretation and analysis of waveform data.
Further, the present disclosure describes an evolution of the Time-Domain Sifrian Inversion (TDSI). This foundational innovation marked a substantial advancement in waveform inversion by integrating a second-order Lagrangian approach tailored to the time domain and paving the way to adeptly overcoming the intrinsic limitations of traditional Newton-type methods. Central to this advancement is the efficient resolution of the Sifr equation and its subsequent implementation through the Sifr update process, which collectively enhanced the waveform inversion process by integrating sophisticated solutions to address cycle-skipping challenges—phenomena occurring when low-frequency information is missing from recorded data, leading to the optimization process becoming trapped in local minima, thereby yielding inadequate and unrealistic modeling due to the oscillatory nature of the wave data.
Further, the enhancement in the refined version of the Sifr optimizer simplifies the innovations by eliminating the necessity for dimensionality reduction, which was an important step in implementations to reduce the computational burden. The updated variant of the Sifr optimizer focuses on direct minimization strategies of the loss/cost function to significantly improve usability and computational efficiency, rendering it exceptionally suitable for extensive modeling and data analysis across both time and frequency domains. Although this is inspired by the original foundational equations, it conceptually diverges from the original Sifr equation, necessitating comprehensive justification for its reclassification as a second-order method with quadratic convergence.
Further, the enhanced Sifr optimizer employs sophisticated mathematical formulations to streamline the computational process effectively by directly targeting the discrepancies between observed and synthetic data instead of looking for a null gradient. This method not only preserves the precision of second-order accuracy essential for detailed imaging and inversion tasks but also simplifies the iterative update mechanism, rendering the process more accessible and reducing computational demands significantly. The present disclosure bypasses the conceptual limitations by offering a novel interpretation of the underlying equations which led to the Sifr characterization of the second-order update, culminating in a re-envisioned Sifr Optimizer. This new optimizer upholds the fundamental second-order characteristics essential for effective waveform inversion applications, applicable across both temporal and spectral domains.
Moreover, the optimized Sifr optimizer method, designed for enhanced efficiency, relies on the performance of a least squares regression to estimate the model update. Such a regression could be also weighted or generalized and could leverage any suitable solver for the task at hand; in particular estimations of the covariance matrix and, preferably, the Gram matrix of gradients, irrespective of the domain (time or frequency) could be computationally advantageous. With the abrogation of dimensionality reduction, the updated Sifr optimizer presents a direct and robust alternative to waveform inversion that is readily deployable across various scales of data and complexity. This approach is particularly beneficial in domains such as but not limited to geophysics, seismology, and medical imaging, where accuracy and efficiency are imperative for delivering concrete outcomes within practical timeframes.
The theoretical foundations of the updated Sifr optimizer are corroborated by proofs of second-order convergence, illustrating its capacity to achieve accelerated convergence rates and enhanced inversion accuracy. The important role of damping is also highlighted to avoid any ill-conditioning. The adaptability of this method to diverse loss functions and inversion scenarios further highlights its utility across a broad spectrum of waveform inversion applications, suggesting its potential for extensive adoption in contemporary scientific and industrial contexts.
Further, the present disclosure describes significant augmentation to the waveform inversion methodology by introducing a more advanced and efficient Sifr optimizer. It builds upon the successes of the TDSI and previous iterations of the Sifr optimizer, furnishing a more potent and efficient tool designed to address the increasingly complex challenges of waveform inversion in modern applications.
Further, The Sifr Optimizer represents a significant leap in waveform inversion technology, building upon the foundational insights of the Time-Domain Sifrian Inversion (TDSI). The present disclosure introduces an enhanced version of the optimizer that simplifies earlier innovations by removing the necessity for dimensionality reduction, focusing instead on direct minimization strategies of the loss/cost function. This approach not only improves usability and computational efficiency but also adheres to the second-order accuracy crucial for detailed imaging and inversion tasks. By re-envisioning the underlying equations of the Sifr optimizer, this technology bypasses the conceptual limitations highlighted in the present disclosure, offering an interpretation and reclassification as a robust second-order method. Additionally, the optimizer's design includes the ability to perform least squares regression to estimate the model update, adaptable to various scales of data and complexity. This makes the Sifr optimizer a powerful tool poised for widespread adoption in addressing the complex challenges of modem waveform inversion applications.
A user 112, such as the one or more relevant parties, may access online platform 100 through a web based software application or browser. The web based software application may be embodied as, for example, but not be limited to, a website, a web application, a desktop application, and a mobile application compatible with a computing device 1100.
Further, at 204, the method 200 may include obtaining, using a processing device, at least one initial model associated with the at least one object. Further, the at least one initial model may include a Sigsbee salt model, a BP2004 Salt Model, etc. Further, the at least one initial model may include an initial image. Further, the at least one initial model may include a computational model.
Further, at 206, the method 200 may include obtaining, using the processing device, at least one synthetic data from the at least one initial model. Further, the at least one synthetic data may be predicted data, modeled data, etc. Further, the at least one synthetic data may include data associated with second waves. Further, the first waves and the second waves may be oscillatory waves.
Further, at 208, the method 200 may include determining, using the processing device, a matching of the at least one synthetic data with the at least one observed data is not within a matching threshold. Further, the matching corresponds to aligning the at least one synthetic data with the at least one observed data within a set threshold.
Further, at 210, the method 200 may include updating, using the processing device, the at least one initial model using the Time-Domain Sifrian Inversion (TDSI) iteratively until the matching threshold for the matching may be achieved. Further, the Time-Domain Sifrian Inversion (TDSI) provides an update for the updating of the at least one initial model iteratively using at least one time-domain cost function. Further, the at least one time-domain cost function may be robust to a cycle skipping. Further, the at least one time-domain cost function may be a time-oriented cost function. Further, the at least one time-domain cost function may be associated with squared Wasserstein distance W2. Further, the update may include at least one second-order update. Further, the update may include a Sifr update. Further, the update may include a Newton update. Further, the update may be a Time-Domain Sifrian Inversion (TDSI) update. Further, in an embodiment, the updating of the at least one initial model using the Time-Domain Sifrian Inversion (TDSI) iteratively may include iteratively refining the at least one initial model until a match between the at least one synthetic data and the at least one observed data may be achieved. Further, in an embodiment, the updating of the at least one initial model using the Time-Domain Sifrian Inversion (TDSI) iteratively may include optimizing the at least one initial model. Further, in an embodiment, the updating of the at least one initial model using the Time-Domain Sifrian Inversion (TDSI) iteratively may include implementing a Time-Domain Sifrian Inversion (TDSI) optimizer. Further, the TDSI optimizer performs a parameter update of at least one parameter of the at least one initial model leading to a Sifr update in a time-domain using a time-domain cost function. Further, the TDSI optimizer offers a streamlined and efficient solution that capitalizes on second-order curvature information and provides a time-based remedy for the cycle skipping. Further, the using of the Time-Domain Sifrian Inversion (TDSI) may include implementing a time domain Sifrian functional for the providing of the update. Further, the time domain Sifrian functional may be a Sifrian functional associated with the time domain.
Further, at 212, the method 200 may include generating, using the processing device, at least one final model based on the updating. Further, the at least one final model may include a final image. Further, the at least one final model may include a computational model.
Further, at 214, the method 200 may include storing, using a storage device, the at least one final model.
Further, in some embodiments, the at least one device may include at least one sensor. Further, the at least one sensor may be configured for generating the at least one observed data associated with the at least one object. Further, the at least one sensor may include a plurality of sensors arranged in at least one configuration. Further, the at least one configuration corresponds to a location of each of the plurality of sensors in relation to the at least one object. Further, the at least one sensor may include a waveform sensor. Further, the waveform sensor may be configured for capturing data related to a waveform signal associated with the at least one object. Further, the waveform signal may include mechanical wave signals such as seismic, sound, acoustic, and elastic, or electrical wave signals such as electromagnetic waves. Further, the at least one sensor may include Oscilloscopes, Seismometers, Microphones, Pressure Transducers, Temperature Sensors, etc.
Further, in some embodiments, the updating of the at least one initial model using the Time-Domain Sifrian Inversion (TDSI) may include augmenting at least one of a wavefield equation, an adjoint wavefield, and a gradient definition of the at least one initial model by introducing at least one augmentation variable to at least one of the wavefield equation, the adjoint wavefield, and the gradient definition. Further, the at least one augmentation variable may include at least one adjoint variable. Further, the updating of the at least one initial model using the Time-Domain Sifrian Inversion (TDSI) may include assembling at least one of an augmented wavefield function, an augmented adjoint wavefield, and an augmented gradient definition of the at least one initial model for creating a time domain Sifrian functional based on the augmenting. Further, the Time-Domain Sifrian Inversion (TDSI) provides the update for the updating of the at least one initial model iteratively using the time domain Sifrian functional. Further, the time domain Sifrian functional and at least one derivative of the time domain Sifrian functional become null when equilibrium conditions for forward, backward, and gradient definitions are met.
Further, in an embodiment, the updating of the at least one initial model using the Time-Domain Sifrian Inversion (TDSI) may include performing a variational perturbation for characterizing the update using the time domain Sifrian functional. Further, the Time-Domain Sifrian Inversion (TDSI) provides the update for the updating of the at least one initial model iteratively based on the characterizing of the update.
Further, in an embodiment, the updating of the at least one initial model using the Time-Domain Sifrian Inversion (TDSI) may include characterizing a second-order Newton direction for the updating of the at least one initial model based on the time domain Sifrian functional. Further, the Time-Domain Sifrian Inversion (TDSI) provides the update for the updating of the at least one initial model iteratively based on the characterizing of the second-order Newton direction.
Further, in an embodiment, the time domain Sifrian functional may be:
Further, in an embodiment, the updating of the at least one initial model using the Time-Domain Sifrian Inversion (TDSI) may include obtaining at least one Sifrian equation from the time domain Sifrian functional. Further, the at least one Sifrian equation may include a system of equations, an enhanced Sifrian system of equations, etc. Further, the at least one Sifrian equation addresses a non-convexity issue inherent in the wave-based inversion. Further, the Time-Domain Sifrian Inversion (TDSI) provides the update for the updating of the at least one initial model iteratively based on the at least one Sifrian equation. Further, in an embodiment, the at least one Sifrian equation may be:
In further embodiments, the method 200 may include initializing, using the processing device, at least one parameter of the at least one initial model based on the obtaining of the at least one initial model. Further, the at least one parameter may include a velocity, a resistivity, etc. Further, the obtaining of the at least one synthetic data may be based on the initializing. Further, the updating of the at least one initial model may include updating the at least one parameter using the Time-Domain Sifrian Inversion (TDSI) iteratively until the matching threshold for the matching may be achieved.
Further, in an embodiment, the updating of the at least one parameter using the Time-Domain Sifrian Inversion (TDSI) iteratively may include employing at least one of a time discretization and a space discretization for reducing a size of a parameter search space. Further, the Time-Domain Sifrian Inversion (TDSI) provides the update for the updating of the at least one initial model iteratively by selecting at least one parameter value for the at least one parameter from the parameter search space.
Further, the communication device 302 may be configured for receiving at least one observed data associated with at least one object from at least one device 402, a shown in
Further, the processing device 304 may be communicatively coupled with the communication device 302. Further, the processing device 304 may be configured for obtaining at least one initial model associated with the at least one object. Further, the processing device 304 may be configured for obtaining at least one synthetic data from the at least one initial model. Further, the processing device 304 may be configured for determining a matching of the at least one synthetic data with the at least one observed data is not within a matching threshold. Further, the processing device 304 may be configured for updating the at least one initial model using the Time-Domain Sifrian Inversion (TDSI) iteratively until the matching threshold for the matching may be achieved. Further, the Time-Domain Sifrian Inversion (TDSI) provides an update for the updating of the at least one initial model iteratively using at least one time-domain cost function. Further, the at least one time-domain cost function may be robust to a cycle skipping. Further, the processing device 304 may be configured for generating at least one final model based on the updating. Further, the processing device 304 may include a computational unit, a data processing unit, etc.
Further, the storage device 306 may be communicatively coupled with the processing device 304. Further, the storage device 306 may be configured for storing the at least one final model.
Further, in some embodiments, the at least one device 402 may include at least one sensor 404, as shown in
Further, in some embodiments, the updating of the at least one initial model using the Time-Domain Sifrian Inversion (TDSI) may include augmenting at least one of a wavefield equation, an adjoint wavefield, and a gradient definition of the at least one initial model by introducing at least one augmentation variable to at least one of the wavefield equation, the adjoint wavefield, and the gradient definition. Further, the updating of the at least one initial model using the Time-Domain Sifrian Inversion (TDSI) may include assembling at least one of an augmented wavefield function, an augmented adjoint wavefield, and an augmented gradient definition of the at least one initial model for creating a time domain Sifrian functional based on the augmenting. Further, the Time-Domain Sifrian Inversion (TDSI) further provides the update for the updating of the at least one initial model iteratively using the time domain Sifrian functional.
Further, in an embodiment, the updating of the at least one initial model using the Time-Domain Sifrian Inversion (TDSI) further may include performing a variational perturbation for characterizing the update using the time domain Sifrian functional. Further, the Time-Domain Sifrian Inversion (TDSI) further provides the update for the updating of the at least one initial model iteratively based on the characterizing of the update.
Further, in an embodiment, the updating of the at least one initial model using the Time-Domain Sifrian Inversion (TDSI) further may include characterizing a second-order Newton direction for the updating of the at least one initial model based on the time domain Sifrian functional. Further, the Time-Domain Sifrian Inversion (TDSI) further provides the update for the updating of the at least one initial model iteratively based on the characterizing of the second-order Newton direction.
Further, in an embodiment, the time domain Sifrian functional may be:
Further, in an embodiment, the updating of the at least one initial model using the Time-Domain Sifrian Inversion (TDSI) further may include obtaining at least one Sifrian equation from the time domain Sifrian functional. Further, the at least one Sifrian equation addresses a non-convexity issue inherent in the wave-based inversion. Further, the Time-Domain Sifrian Inversion (TDSI) further provides the update for the updating of the at least one initial model iteratively based on the at least one Sifrian equation. Further, in an embodiment, the at least one Sifrian equation may be:
Further, in some embodiments, the processing device 304 may be further configured for initializing at least one parameter of the at least one initial model based on the obtaining of the at least one initial model. Further, the obtaining of the at least one synthetic data may be based on the initializing. Further, the updating of the at least one initial model may include updating the at least one parameter using the Time-Domain Sifrian Inversion (TDSI) iteratively until the matching threshold for the matching may be achieved.
Further, in an embodiment, the updating of the at least one parameter using the Time-Domain Sifrian Inversion (TDSI) iteratively may include employing at least one of a time discretization and a space discretization for reducing a size of a parameter search space. Further, the Time-Domain Sifrian Inversion (TDSI) provides the update for the updating of the at least one initial model iteratively by selecting at least one parameter value for the at least one parameter from the parameter search space.
With reference to
Computing device 1100 may have additional features or functionality. For example, computing device 1100 may also include additional data storage devices (removable and/or non-removable) such as, for example, magnetic disks, optical disks, or tape. Such additional storage is illustrated in
Computing device 1100 may also contain a communication connection 1116 that may allow device 1100 to communicate with other computing devices 1118, such as over a network in a distributed computing environment, for example, an intranet or the Internet.
Communication connection 1116 is one example of communication media. Communication media may typically be embodied by computer readable instructions, data structures, program modules, or other data in a modulated data signal, such as a carrier wave or other transport mechanism, and includes any information delivery media. The term “modulated data signal” may describe a signal that has one or more characteristics set or changed in such a manner as to encode information in the signal. By way of example, and not limitation, communication media may include wired media such as a wired network or direct-wired connection, and wireless media such as acoustic, radio frequency (RF), infrared, and other wireless media. The term computer readable media as used herein may include both storage media and communication media.
As stated above, a number of program modules and data files may be stored in system memory 1104, including operating system 1105. While executing on processing unit 1102, programming modules 1106 (e.g., application 1120 such as a media player) may perform processes including, for example, one or more stages of methods, algorithms, systems, applications, servers, databases as described above. The aforementioned process is an example, and processing unit 1102 may perform other processes. Other programming modules that may be used in accordance with embodiments of the present disclosure may include machine learning applications.
Generally, consistent with embodiments of the disclosure, program modules may include routines, programs, components, data structures, and other types of structures that may perform particular tasks or that may implement particular abstract data types. Moreover, embodiments of the disclosure may be practiced with other computer system configurations, including hand-held devices, general purpose graphics processor-based systems, multiprocessor systems, microprocessor-based or programmable consumer electronics, application specific integrated circuit-based electronics, minicomputers, mainframe computers, and the like. Embodiments of the disclosure may also be practiced in distributed computing environments where tasks are performed by remote processing devices that are linked through a communications network. In a distributed computing environment, program modules may be located in both local and remote memory storage devices.
Furthermore, embodiments of the disclosure may be practiced in an electrical circuit comprising discrete electronic elements, packaged or integrated electronic chips containing logic gates, a circuit utilizing a microprocessor, or on a single chip containing electronic elements or microprocessors. Embodiments of the disclosure may also be practiced using other technologies capable of performing logical operations such as, for example, AND, OR, and NOT, including but not limited to mechanical, optical, fluidic, and quantum technologies. In addition, embodiments of the disclosure may be practiced within a general-purpose computer or in any other circuits or systems.
Embodiments of the disclosure, for example, may be implemented as a computer process (method), a computing system, or as an article of manufacture, such as a computer program product or computer readable media. The computer program product may be a computer storage media readable by a computer system and encoding a computer program of instructions for executing a computer process. The computer program product may also be a propagated signal on a carrier readable by a computing system and encoding a computer program of instructions for executing a computer process. Accordingly, the present disclosure may be embodied in hardware and/or in software (including firmware, resident software, micro-code, etc.). In other words, embodiments of the present disclosure may take the form of a computer program product on a computer-usable or computer-readable storage medium having computer-usable or computer-readable program code embodied in the medium for use by or in connection with an instruction execution system. A computer-usable or computer-readable medium may be any medium that can contain, store, communicate, propagate, or transport the program for use by or in connection with the instruction execution system, apparatus, or device.
The computer-usable or computer-readable medium may be, for example but not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, device, or propagation medium. More specific computer-readable medium examples (a non-exhaustive list), the computer-readable medium may include the following: an electrical connection having one or more wires, a portable computer diskette, a random-access memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or Flash memory), an optical fiber, and a portable compact disc read-only memory (CD-ROM). Note that the computer-usable or computer-readable medium could even be paper or another suitable medium upon which the program is printed, as the program can be electronically captured, via, for instance, optical scanning of the paper or other medium, then compiled, interpreted, or otherwise processed in a suitable manner, if necessary, and then stored in a computer memory.
Embodiments of the present disclosure, for example, are described above with reference to block diagrams and/or operational illustrations of methods, systems, and computer program products according to embodiments of the disclosure. The functions/acts noted in the blocks may occur out of the order as shown in any flowchart. For example, two blocks shown in succession may in fact be executed substantially concurrently or the blocks may sometimes be executed in the reverse order, depending upon the functionality/acts involved.
While certain embodiments of the disclosure have been described, other embodiments may exist. Furthermore, although embodiments of the present disclosure have been described as being associated with data stored in memory and other storage mediums, data can also be stored on or read from other types of computer-readable media, such as secondary storage devices, like hard disks, solid state storage (e.g., USB drive), or a CD-ROM, a carrier wave from the Internet, or other forms of RAM or ROM. Further, the disclosed methods' stages may be modified in any manner, including by reordering stages and/or inserting or deleting stages, without departing from the disclosure.
According to some aspects, a method for wave-based inversion using Time-Domain Sifrian Inversion (TDSI), is disclosed. Further, the method may include receiving an initial model of a surveyed medium. Further, the method may include implementing the TDSI optimizer that performs parameter updates leading to the Sifr update in the time-domain using a time-domain cost function. Further, the method may include optimizing the model parameters until convergence of the optimization process or a stopping criterion is met.
Further, according to some aspects, the TDSI optimizer mitigates computational complexities associated with second-order optimization by solving for a reduced number of parameters equivalent to the number of time steps or less, irrespective of the model parameter dimensions.
Further, according to some aspects, the TDSI optimizer utilizes cost functions that are robust to cycle skipping, employs the Sifr equation, and uses damping to enhance the conditioning of its inherent system, thereby facilitating more effective model inversion.
Further, according to some aspects, the TDSI optimizer employs time and space discretization to reduce the parameter search space, leading to improved computational efficiency.
Further, according to some aspects, the TDSI optimizer performs the inversion across various wave-based survey architectures, including but not limited to seismic, electromagnetic, and other wave-based surveys.
Further, according to some aspects, the TDSI optimizer accomplishes a rapid reduction in the cost function per iteration while maintaining competitive computational performance.
Further, according to some aspects, the TDSI optimizer enables accelerated convergence and improved fitting in wave-based inversion by considering parameter search with a suitable cost function to create a cycle-skipping robust Sifr update in the time domain.
Further, according to some aspects, the TDSI optimizer effectively performs inversion for surveys with a large number of sensors and parameters, demonstrating its scalability and adaptability to various scales of survey implementations.
Further, according to some aspects, the TDSI optimizer effectively performs inversion for various tasks, including but not limited to subsurface imaging, resource exploration, medical imaging, material investigation, and environmental monitoring.
Further, according to some aspects, the TDSI optimizer effectively performs second-order inversion with various types of data, including but not limited to seismic, electromagnetic, and other oscillatory data, illustrating its versatility across different data types and tasks.
Further, according to some aspects, the TDSI optimizer addresses non-convexity issues that are inherent in waveform inversion methods by including quadratic terms when canceling the gradient, thereby solving the non-convexity issue and further enhancing the inversion process.
Further, according to some aspects, the TDSI optimizer applies theoretical considerations that depart from standard inversion methods, yielding an optimizer that performs second-order updates without the computational complexity of forming the Hessian or its inverse.
Further, according to some aspects, the TDSI optimizer achieves dimensionality reduction during the inversion process by leveraging the differential properties inherent in wave-based phenomena and adjoint wavefields in wave-based surveys.
Further, according to some aspects, the TDSI optimizer effectively optimizes various survey architectures regardless of their size, demonstrating the scalability and versatility of the method.
According to some aspects, a device for wave-based inversion is disclosed. Further, the device is configured with standard components and potentially benefiting from additional resources such as multiple CPUs, GPUs, and extra RAM. Further, the device may include an interface for inputting the initial model of the surveyed medium. Further, the device may include a data processing unit implementing the TDSI optimizer for performing second-order, cycle-skipping robust inversion.
Further, according to some aspects, the TDSI optimizer in the data processing unit characterizes the TDSI update for optimization by reducing the parameter space to a vector of maximum size equivalent to the number of time steps or less.
Further, according to some aspects, the TDSI optimizer in the data processing unit performs second-order updates, accommodating time and space discretization, and reducing dimensionality, leading to an efficient and accurate inversion process.
Further, according to some aspects, the TDSI optimizer in the data processing unit optimizes the inversion across various survey architectures, providing enhanced second-order inversion functionality across different survey designs.
Further, according to some aspects, the data processing unit is configured with specific components and settings to facilitate the implementation of the TDSI optimizer, providing a hardware solution that supports the efficient inversion of oscillatory data in the time domain.
Further, according to some aspects, the TDSI optimizer in the data processing unit can be efficiently implemented on multiple GPUs and with extensive RAM, demonstrating the flexibility of the method with respect to hardware configurations.
Accordingly, at 1202, the method 1200 may include a step of receiving an initial model m0.
Further, at 1204, the method 1200 may include a step of obtaining a model m.
Further, at 1206, the method 1200 may include a step of receiving at least one input from a source Q.
Further, at 1208, the method 1200 may include a step of performing forward propagation in the model m based on the at least one input. Further, the forward propagation may be represented as L(m)u+q=0. Further, performing of the forward propagation may output a forward wavefield u.
Further, at 1210, the method 1200 may include a step of applying a sampling operator R to the forward wavefield and generating Ru.
Further, at 1212, the method 1200 may include a step of receiving at least one sensor data.
Further, at 1214, the method 1200 may include a step of obtaining a cost function based on the at least one sensor data and Ru. Further, the cost function may be represented as:
Further, at 1216, the method 1200 may include a step of determining a residual based on the cost function. Further, the residuals may be represented as
Further, at 1218, the method 1200 may include a step of performing backward propagation for determining an adjoint wavefield λ. Further, the backward propagation may be represented as:
Further, at 1220, the method 1200 may include a step of performing gradient computation based on the adjoint wavefield λ. Further, the gradient computation may include providing an update with δm. Further, the δm may be:
Further, at 1222, the method 1200 may include a step of performing step length search and providing the updates to the model m.
Further, at 1224, the method 1200 may include a step of determining a meeting of convergence criteria.
Further, at 1226, the method 1200 may include a step of generating an inverted model m if the convergence criteria are met.
Further,
Further, at 1304, the method 1300 may include a step of obtaining a model m.
Further, at 1306, the method 1300 may include a step of receiving at least one input from a source Q.
Further, at 1308, the method 1300 may include a step of performing forward propagation in the model m based on the at least one input. Further, the forward propagation may be represented as L(m)u+q=0. Further, the performing of the forward propagation may output a forward wavefield u.
Further, at 1310, the method 1300 may include a step of applying a sampling operator R to the forward wavefield and generating Ru.
Further, at 1312, the method 1300 may include a step of receiving at least one sensor data.
Further, at 1314, the method 1300 may include a step of obtaining a cost function based on the at least one sensor data and Ru. Further, the cost function may be represented as:
Further, at 1316, the method 1300 may include a step of determining a residual based on the cost function. Further, the residuals may be represented as
Further, at 1318, the method 1300 may include a step of performing backward propagation for determining an adjoint wavefield λ. Further, the backward propagation may be represented as:
Further, at 1320, the method 1300 may include a step of gradients computation based on the adjoint wavefield λ. Further, the gradients computation may be based on:
Further, at 1322, the method 1300 may include a step of determining a sifr equation based on an output from the gradients computation and the cost function. Further, the sifr equation may include
Further, at 1324, the method 1300 may include a step of implementing a solver for solving the Sifr equation. Further, the solver provides updates with δm.
Further, at 1326, the method 1300 may include a step of performing step length search for updating the model m based the updates with δm.
Further, at 1328, the method 1300 may include a step of a meeting of convergence criteria.
Further, at 1330, the method 1300 may include a step of generating an inverted model m if the convergence criteria are met.
Further,
Further, at 1504, the method 1500 may include obtaining, using a processing device, at least one initial model associated with the at least one object. Further, the at least one initial model may include a Sigsbee salt model, a BP2004 Salt Model, etc. Further, the at least one initial model may include an initial image. Further, the at least one initial model may include a computational model.
Further, at 1506, the method 1500 may include obtaining, using the processing device, at least one synthetic data from the at least one initial model. Further, the at least one synthetic data may be predicted data, modeled data, etc. Further, the at least one synthetic data may include data associated with second waves. Further, the first waves and the second waves may be oscillatory waves.
Further, at 1508, the method 1500 may include determining, using the processing device, a discrepancy between the at least one observed data and the at least one synthetic data. Further, the discrepancy corresponds to an unalignment, an anomaly, a mismatch, etc., in the at least one synthetic data with the at least one observed data.
Further, at 1510, the method 1500 may include creating, using the processing device, a cost function based on the determining of the discrepancy. Further, the cost function may include a loss function. Further, the cost function may include a function used to map an event or values of one or more variables onto a real number intuitively representing some “cost/loss” associated with the event. Further, the cost function may include at least one of an integral and a summation over at least one sample of the domain representation. Further, the domain representation may include a unified representation of time/frequency with a unified sample.
Further, at 1512, the method 1500 may include updating, using the processing device, the at least one initial model iteratively using the Sifr optimizer until at least one condition is met. Further, the at least one condition may include at least one of a minimization of the cost function below a threshold, an elapsing of a given number of iterations, and a meeting of a convergence criterion. Further, the Sifr optimizer may be used for waveform inversion. Further, the minimization of the cost function may mean substantially reducing the cost function, minimizing the cost function, etc. Further, the Sifr optimizer provides an update for the updating of the at least one initial model iteratively based on a regression comprising a least squares resolution of at least one Sifr equation. Further, the least squares resolution may include a parameter estimation method in regression analysis based on minimizing the sum of the squares of the residuals made in the results of the equation. Further, the at least one sifr equation may include
Further, at 1514, the method 1500 may include generating, using the processing device, at least one final model based on the updating. Further, the at least one final model may include a second computational model.
Further, at 1516, the method 1500 may include storing, using a storage device, the at least one final model.
Further, in some embodiments, the at least one device may include at least one sensor. Further the at least one sensor may be configured for generating the at least one observed data associated with the at least one object. Further, the at least one sensor may include a plurality of sensors arranged in at least one configuration. Further, the at least one configuration corresponds to a location of each of the plurality of sensors in relation to the at least one object. Further, the at least one sensor may include a waveform sensor. Further, the waveform sensor may be configured for capturing data related to a waveform signal associated with the at least one object. Further, the waveform signal may include mechanical wave signals such as seismic, sound, acoustic, and elastic, or electrical wave signals such as electromagnetic waves. Further, the at least one sensor may include Oscilloscopes, Seismometers, Microphones, Pressure Transducers, Temperature Sensors, etc.
Further, in some embodiments, the at least one Sifr equation may include:
Further, in some embodiments, the at least one observed data may include at least one sample associated with at least one domain representation of at least one domain. Further, the at least one sample may include a value set associated with the at least one domain representation. Further, the at least one domain representation may include a representation associated with at least one of time domain and frequency domain.
Further, in an embodiment, the cost function involves at least one of an integral and a summation over at least one of the at least one sample associated with the at least one domain representation. Further, the at least one sample may include a time domain value set, a frequency domain value set, a unified sample p, etc.
In further embodiments, the method 1500 may include initializing, using the processing device, at least one parameter of the at least one initial model based on the obtaining of the at least one initial model. Further, the obtaining of the at least one synthetic data may be based on the initializing. Further, the updating of the at least one initial model may include updating the at least one parameter iteratively until the at least one condition may be met. Further, the at least one parameter may include a velocity, a resistivity, etc. Further, the obtaining of the at least one synthetic data may be based on the initializing. Further, the updating of the at least one initial model may include updating the at least one parameter.
Further, in an embodiment, the updating of the at least one initial model further may include dynamically adjusting an update magnitude for the at least one parameter for each of a plurality of iterations using a step length search. Further, the updating of the at least one parameter iteratively until the at least one condition may be met may be based on the dynamically adjusting of the update magnitude. Further, the step length search may be configured for dynamically adjusting the update magnitude.
Further, in some embodiments, the updating of the at least one initial model iteratively using the Sifr optimizer until the at least one condition may be met may include performing the regression by the Sifr optimizer for estimating the update for the at least one initial model. Further, the providing of the update may be further based on the performing of the regression.
Further, in an embodiment, the least squares regression may include at least one of a least squares regression, a weighted least squares regression, and a generalized least squares regression. Further, the weighted least squares regression may include a generalization of least squares and linear regression in which knowledge of the unequal variance of observations (heteroscedasticity) is incorporated into the regression. Further, the generalized least squares regression may include a method used to estimate the unknown parameters in a linear regression model. Further, the least squares regression may be a parameter estimation method in regression analysis based on minimizing the sum of the squares of the residuals made in the results of the equation.
Further, in an embodiment, the updating of the at least one initial model iteratively using the Sifr optimizer until the at least one condition may be met further may include leveraging at least one solver from a plurality of solvers by the Sifr optimizer. Further, the performing of the regression by the Sifr optimizer may include performing the regression by the Sifr optimizer for the estimating of the update for the at least one initial model using the at least one solver based on the leveraging of the at least one solver. Further, the at least one solver may include a forward solver configured for providing losses and instantaneous gradient.
Further, the communication device 1602 may be configured for receiving at least one observed data associated with at least one object from at least one device 1702, as shown in
Further, the processing device 1604 may be communicatively coupled with the communication device 1602. Further, the processing device 1604 may be configured for obtaining at least one initial model associated with the at least one object. Further, the at least one initial model may include a Sigsbee salt model, a BP2004 Salt Model, etc. Further, the at least one initial model may include an initial image. Further, the at least one initial model may include a computational model.
Further, the processing device 1604 may be configured for obtaining at least one synthetic data from the at least one initial model. Further, the at least one synthetic data may be predicted data, modeled data, etc. Further, the at least one synthetic data may include data associated with second waves. Further, the first waves and the second waves may be oscillatory waves.
Further, the processing device 1604 may be configured for determining a discrepancy between the at least one observed data and the at least one synthetic data. Further, the discrepancy corresponds to an unalignment in the at least one synthetic data with the at least one observed data.
Further, the processing device 1604 may be configured for creating a cost function based on the determining of the discrepancy. Further, the cost function may include a function used to map an event or values of one or more variables onto a real number intuitively representing some “cost/loss” associated with the event. Further, the cost function may include at least one of an integral, a summation over at least one sample of the domain representation. Further, the domain representation may include a unified representation of time/frequency with a unified sample.
Further, the processing device 1604 may be configured for updating the at least one initial model iteratively using the Sifr optimizer until the at least one condition may be met. Further, the at least one condition may include at least one of a minimization of the cost function below a threshold, an elapsing of a given number of iterations, and a meeting of a convergence criterion. Further, the Sifr optimizer provides an update for the updating of the at least one initial model iteratively based on a regression comprising a least squares resolution of at least one Sifr equation. Further, the least squares resolution may include a parameter estimation method in regression analysis based on minimizing the sum of the squares of the residuals made in the results of each individual equation. Further, the at least one sifr equation may include
Further, the processing device 1604 may be configured for generating at least one final model based on the updating. Further, the at least one final model may include a second computational model. Further, the cost function associated with the second computational model and the first computational model may be a null value.
Further, the storage device 1606 may be communicatively coupled with the processing device 1604. Further, the storage device 1606 may be configured for storing the at least one final model.
Further, in some embodiments, the at least one device 1702 may include at least one sensor 1704, as shown in
Further, in some embodiments, the at least one Sifr equation may include:
Further, in some embodiments, the at least one observed data may include at least one sample associated with at least one domain representation of at least one domain. Further, the at least one sample may include a value set associated with the at least one domain representation. Further, the at least one domain representation may include a representation associated with at least one of time domain and frequency domain.
Further, in some embodiments, the cost function involves at least one of an integral and a summation over at least one of the at least one sample associated with the at least one domain representation. Further, the at least one sample may include a time domain value set, a frequency domain value set, a unified sample p, etc.
Further, in some embodiments, the processing device 1604 may be further configured for initializing at least one parameter of the at least one initial model based on the obtaining of the at least one initial model. Further, the obtaining of the at least one synthetic data may be based on the initializing. Further, the updating of the at least one initial model may include updating the at least one parameter iteratively until the at least one condition may be met. Further, the at least one parameter may include a velocity, a resistivity, etc. Further, the obtaining of the at least one synthetic data may be based on the initializing. Further, the updating of the at least one initial model may include updating the at least one parameter.
Further, in some embodiments, the updating of the at least one initial model further may include dynamically adjusting an update magnitude for the at least one parameter for each of a plurality of iterations using a step length search. Further, the updating of the at least one parameter iteratively until the at least one condition may be metmay be based on the dynamically adjusting of the update magnitude. Further, the updated magnitude may include a new mathematical value associated with the at least one parameter. Further, the step length search may be configured for dynamically adjusting the update magnitude.
Further, in some embodiments, the updating of the at least one initial model iteratively using the Sifr optimizer until the at least one condition may be met may include performing the regression by the Sifr optimizer for estimating the update for the at least one initial model. Further, the providing of the update may be further based on the performing of the regression.
Further, in some embodiments, the least squares regression may include at least one of a least squares regression, a weighted least squares regression, and a generalized least squares regression. Further, the weighted least squares regression may include a generalization of least squares and linear regression in which knowledge of the unequal variance of observations (heteroscedasticity) is incorporated into the regression. Further, the generalized least squares regression may include a method used to estimate the unknown parameters in a linear regression model. Further, the least squares regression is a parameter estimation method in regression analysis based on minimizing the sum of the squares of the residuals made in the results of each individual equation.
Further, in some embodiments, the updating of the at least one initial model iteratively using the Sifr optimizer until the at least one condition may be met further may include leveraging at least one solver from a plurality of solvers by the Sifr optimizer. Further, the performing of the regression by the Sifr optimizer may include performing the regression by the Sifr optimizer for the estimating of the update for the at least one initial model using the at least one solver based on the leveraging of the at least one solver. Further, the at least one solver may include a forward solver configured for providing losses and instantaneous gradient.
Although the present disclosure has been explained in relation to its preferred embodiment, it is to be understood that many other possible modifications and variations can be made without departing from the spirit and scope of the disclosure.
The current application is a continuation-in-part (CIP) application of the U.S. non-provisional application Ser. No. 18/506,736 filed on Nov. 10, 2023.
Number | Date | Country | |
---|---|---|---|
Parent | 18506736 | Nov 2023 | US |
Child | 18756689 | US |