The present disclosure is directed at methods, systems, and techniques for simulating an event.
The combination of ubiquitous Internet connectivity, the prevalence of various types of electronic sensors, and the dropping cost of data storage has resulted in seemingly ever increasing amounts of data being collected and stored. This data is often analyzed to determine whether it contains certain “events” as measured by the sensors, the nature of which are context-dependent. Given the amount of data that is collected and consequently needs to be analyzed, that analysis is often done using computer-implemented methods, such as by applying machine learning.
According to a first aspect, there is provided a method comprising obtaining simulated event data comprising a simulated event and authentic raw data; and combining the simulated event data and the authentic raw data to form blended data that comprises the simulated event.
The method may further comprise subsequently processing the blended data and identifying the simulated event therein.
Combining the simulated event data and the raw data to form blended data may comprise: respectively converting the simulated event data and the authentic raw data into frequency domain representations thereof; summing the frequency domain representations of the simulated event data and the authentic raw data together to form a frequency domain representation of the blended data; and converting the frequency domain representation of the blended data into a time domain representation of the blended data.
The blended data may be expressed as a power spectral density.
The simulated event data may be expressed as a power spectral density when combined with the authentic raw data.
The simulated event data may comprise recorded authentic events.
The method may further comprise generating the simulated event data using a generative adversarial network, and some of the authentic raw data may be input to the generative adversarial network to permit generation of the simulated event data.
The generative adversarial network may comprise a generator and a discriminator, all layers except an output layer of the discriminator may use leaky rectified linear unit activation, the output layer of the discriminator may use tan h activation, and all layers of the generator may use leaky rectified linear unit activation.
The authentic raw data may comprise acoustic data.
Obtaining the authentic raw data may comprise performing optical fiber interferometry using fiber Bragg gratings.
Obtaining the authentic raw data may comprise performing distributed acoustic sensing.
The authentic raw data may be obtained and combined with the simulated event data in real-time.
The authentic raw data may be obtained by recording acoustics proximate a pipeline, and the simulated event may comprise a pipeline leak.
According to another aspect, there is provided a system comprising: a processor; a database that is communicatively coupled to the processor and that has simulated event data stored thereon; a memory that is communicatively coupled to the processor and that has stored thereon computer program code that is executable by the processor and that, when executed by the processor, causes the processor to perform the foregoing method.
According to another aspect, there is provided a non-transitory computer readable medium having stored thereon computer program code that is executable by a processor and that, when executed by the processor, causes the processor to perform the foregoing method.
This summary does not necessarily describe the entire scope of all aspects. Other aspects, features and advantages will be apparent to those of ordinary skill in the art upon review of the following description of specific embodiments.
In the accompanying drawings, which illustrate one or more example embodiments:
A variety of types of data can be electronically collected using sensors for processing, which processing may be performed in real-time and/or in a delayed fashion if the collected data is stored for subsequent processing. This data may be processed by a computer system in order to detect “events” in that data, with the nature of the event depending at least in part on the type of data being processed.
For example, the data may be collected using sensors that comprise part of:
1. a pipeline leak detection system, and the event may comprise a leak in the pipeline;
2. a perimeter security system, and the event may comprise an intrusion event;
3. a patient monitoring system, and the event may comprise a cardiac event;
4. a geotechnical monitoring system, and the event may comprise a strain event; or
5. a camera-based event detection system, and the event may comprise any sort of anomalous visual event captured by the camera, such as the presence of a particular gas in the environment.
A machine learning based system, such as an object classifier used to classify images captured from a camera that is implemented using an artificial neural network, such as a convolutional neural network, may be used for computer-implemented event detection.
An event that is represented in data recorded by the sensors is herein referred to as an “authentic event”. Depending on the nature of the system, authentic events may be very rare. In fact, for some systems such as a pipeline leak detection system, authentic events may ideally never occur. Regardless, it is important to periodically test the event detection system, particularly when authentic events are infrequent but serious.
Testing may in some situations be done by physically creating an authentic event and then monitoring the event detection system's response to it. However, this can be costly and risky, such as in the context of a pipeline leak. Moreover, in some situations it may not be possible: sensors may be inaccessible (e.g., they may be buried), or an authentic event may simply not be able to be created (e.g., for a patient monitoring system, it is not feasible to test using an authentic, acute medical event).
Consequently, in at least some embodiments herein, a “simulated event” may at least in part be computer generated and input to the event detection system. The simulated event is designed to mimic an authentic event. The response of the event detection system to one or more simulated events may be monitored, and the event detection system may consequently be tested in this manner. The simulated event may be generated using, for example, an artificial neural network such as a generative adversarial network (“GAN”). Additionally or alternatively, the simulated event may be generated based off a recording of an authentic event and blended with a data stream for processing by the event detection system. The integration of the simulated event with the data stream is performed in a manner that prevents the event detection system from recognizing the simulated event by virtue of how it is integrated into the data stream.
By using simulated events in this manner, the need to go into the field and replicate authentic events for testing is eliminated. Further, simulated events can be introduced into the event detection system's data stream under any number of ambient conditions, which may be difficult to replicate in the field. Simulated events may also be easily triggered as desired: they may, for example, be used on demand; periodically according to a preset schedule, or upon detection or determination of a certain condition (e.g., when measured noise levels go high, a simulated event may be introduced to see if the event is still detectable despite the event detection system having to perform processing with a higher noise floor).
As mentioned above, in at least some example embodiments the event detection system comprises a pipeline leak detection system. In those embodiments, leaks may be detected as acoustic events. Fiber optic cables are often used as distributed measurement systems in acoustic sensing applications. Pressure changes, due to sound waves for example, in the space immediately surrounding an optical fiber and that encounter the optical fiber cause dynamic strain in the optical fiber. Optical interferometry may be used to detect the dynamic strain along a segment of the fiber. Optical interferometry is a technique in which two separate light pulses, a sensing pulse and a reference pulse, are generated and interfere with each other. The sensing and reference pulses may, for example, be directed along an optical fiber that comprises fiber Bragg gratings. The fiber Bragg gratings partially reflect the pulses back towards an optical receiver at which an interference pattern is observed.
The nature of the interference pattern observed at the optical receiver provides information on the optical path length the pulses traveled, which in turn provides information on parameters such as the strain experienced by the segment of optical fiber between the fiber Bragg gratings. Information on the strain then provides information about the event that caused the strain.
Referring now to
The optical fiber 112 comprises one or more fiber optic strands, each of which is made from quartz glass (amorphous SiO2). The fiber optic strands are doped with various elements and compounds (including germanium, erbium oxides, and others) to alter their refractive indices, although in alternative embodiments the fiber optic strands may not be doped. Single mode and multimode optical strands of fiber are commercially available from, for example, Corning® Optical Fiber. Example optical fibers include ClearCurve™ fibers (bend insensitive), SMF28 series single mode fibers such as SMF-28 ULL fibers or SMF-28e fibers, and InfiniCor® series multimode fibers.
The interrogator 106 generates the sensing and reference pulses and outputs the reference pulse after the sensing pulse. The pulses are transmitted along optical fiber 112 that comprises a first pair of FBGs. The first pair of FBGs comprises first and second FBGs 114a,b (generally, “FBGs 114”). The first and second FBGs 114a,b are separated by a certain segment 116 of the optical fiber 112 (“fiber segment 116”). The optical length of the fiber segment 116 varies in response to dynamic strain that the fiber segment 116 experiences.
The light pulses have a wavelength identical or very close to the center wavelength of the FBGs 114, which is the wavelength of light the FBGs 114 are designed to partially reflect; for example, typical FBGs 114 are tuned to reflect light in the 1,000 to 2,000 nm wavelength range. The sensing and reference pulses are accordingly each partially reflected by the FBGs 114a,b and return to the interrogator 106. The delay between transmission of the sensing and reference pulses is such that the reference pulse that reflects off the first FBG 114a (hereinafter the “reflected reference pulse”) arrives at the optical receiver 103 simultaneously with the sensing pulse that reflects off the second FBG 114b (hereinafter the “reflected sensing pulse”), which permits optical interference to occur.
While
The interrogator 106 emits laser light with a wavelength selected to be identical or sufficiently near the center wavelength of the FBGs 114, and each of the FBGs 114 partially reflects the light back towards the interrogator 106. The timing of the successively transmitted light pulses is such that the light pulses reflected by the first and second FBGs 114a,b interfere with each other at the interrogator 106, which records the resulting interference signal. The strain that the fiber segment 116 experiences alters the optical path length between the two FBGs 114 and thus causes a phase difference to arise between the two interfering pulses. The resultant optical power at the optical receiver 103 can be used to determine this phase difference. Consequently, the interference signal that the interrogator 106 receives varies with the strain the fiber segment 116 is experiencing, which allows the interrogator 106 to estimate the strain the fiber segment 116 experiences from the received optical power. The interrogator 106 digitizes the phase difference (“output signal”) whose magnitude and frequency vary directly with the magnitude and frequency of the dynamic strain the fiber segment 116 experiences.
The signal processing device 118 is communicatively coupled to the interrogator 106 to receive the output signal. The signal processing device 118 includes a processor 102 and a non-transitory computer-readable medium 104 that are communicatively coupled to each other. An input device 110 and a display 108 interact with the signal processing device 118. The computer-readable medium 104 has stored on it program code to cause the processor 102 (and consequently the signal processing device 118) to perform any suitable signal processing methods to the output signal. For example, if the fiber segment 116 is laid adjacent a region of interest that is simultaneously experiencing vibration at a rate under 20 Hz and acoustics at a rate over 20 Hz, the fiber segment 116 will experience similar strain and the output signal will comprise a superposition of signals representative of that vibration and those acoustics. The signal processing device 118 may apply to the output signal a low pass filter with a cut-off frequency of 20 Hz, to isolate the vibration portion of the output signal from the acoustics portion of the output signal. Analogously, to isolate the acoustics portion of the output signal from the vibration portion, the signal processing device 118 may apply a high-pass filter with a cut-off frequency of 20 Hz. The signal processing device 118 may also apply more complex signal processing methods to the output signal; example methods include those described in PCT application PCT/CA2012/000018 (publication number WO 2013/102252), the entirety of which is hereby incorporated by reference.
Any changes to the optical path length of the fiber segment 116 result in a corresponding phase difference between the reflected reference and sensing pulses at the interrogator 106. Since the two reflected pulses are received as one combined interference pulse, the phase difference between them is embedded in the combined signal. This phase information can be extracted using proper signal processing techniques, such as phase demodulation. The relationship between the optical path of the fiber segment 116 and that phase difference (θ) is as follows:
θ=2πnL/λ,
where n is the index of refraction of the optical fiber, L is the physical path length of the fiber segment 116, and λ is the wavelength of the optical pulses. A change in nL is caused by the fiber experiencing longitudinal strain induced by energy being transferred into the fiber. The source of this energy may be, for example, an object outside of the fiber experiencing dynamic strain, undergoing vibration, or emitting energy. As used herein, “dynamic strain” refers to strain that changes over time. Dynamic strain that has a frequency of between about 5 Hz and about 20 Hz is referred to by persons skilled in the art as “vibration”, dynamic strain that has a frequency of greater than about 20 Hz is referred to by persons skilled in the art as “acoustics”, and dynamic strain that changes at a rate of <1 Hz, such as at 500 μHz, is referred to as “sub-Hz strain”.
Another way of determining ΔnL is by using what is broadly referred to as distributed acoustic sensing (“DAS”). DAS involves laying the fiber 112 through or near a region of interest and then sending a coherent laser pulse along the fiber 112. As shown in
DAS accordingly uses Rayleigh scattering to estimate the magnitude, with respect to time, of the strain experienced by the fiber during an interrogation time window, which is a proxy for the magnitude of the vibration or acoustics emanating from the region of interest. In contrast, the embodiments described herein measure dynamic strain using interferometry resulting from laser light reflected by FBGs 114 that are added to the fiber 112 and that are designed to reflect significantly more of the light than is reflected as a result of Rayleigh scattering. This contrasts with an alternative use of FBGs 114 in which the center wavelengths of the FBGs 114 are monitored to detect any changes that may result to it in response to strain. In the depicted embodiments, groups of the FBGs 114 are located along the fiber 112. A typical FBG can have a reflectivity rating of between 0.1% and 5%. The use of FBG-based interferometry to measure dynamic strain offers several advantages over DAS, in terms of optical performance.
In various embodiments herein in which the event detection system comprises a pipeline leak detection system, either FBG-based sensors or DAS-based sensors as described above may be used to generate a data stream of sensor readings. Simulated events are blended into that data stream for processing by the event detection system, as described below.
More particularly and with reference to
In
Referring now to
Analogously, simulated data 206 is recorded at block 302. In this example embodiment, the simulated data 206 is recorded from a field-simulated pipeline leak; in at least some other embodiments as discussed below, the simulated data may be generated using an artificial neural network such as a generative adversarial network. As another example, the simulated data 206 may be based on an actual authentic event, such as a pipeline leak, and used as the basis for one or more simulated events for the same event detection system 204 used to obtain the event or for other event detection systems 204. Analogous to block 308, the simulated data 206 is converted into the frequency domain at block 304 by, for example, determining its Fourier transform.
The frequency domain representations of the simulated data 206 and authentic raw data 205 are added together to result in the blended data 208 (not depicted in
At block 310 the blended data 208 is converted from the frequency domain to the time domain. This may be done by determining the inverse Fourier transform of the blended data 208. The resulting data is a time domain representation of the blended data 208 comprising the ambient pipeline acoustics and the simulated leak.
The data may then be processed using the event detection system 204 at block 312. The event detection system 204 may perform event analysis such as feature extraction and classification by applying, for example, machine learning.
In
The method of
Blocks 302 and 306 of
The computer system 1200 also comprises a database 402 that is communicatively coupled to the processor 1202 via the network interface 1216. As shown in
While various types of leak files 404a-c are shown in
The database 402 in
More particularly, the real time PSD data 502 in
The architecture of
Alternatively, the simulated event may be generated in PSD form as opposed to converted to PSD form. In this case, the simulated event is represented in the simulated PSD data 504, which is combined with the real time PSD data 502 that is the PSD representation of the authentic raw data 205. This real time PSD data 502 and the simulated PSD data 504 are combined together to result in the blended PSD data 506, which is fed to the PSD classifier 508. A reference to combining two PSD images together herein refers to summing the frequency domain data that is used to generate the PSD images analogous to how frequency domain data is combined in respect of
Images depicting simulated PSD events (“simulated PSD images”) may be generated using a suitable generative adversarial network (“GAN”), such as that depicted in
Blocks 802-814 of
Blocks 912-914 of
Generally speaking, when applying a DCGAN in at least some example embodiments Batch Norm is used in both the generator 604 and discriminator 608; fully connected hidden layers may be removed for a deeper architecture; Leaky ReLU activation may be used in the generator 604 for all layers except the output, which uses Tan h; and Leaky ReLU activation may be used in all layers of the discriminator 608.
When training a GAN in at least some example embodiments, the generator 604 may be trained to maximize the final classification error between authentic and simulated PSD images, while the discriminator 608 is trained to reduce that error. The generator 604 reaches equilibrium when the generated data 606 produces samples that follow the probability distribution of the authentic PSD images stored in the database 602 and used to train the discriminator 608. The discriminator 608 is trained directly on the authentic PSD images in the database 602 and on simulated PSD images generated by the generator 604, while the generator 604 is trained via the discriminator 608. Convergence of the generator 604 and discriminator 608 signals the end of training. When the discriminator 608 is being trained, the loss 612 returned to the generator 604 is ignored and only the discriminator 608 loss is used; this penalizes the discriminator 608 for misclassifying authentic leaks as simulated or simulated leaks as authentic (the generator's 604 weights are not updated during discriminator 608 training). When the generator 604 is being trained, the loss 612 returned to the generator 604 is used, which penalizes the generator 604 for failing to fool the discriminator 608 using the simulated PSD images it generates.
The various embodiments of the GAN depicted in
The processor 1202 may comprise any suitable processing unit such as a processor, microprocessor, artificial intelligence accelerator, or programmable logic controller, or a microcontroller (which comprises both a processing unit and a non-transitory computer readable medium), or system-on-a-chip (SoC). Examples of computer readable media that are non-transitory include disc-based media such as CD-ROMs and DVDs, magnetic media such as hard drives and other forms of magnetic disk storage, semiconductor based media such as flash media, random access memory (including DRAM and SRAM), and read only memory. As an alternative to an implementation that relies on processor-executed computer program code, a hardware-based implementation may be used. For example, an application-specific integrated circuit (ASIC), field programmable gate array (FPGA), or other suitable type of hardware implementation may be used as an alternative to or to supplement an implementation that relies primarily on a processor executing computer program code stored on a computer medium.
The embodiments have been described above with reference to flow, sequence, and block diagrams of methods, apparatuses, systems, and computer program products. In this regard, the depicted flow, sequence, and block diagrams illustrate the architecture, functionality, and operation of implementations of various embodiments. For instance, each block of the flow and block diagrams and operation in the sequence diagrams may represent a module, segment, or portion of code, which comprises one or more executable instructions for implementing the specified action(s). In some alternative embodiments, the action(s) noted in that block or operation may occur out of the order noted in those figures. For example, two blocks or operations shown in succession may, in some embodiments, be executed substantially concurrently, or the blocks or operations may sometimes be executed in the reverse order, depending upon the functionality involved. Some specific examples of the foregoing have been noted above but those noted examples are not necessarily the only examples. Each block of the flow and block diagrams and operation of the sequence diagrams, and combinations of those blocks and operations, may be implemented by special purpose hardware-based systems that perform the specified functions or acts, or combinations of special purpose hardware and computer instructions.
The terminology used herein is for the purpose of describing particular embodiments only and is not intended to be limiting. Accordingly, as used herein, the singular forms “a”, “an”, and “the” are intended to include the plural forms as well, unless the context clearly indicates otherwise (e.g., a reference in the claims to “a challenge” or “the challenge” does not exclude embodiments in which multiple challenges are used). It will be further understood that the terms “comprises” and “comprising”, when used in this specification, specify the presence of one or more stated features, integers, steps, operations, elements, and components, but do not preclude the presence or addition of one or more other features, integers, steps, operations, elements, components, and groups. Directional terms such as “top”, “bottom”, “upwards”, “downwards”, “vertically”, and “laterally” are used in the following description for the purpose of providing relative reference only, and are not intended to suggest any limitations on how any article is to be positioned during use, or to be mounted in an assembly or relative to an environment. Additionally, the term “connect” and variants of it such as “connected”, “connects”, and “connecting” as used in this description are intended to include indirect and direct connections unless otherwise indicated. For example, if a first device is connected to a second device, that coupling may be through a direct connection or through an indirect connection via other devices and connections. Similarly, if the first device is communicatively connected to the second device, communication may be through a direct connection or through an indirect connection via other devices and connections. The term “and/or” as used herein in conjunction with a list means any one or more items from that list. For example, “A, B, and/or C” means “any one or more of A, B, and C”.
It is contemplated that any part of any aspect or embodiment discussed in this specification can be implemented or combined with any part of any other aspect or embodiment discussed in this specification.
The scope of the claims should not be limited by the embodiments set forth in the above examples, but should be given the broadest interpretation consistent with the description as a whole.
It should be recognized that features and aspects of the various examples provided above can be combined into further examples that also fall within the scope of the present disclosure. In addition, the figures are not to scale and may have size and shape exaggerated for illustrative purposes.
This application is related to and claims priority to U.S. Provisional Patent Application No. 63/288,763 filed on Dec. 13, 2021, the contents of which are incorporated by reference herein.
Number | Date | Country | |
---|---|---|---|
63288763 | Dec 2021 | US |