The present invention is generally directed to a method and system for analyzing a large volume of complex signal-vector data for pattern recognition, particularly from first order sensors.
Most conventional sensors for the detection of chemicals and/or fire are electrochemical sensors. Electrochemical sensors operate at a fixed potential. Changes in that fixed potential, caused by the presence of an analyte, signal the analytes presence. Most often these sensors are directed towards the recognition of a single analyte and utilize a single sensor or a simple array of sensors.
Current concerns about homeland security, battlefield protection and terrorist activities have created an interest in sensors that detect a variety of chemicals, such as a variety of blood agents, toxic industrial chemicals, explosives, etc, preferably with specificity for the identification of type and quantity of chemicals detected.
Also, an interest has developed in fire prevention and fire detection. Along with fire detection, early detection of various gases or of precursors of fires, before the presence of flame and smoke, which are necessary for detection using conventional smoke detection/fire detection devices, is also desirable. For example, with increased automation on ships, the Navy has sought fire detection systems capable of improved performance over conventional smoke detectors. To the extent such precursors can be identified, fire prevention can occur in advance of mere fire detection to avoid damage to expensive infrastructure.
Although the use of multiple conventional electrochemical sensors monitoring multiple criteria has been studied for this type of chemical identification with varying degree of success, most of the sensors did not have the specificity desired. For example, in fire detection, multiple criteria electrochemical sensors have difficulty discriminating fire-like nuisance sources, such as welding, grinding steel, and cutting with a torch. Further, multiple criteria electrochemical sensors are typically limited to only a few standardized test sources, without providing the range of detection capability often desired.
Cermet sensors are well known in the art that use various ceramic metallic (cermet) films. See U.S. Pat. No. 6,218,687, which is incorporated herein by reference in its entirety. Rather than operating based on a single potential, some cermet sensors characterize data using cyclic voltammetry, in which current is measured while manipulating the potential in a cyclical fashion. Thus, the data provided by such sensors is provided in a cyclical waveform.
Cermet sensor arrays have been used for more realistic and difficult conditions than prior electrochemical sensors and have been used to identify chemical analytes and fires, particularly distinguishing nuisance sources from harmful events. Such cermet sensors are capable of high temperature operation and can be fabricated using both thick and thin film techniques. Cermet sensors are small, lightweight and low cost alternatives to conventional electrochemical sensors. Plus, cermet sensors have design flexibility since different cermet sensors may respond to the same gas in different ways. Using conventional methods for data analysis of cyclic voltammetry, cermet sensors are capable of detecting very low level concentrations (part per billion) for a variety of analytes using aqueous electrolytes, but only higher levels (parts per million) for gas phase detection.
However, the use of cyclic voltammetry creates complex data waveforms having a great deal of information, such as peak position, magnitude, and shape, all of which can be exploited for analytical purposes. Such waveforms contain data that incorporates variations in the presence, concentration, temperature, accessibility to the sensor, function of the sensor etc. which must be selectively analyzed before any meaningful information about the identity and/or quantity of a certain analyte or fire/non-fire condition can be ascertained.
Previously pattern recognition software has been used to analyze waveform data. However, the shear volume of data generated by cyclic voltammetry has caused concern about accuracy of probable identification of analyte sources, particular in the presence of quite similar nuisance sources. Further, the volume of data required may be more than a microprocessor embedded within a portable device used for field applications is capable of processing in a fast, real-time manner.
Wavelet analysis of analytical data prior to pattern recognition has been used in a variety of different applications. For example, wavelet analysis has been used in medicine to specifically look at identifying conditions from time-dependant data such as EEG and ECG scans and in identifying features of interest in medical imagery. Non-medical applications also included image analysis, as well as optical character recognition, and acoustic pattern recognition.
In sensing applications, wavelet analysis of linear voltammetric data of analytical chemical data, has been demonstrated as a means of de-noising linear voltammetric data. An “electric tongue” system has been developed for monitoring industrial liquid process streams. They utilize a “Wavelet Neural Network” as a pattern recognition technique which incorporates wavelet transformation into a feed forward multilayer neural network architecture. The parameters of the wavelet transform are thus modified by the error propagation structure of the neural network during training. Wavelet packet transforms have also been used for pattern recognition. The wavelet packet transforms have included the use of a Euclidean distance measure to generate the best wavelet basis set of an analytical signal for classification purposes. This approach has been generally applied only to spectroscopic and chromatographic data. Wavelet transformation has also been used for semi-conductor based gas sensors, particularly the use of discrete wavelet transform as a data pretreatment step prior to pattern recognition via a number of different techniques. Wavelet transform followed by classification by a PNN has been done by HPLC-DAD data in the classification of Chilean wines, peak detection in LC-TOF MS data and to capture features in NIR data.
In these various known data analysis techniques, the wavelet-transformed data are not down-selected or narrowed based on any particular classifying ability. As such, the data continues to include random variables and data points which are not necessary for classification purposes. The inclusion of unnecessary data points increases processing time and creates additional variables which can unnecessarily complicate or interfere with pattern recognition. In some cases, the wavelet coefficients may be selected based upon mere visual inspection. However, it is particularly difficult to discern narrow differences by mere visual inspection. In other cases, another “quick and dirty” method has been used to reduces the wavelet data, but not by any particular method particularly suitable for classification purposes. These “quick and dirty” methods do not ensure selection of the most desirable features of a wavelet, which can lead to lack of accuracy in pattern recognition.
The present application is generally directed to a method for characterizing signal-vector data using automatic feature selection techniques on wavelet-transformed data to enhance the use of pattern recognition techniques for classification purposes. Automatic feature selection has the advantage of extracting only the relevant information from wavelet-transformed data for improved classification of signal-vector data, which leads to increased effectiveness in classification of first order sensor data. By reducing data to only relevant information, data processing times and microprocessor memory requirements are reduced, which results in better response times for first order sensors, as well as an increase in accuracy of sensing applications. For example, in chemical sensing application, the improved classification creates better sensitivity and increased recognition of unknown analytes in real-time, and, in fire detection, improved classification enhances fire detection by better distinguishing fires from nuisance sources.
An embodiment of the present invention is a method of analyzing signal-vector data from at least one first order sensor, which includes providing a training data set that includes signal-vector data from at least one first order sensor, wherein the training data set includes background data and sample data collected under known conditions. The sample data of the training data set is then adjusted to account for variations between the background data and the sample data using a background adjustment technique. The method then includes normalizing the training data set and transforming the normalized training data set into wavelet coefficients using a wavelet transformation technique. Using an automated analysis of variance feature selection technique, desirable wavelet coefficients for classification are then automatically selected. Then, a pattern recognition technique is used to classify the training data set. A classification model generated from the pattern recognition technique is then output.
This embodiment may also include a prediction step which includes providing an unknown sample data set that includes signal-vector data from a first order sensor, wherein the sample data set includes unknown sample data collected under unknown conditions. The unknown sample data of the unknown sample data set is adjusted to account for variations between the background data and the unknown sample data using the same method used for the training data set. The unknown sample data set is normalized using the same method used for the training data set, and the normalized unknown sample data set is transformed into wavelet coefficients using the same wavelet transformation technique used for the training data set. The method also includes automatically selecting the same desirable wavelet coefficients selected from the wavelet-transformed data of the training data set in the wavelet-transformed data of the unknown sample data set. The same pattern recognition technique for the training data set is used to compare the unknown sample data set to the classification model. An output is provided including an identity of the unknown conditions associated with the unknown sample data set.
Another embodiment of the present invention is a computer system that includes a training data set comprising signal-vector data, including background data and sample data provided by at least one first order sensor under known conditions and a processor that receives the training data set and memory capable of storing a classification model generated from a pattern recognition technique. The processor includes software suitable for each of the following applications: adjusting the sample data to account for variances between the background data and the sample data, normalizing the training data set, transforming the normalized training data set into wavelet-transformed data using a wavelet transformation technique, automatically selecting desirable wavelet coefficients for classification using analysis of variance feature selection technique and classifying the training data set using the pattern recognition technique.
This embodiment may also include an unknown sample data set comprising signal-vector data, including unknown sample data provided by the at least one first order sensor under unknown conditions. Also, the processor may further included software suitable for each of the following applications: adjusting the unknown sample data to account for variances between the background data and the unknown sample data in the same manner as the training data set, normalizing the unknown sample data set in the same manner as the training data set, transforming the normalized unknown sample data set into wavelet coefficients using the wavelet transformation technique used for as the training data set, automatically selecting the same desirable wavelet coefficients from the unknown sample data set as were selected for the training data set; and comparing the wavelet coefficients from the unknown sample data set with the classification model using the pattern recognition technique. The system may also include an output device for outputting an identity of the unknown conditions associated with the unknown sample data set.
Another embodiment of the present invention is a sensing system that includes at least one first order sensor, wherein the sensor senses the presence of a condition and produces signal-vector data, an unknown sample data set comprising signal-vector data, including unknown sample data provided by the at least one sensor under unknown conditions, a microprocessor connected with the at least one first order sensor and an outputting device for outputting an identity of the unknown conditions associated with the unknown sample data set. The microprocessor includes a stored classification model previously compiled from a training data set previously collected and software. The software is suitable for each of the following applications: adjusting the unknown sample data to account for variances between background data and the unknown sample data, normalizing the unknown sample data set, transforming the normalized unknown sample data set into wavelet coefficients using a wavelet transformation technique used for the training data set, selecting the same desirable wavelet coefficients from the unknown sample data set as were selected for the training data set using an analysis of variance feature selection technique; and comparing the wavelet coefficients from the unknown sample data set with the classification model using a pattern recognition technique that generated the classification model.
The foregoing and other features and advantages of the present invention will be apparent from the following, more particular description of a preferred embodiment of the invention, as illustrated in the accompanying drawings.
Preferred embodiments of the present invention are now described with reference to the Figures, in which like reference numerals are generally used to indicate identical or functionally similar elements. Also in the Figures, the left most digit of each reference numeral generally corresponds to the Figure in which the reference numeral appears. While specific details of the preferred embodiments are discussed, it should be understood that this is done for illustrative purposes only. A person skilled in the relevant art will recognize that other configurations and arrangements can be used without departing from the spirit and scope of the invention. It will also be apparent to a person skilled in the relevant art that this invention can also be employed in other applications.
Cermet sensors employ an electrochemical (particularly voltammetric) measurement technique to generate their complex response signal-vector data, generally in the form of cyclic waveforms. Voltammetry involves applying a varying potential (typically a ±triangular waveform) across an electrochemical cell and measuring the resultant current produced. This technique is commonly referred to as “cyclic voltammetry.” Other types of sensors or other devices may also use cyclic voltammetry or may utilize other similar methods for producing an output in the form of cyclic waveforms. The cyclical waveform is an example of signal-vector data, wherein the signal data is generated as an information-rich vector rather than merely a data point. The cermet sensor, therefore, is an example of a first order sensor, which is capable of producing signal-vector data.
The present invention is generally directed to a method for analyzing signal-vector data from first order sensors and a sensor system which analyzes signal-vector data via this method. Thus, a cermet sensor is merely a non-limiting example of a first order sensor and a cyclic waveform is merely an example of signal-vector data that may be available therefrom.
In particular, the present invention focuses on using information rich wavelet that are transformed from the signal-vector data, but reducing the wavelet to only the most relevant wavelet coefficients via an automated feature selection technique that is particularly suited for classification purposes. Automated feature selection has an advantage over any other wavelet analysis in that it selects only the most relevant information, increasing accuracy of classification via pattern recognition, which reducing the time and microprocessing requirements needed for the rigorous data analysis of signal-vector data. Thus, smaller and more lightweight field sensors can be made with increased accuracy and reducing microprocessing requirements, which is a particular benefit for sensors used, for example, in public areas or battlefield environments.
Sensor arrays offer even greater selectivity than single sensors, such that a plurality of sensors is available to sense a particular environment under the same conditions. The use of a plurality of sensors, for example a plurality of cermet sensors with different electrode-electrolyte-electrode configurations, provides different sensing criteria. The combination of data from each sensor produces a signature or fingerprint that is unique for exposure to each of a plurality of conditions. For example, they may provide a signature for each of several different analytes and/or different fire conditions as they are detected by each of the plurality of sensors.
When a plurality of different cermet sensors are utilized in a sensor array to sense different conditions, an even greater volume of signal-vector data is developed. Thus, signal-vector data can include a large number of data points, which must be distinguished based on the particular condition.
Thus, the present invention is generally directed to a method and system for analyzing the large volume of signal-vector data that is generated from first order sensors, for the particular purpose of classifying the conditions under which the signal-vector data was generated. Generally, this method includes providing a training data set generated from known conditions. For sensing applications, this training data set generally includes both baseline data, or background data, and sample data from a specifically known source under known conditions. In order to characterize only the actual variations of the sample data over the background data, the training data set must first be adjusted. The most simple background adjustment technique is merely subtracting the baseline data from the sample data. Other non-limiting examples of background adjusting techniques include, but are not limited to, baseline adjustment, calculating a slope to remove overall baseline drift and rigorous dynamic background normalization to account for particular locations of substantial baseline drift.
Additionally, the data should be normalized using known normalization techniques. One non-limiting example of a normalization technique is the most common: using the square root of the sum of the squares. For certain sensing applications, these normalization techniques remove certain known variables from consideration, such as concentration and temperature changes, so as to focus the data on merely the presence or absence of a specific material or condition.
Wavelet transformation techniques are then applied to the training data set, which may be presented in as spectra that has been extended to a dyadic length. Various known wavelet transformation techniques may be used, including but not limited to discrete wavelet transformation, particularly a fast wavelet transformation. Daubechies, Haar, Symmlet, and Coiflet are non-limiting examples of wavelets suitable for use in the present invention. A preferred wavelet is the Daubechies 8 technique, which is well known. Wavelet coefficients may not be visually interpretable to an analyst but are an efficient way to express the data in a compressed form that is amenable to mathematical data analysis. The wavelet transform extracts information about signal-vector data within an output response that varies in frequency and is localized in time. In other words, the wavelet coefficients provide information in a different format which is particularly suitable for feature selection and pattern recognition.
A feature selection technique is performed on the wavelet-transformed data, in which particularly significant or relevant wavelet coefficients are selected. The feature selection technique of the present invention is an automated feature selection technique. While several non-automated feature selection techniques may be known, automated feature selection has the benefit of reducing human error in the selection of the most relevant features of wavelet-transformed data, particularly for classification purposes. The preferred automatic feature selection technique of the present invention is the application of analysis of variance (ANOVA) combined with cross-validation for parameter optimization and to minimize error. Cross-validation automatically determines and optimal Fisher-ratio or “f-ratio” threshold value with which the wavelet coefficients are selected.
The training data set then may be used to form a classification model using classifying or pattern recognition techniques. Examples of pattern recognition techniques include but are not limited to linear discriminant analysis, nearest neighbor analysis or the application of a neural network. Preferably, a probabilistic neural network, such as that described in U.S. Pat. Nos. 6,289,328 and 7,034,701, which are incorporated herein by reference in their entirety, is used such that unknown conditions are not only classified, but classified with a level of probability of accuracy. Using a probabilistic neural network, the data is classified by computing the distances of each data point in the model to each data point of the unknown sample data set, also taking into account a probability density function that is superimposed upon each sample data set.
The classification model may be output by a variety of methods, including but not limited to, displaying, printing, storing for further retrieval or processing, making available for download, and/or transmitting the results generated by the equations, the system and methods discussed herein. The present invention also includes any computer systems and computer readable media having programs or applications for receiving, displaying, printing further processing, and transmitting the classification model or information derived therefrom, as well as for making the results available for retrieval by other users.
Via the feature selection techniques of the present invention, large volume of signal-vector data can be analyzed on any computer system, including a small microprocessor, suitable for use in a small a lightweight field sensor device. Without this feature selection technique applied to the wavelet-transformed data, the large volume of the training data, if required to be analyzed in the number of dimensions that would be required with such a condition specific data set, may be too much data for a field microprocessor and instead require the full capabilities of a conventional desktop PC, which would not be suitable for hand-held or battlefield sensing applications.
The same method is also applied to real-time sensing operations. An unknown sample data set can be provided in which the conditions are unknown. The unknown sample data set includes the same signal-vector data from the same first order sensor. The unknown sample data can then be correlated with the training data set so as to identify the unknown conditions. For example, an unknown sample data set may be generated that represents exposure to an unknown analyte. As another example, the unknown sample data set may represent an unknown fire/fire precursor/non-fire condition that is sensed in real-time.
The signal-vector data of the unknown sample data set is treated in the same way as the training data, including background adjustment, normalization and wavelet transformation, using the same techniques applied to the training data set. Then, the same wavelet features that were selected for the training data set are also selected for the unknown sample data set using ANOVA. The unknown sample data set is then compared to the classification model from the training data set via the classifying or pattern recognition technique. The pattern recognition technique then is used to identify the unknown conditions under which the unknown sample data set was prepared. For example, the pattern recognition technique would identify the presence of a particular analyte or indicate whether or not a fire or fire pre-cursor was occurring or not. Again, the probabilistic neural network is the preferred classification or pattern recognition technique, since it provides not only an identity of the unknown condition, but also the probability that the identified condition is accurate.
As with the classification model, the identification of the unknown condition and/or the probability of accuracy of the identified unknown condition may be output by a variety of methods, including but not limited to, displaying, printing, storing for further retrieval or processing, making available for download, and/or transmitting the results generated by the equations, the system and methods discussed herein. The present invention also includes any computer systems and computer readable media having programs or applications for receiving, displaying, printing further processing, and transmitting the identification and/or probability of accuracy of the unknown condition or information derived therefrom, as well as for making the results available for retrieval by other users.
For example, in a real-time sensing application, it may be useful to have an audible or visual alarm when results dictate the presence and/or high probability of a harmful substance or fire. Where probability information exists, the alarm may only occur when the probability exceeds a predetermined percentage. For example, if the probability of a harmful event is greater than 50%, it is more likely than not that the event is occurring. However, for increased certainty, an alarm could be set to active only when higher probabilities, such as 60% or even 90% is exceeded. Knowing the level of certainty allows the analyst to set any probability threshold designed to lower the rate of false alarms.
Examples 1 and 2 below utilize an array of four conventional cermet sensors which have different electrode-electrolyte-electrode configurations to provide slightly different sensing criteria. For example, in Example 1, sensor A is platinum-yttria stabilized zirconia-platinum/palladium (Pt—YSZ—Pt/Pd), sensor B is platinum-yttria stabilized zirconia-platinum (Pt—YSZ—Pt), sensor C is platinum-yttria stabilized zirconia-platinum-tungsten bismuth oxide (Pt—YSZ—Pt—WBO), and sensor D is platinum-yttria stabilized zirconia-platinum/palladium-tungsten bismuth oxide (Pt—YSZ—Pt/Pd—WBO). However, one skilled in the art can appreciate that data from only one sensor or from any plurality of sensors may be analyzed according to the present invention. Further, any data exhibiting such a cyclic waveform under various conditions should be suitable for analysis according to the present invention.
The following examples further describe the system and method of the present invention. In each Example, there generally exists a training step, which includes the collection of a training data set (for example, sensor responses to clean air and known analyte gases or sensors responses under known fire or non-fire conditions). The training data set is then used to generate a classification model. There also exists a prediction step, which includes a measurement of an unknown sample data set collected under unknown conditions (for example, an unknown gas sample is taken or the potential fire conditions of a room are not known). The unknown sample data set is compared via the pattern recognition technique, from which a classification model has been “trained” using the training data set, in order to predict the classification of the unknown sample data set.
In this example, the four sensors are exposed to clean air and then to a variety of analytes of interest. Analyte streams were produced by diluting certified gas cylinders of carbon monoxide (CO), nitrogen dioxide (NO2), sulfur dioxide (SO2), hydrogen sulfide (H2S), chlorine (Cl2), hydrogen chloride (HCl) (Air Gas, Los Angeles, Calif.) ammonia (NH3), carbon disulfide (CS2), benzene (C6H6), formaldehyde (HCHO), hydrogen cyanide (HCN) (Tex-La Gas, Sulphur, La.), and cyanogen chloride (CNCl) (HP Gas Products, Baytown, Tex.). Analyte streams of the nerve agent simulants dimethyl methylphosphonate (DMMP), diisopropyl methylphosphonate (DIMP), and 2-chloroethyl ethyl sulfide (CEES) were generated by certified permeation tubes (KIN-TEK, La Marque, Tex.). Various controllers were used such that the flow, temperature, and humidity of the purified air and analyte/air streams to create the known concentration of analyte needed under these known conditions. The sensors were exposed to each of analytes at five different concentrations from low to high, high to low, and low to high, making three replicates of each concentration. A training data set was also collected for two analytes at various temperatures and at various scan rates. The sensors were flash heated to 300° C. between experiments to help clear out any remaining analyte from the sensor and restore it to a clean background condition.
Data is collected during the experiment at a preset data acquisition rate, and with preset cermet sensor parameters (voltage limits, voltage sweep frequency, operating temperature, etc.). This data is collected in the form of one vector sensor current readings per sensing element per voltage sweep cycle, as an example of signal-vector data of the present invention. For the purpose of data analysis, the signal-vector data for each of the four sensing elements are joined end to end. Such an augmented spectrum is collected for each voltage sweep cycle, and these form the rows of the data matrix that is eventually acquired over the duration of the experiment. This data is saved a text file on a workstation PC. However, the data could be saved in a variety of ways on a variety of computer systems or memory devices.
The training data set was collected at 2 mV intervals resulting in at least about 3000 points for each cyclic waveform and greater than about 12,000 points for cyclic waveforms from all four sensors. For each 5 min cycle, about 20 cyclic waveforms were collected with about the last 10 cyclic waveforms of each air/analyte run used in the analysis to allow for a complete exchange in the chamber and the sensor to reach a steady state. A matrix of about 10×12,000 was used from each initial clean air exposure of each analyte to model the background.
In particular, a row of the data matrix (or, alternatively, an average of multiple rows) corresponding to a measurement taken from clean air is subtracted from each remaining row of the training data set matrix, forming a new matrix of background-subtracted data. Following this, the data is then normalized on a row-wise basis using the square root of the sum of the squares of each row as the normalizing factor (i.e., the Euclidean norm). Normalization makes the patterns observed for each analyte gas mostly invariant with respect to concentration.
The data is imported into a MATLAB program, specifically MATLAB, version 7.0 (Mathworks, Inc., Natick, Mass.). MATLAB is a computational environment that can be facilitated on a conventional workstation PC or microprocessor. When the training data set was background subtracted about a 30×12,000 matrix for the three replicates of each concentration was generated. A data matrix of about 2400×12,000 were generated using the data collected (150 air, 150×15 analyte).
Principal component analysis was used merely to provide an interpretable view (or graphical display) of the multidimensional data space. However, principal component analysis was also used for comparison purposes to demonstrate the advantage of the present invention over other data analysis techniques, as discussed in further detail below. The MATLAB routines used for principal component analysis were provided in the PLS Toolbox, version 3.0.4 (Eigenvector Technologies, Inc., Manson, Wash.).
Following normalization, wavelet transformation was also used. A wavelet transformation technique is row-wise applied to the normalized training data set. The wavelet transformation technique used in this example is discrete wavelet transformation. The discrete wavelet transformation is a fast wavelet transformation. In this case, “Daubechies 8” was the preferred wavelet. However, other wavelet transformation techniques and/or other wavelets may be selected for use in the present invention, as would be apparent to one skilled in the art. The particular choice of wavelet transformation technique to use will depend on the signal characteristics of the measured data. The Daubechies wavelet is commonly used in similar applications and is well suited to describing the broad features located in the spectra output by a cermet microsensor array. However, as previously discussed, prior to transformation, the spectra from each sensing element are separated and extended to dyadic length as required by the fast wavelet transform algorithm of Daubechies 8. As such, the waveform was reduced from 12,000 to 1024 points. After wavelet transformation, a wavelet-transformed training data matrix is developed of the same size as the normalized training data matrix, but with wavelet coefficients occupying the rows rather than time-dependant current measurements. For an example of the effects of wavelet transformation see
Algorithm implementations utilized for wavelet transformation of the training data set are taken from the WaveLab toolbox for MATLAB (particularly WaveLab802), which is available from Stanford University (http://www-stat.stanford.edu/˜wavelab/). Wavelet transformation of cermet microsensor array data provides two principal advantages: data compression and the ability to apply enhanced feature selection. A wavelet transformation takes data from a time domain to a scale-dependant frequency domain. Varying this scale allows for a series of wavelet coefficients to describe frequency-based features in the data that are also localized at specific times.
Further data compression is achieved by selectively filtering out coefficients from the wavelet transform that contain little or no signal. Feature selection is achieved by locating coefficients that contain information relevant to a desired classification. The wavelet coefficients are automatically processed in order to select for relevant features. Coefficients that are less than an optimized threshold in every spectrum are automatically discarded. In this example, an analysis of variance (ANOVA) calculation is performed column-wise on the wavelet coefficients in order to locate coefficients that are effective at discriminating between different analyte gases.
ANOVA examines the inter- and intra-class variance between the data for the various analytes. As the training data set is generated using known gas analytes, it is known which analyte gas corresponds to each row of the wavelet-transformed data matrix. The value calculated by the ANOVA equation is a Fischer-ratio or an “f-ratio”, which is the ratio of the variance observed over the measurements of different analytes to the variance observed in the measurement of only one analyte. The magnitude of the f-ratio provides an indicator of class separation for each calculated wavelet coefficient over the wavelet-transformed data matrix. Data that produce an f-ratio greater than a set threshold are retained as features while the remaining data are discarded. The threshold selected is one that provides the greatest between class variance when compared to the variance within a class. This threshold is determined through cross-validation to maximize correct classification by optimizing the classifier's performance. The set of wavelet coefficients thus chosen comprises the set of selected features to be used in subsequent classification of unknown gas samples. As one mode of output, the waveform data can then be reconstructed using the selected coefficients.
Using ANOVA, the wavelet-transformed training data set was then reduced to 18 wavelet coefficients.
The MATLAB script for ANOVA calculation is a custom implementation by personnel at the Naval Research Laboratory. Attached at the end of this specification is the particular MATLAB script used for the ANOVA calculations in the examples herein.
Next, the known analyte feature-selected wavelet coefficients are used to build a classification model. The function of the classifier is to provide an optimized decision regarding sample classification when presented with new data. For comparison purposes, two different types of classifiers were used in this example, a known probabilistic neural network (PNN) developed at the Naval Research Laboratory and linear discriminant analysis (LDA) based on the Mahalanobis distance metric, both with good results. Other types of classifiers or pattern recognition techniques may be utilized, including, but not limited to, conventional nearest neighbor analysis.
Generally, supervised training classification techniques use training data sets to develop classification rules. The rules predict classification of future unknown sample data sets. Given both the unknown sample data sets and the classification model results, the classifier, or pattern recognition technique, generates mathematical functions to define the classes. Contiguous block cross-validation, testing all the replicates of each concentration level, was used for classification analysis. In this example, all patterns at a given concentration are withheld from the training data set being classified.
An unknown sample data set is generated in the same manner as the training data set. The unknown sample data set is normalized in the same row-wise fashion as the training data set (i.e., the unknown sample data set is background subtracted and then normalized by the Euclidean norm). The same wavelet transformation technique as used previously for the training data set is row-wise applied to the unknown sample data set. The same feature-selected wavelet coefficients are extracted as for the training data set. These feature-selected wavelet coefficients are then presented to the classification model generated from the training data set for sample analyte identification.
The results from Example 1 demonstrate the improvement over other method for data analysis conventionally used, which do not incorporate wavelet transformation and feature selection techniques. For example, Table 1 shows the results using the probabilistic neural network pattern recognition technique, where principal component analysis was used for data reduction rather than wavelet transformation. Table 2 shows the results using the linear discriminant analysis pattern recognition technique, where principal component analysis was used for data reduction rather than wavelet transformation. The overall correct classification for all five concentrations is 88/90%, respectively. The poorest performances are observed for the initial run of the lowest concentrations through the sample chamber.
Meanwhile, Table 3 shows the results using the probabilistic neural network pattern recognition technique, where wavelet transformation and feature selection techniques were used. Table 4 shows the results using the linear discriminant analysis pattern recognition technique, where wavelet transformation and feature selection techniques were used. The overall classification for all five concentrations is 94/93%, respectively, which is a significant improvement over the methods of Tables 1 and 2, with probabilistic neural network demonstrating slight improvement over linear discriminant analysis.
Improvements have been demonstrated in extracting the relevant information for selection identification of a variety of analytes with the data analysis method of the present invention. The use of wavelet transformation and feature selection techniques provided improved classification. All of the vapors of interest are readily detected at the threshold limit value, which indicates when safe levels of exposure to various hazards have been exceeded. These data analysis methods show much promise for selective detection of a wide variety of vapors, particularly as improvements continue to be made in the cermet sensors themselves.
Instead of sensing analytes for the detection of particular chemicals, cermet sensors were also used to detect fire or pre-cursors of a fire under various conditions using the data analysis methods of the previous invention as demonstrated by Example 1.
In both laboratory and shipboard environments, fire and nuisance sources were created to expose the detection systems to a range of potential shipboard scenarios. For example, the following fire sources were used in this example either in the laboratory environment, in the shipboard environment or in both environments: flaming cardboard box (with and without polystyrene pellets therein), flaming trash can, flaming shipping supplies, flaming mattress and bedding, flaming IPA spill, flaming heptane pan, flaming diesel-soaked rags, smoldering bag of trash, smoldering mattress and bedding, smoldering cable bundle, smoldering laundry, smoldering oily rags, heated painted bulkhead, IPA pan in the interior of a cabinet and smoldering computer monitor. The following nuisance sources were used in this example either in the laboratory environment, in the shipboard environment or in both environments: torch cut steel, cutting steel, welding, grinding painted steel, normal toasting of bread at the darkest setting, engine exhaust, grinding cinder block, spray aerosol, spilling metal bolts, space heater and heat gun. Thus, a training data set was generated for each of the fire sources and nuisance sources under a variety of controlled conditions. Small fires were used to challenge the detection systems and to provide performance results for early detection.
The same array of four sensors was used in the laboratory environment. However, in the shipboard environment, Sensor C and Sensor D were replaced with platinum-tungsten bismuth oxide-yttria stabilized zirconia-platinum (Pt—WBO—YSZ—Pt) and platinum-tungsten bismuth oxide-yttria stabilized zirconia-platinum/palladium (Pt—WBO—YSZ—Pt/Pd) sensors, respectively. The use of different sensors in the various environments illustrates that the type or number of sensors utilized is not critical to the data analysis performed by the methods of the present invention. In fact, the present invention is suitable for classifying signal-vector data from a variety of sensors.
The primary data acquisition system was started and allowed to collect background data for a minimum of 300 seconds. After the background data was collected, the fire or nuisance source was initiated and allowed to continue until fully consumed or until all systems were in alarm or showed no change in detection due to quasi-steady state conditions.
Waveform sample data sets, such as those shown in
In both the laboratory and shipboard environments, the training data set was designed to provide a range of fires sources, source locations, and ventilation conditions to comprehensively evaluate the detection systems against likely shipboard fire scenarios. The data was collected at 8 mV intervals resulting in 750 points for each voltammogram with a total of 3000 points for all four sensors.
The data was background subtracted using the five waveforms generated prior to ignition.
As with Example 1, a Daubechies 8 wavelet function was chosen for this example. As such, the waveform was expanded from 3000 to 4096 points.
An analysis of variance (ANOVA) was used to select down the total wavelet coefficients to only 64 wavelet coefficients. The training data set was divided into three classes: non-fires, flaming fires and smoldering fires. The fires were split into two categories so that differences between flaming and smoldering fires could be maintained in the data space.
The probabilistic neural network computes the probability that allows a level of certainty to be associated with the prediction.
For the laboratory environment, the waveform data from the first 47 tests, when the sensors were clustered together, was used to build a training data set. Waveform data from the next 58 tests were used as unknown sample data sets. From the unknown sample data sets a probability cutoff of 90% was seen as giving the best fire classification.
The cermet sensors were able to pick up most of the fires that were tested, especially the smoldering fires. Table 5 represents the percent of correct classification of each source type for the laboratory environment. Table 6 represents the percent correct classification of each system by source type for the shipboard environment. The percent correct classification represents the number of sources the system correctly classified. It should be noted that when calculating the percent of correctly classified nuisance alarms the number of times a given system did not alarm divided by the number of nuisance sources tested results in the percent correct classification. Overall the results show that data from the cermet sensors could be used to provide good results for classifying fire conditions using the methods of the present invention.
The data analysis methods were run in near real-time. The fires were monitored as they occurred by running the algorithm on the sample data sets as they were being saved. Thus, the method of the present invention may be incorporated into the data collection software to form a system for analyzing data for signal-vector data as it is generated. Further, the present invention may be utilized in real-time in various applications.
The training data set may be collected and the classification model prepared on a conventional computer or PC. The classification model could then be saved or uploaded to a microprocessing system associated with a sensor device incorporating the sensors as well as the software programs for the method of the present invention. Thus, only raw data from the unknown sample data sets need be loaded in the microprocessor for analysis, while the larger volume of raw data for the training data set could remain stored on the desktop computer or PC. Thus, the sensor device can be made very small and lightweight.
Instrumental drift associated with first order sensors or changes in the sampling environmental may cause future data to be sufficiently different from the training data set such that a new unknown same data may be misclassified or even lie entirely outside the classification model. Thus, the present invention contemplates the prudent (and often necessary) periodic update of the classification model with new training data sets. Since the generation of training data sets is generally costly in terms of time and effort, classification models may be updated by comparison of with only “clean air” or “background” data or with comparison of training data sets from only one source. For updating data associated with fire detection, the use of ammonia as a source could be used to calibrate and monitor the effectiveness of the sensors and training data sets, without having to reenact fire/non-fire conditions in order to update the training set data.
The present invention contemplates the use of computer software and computer programs, as well as computer systems and computer readable media having programs for implementing the methods discussed above. With reference to
It should be understood that the term “computer system” is intended to encompass any device having a processor or microprocessor that executes instructions from a memory medium. The memory medium preferably stores instructions (also known as a “software program”) for implementing various embodiments of a method in accordance with the present invention. In various embodiments the one or more software programs are implemented in various ways, including procedure-based techniques, component-based techniques, and/or object-oriented techniques, among others. Specific examples include but are not limited to FORTRAN, XML, C++ objects, Java and Microsoft Foundation Classes (MFC).
“Outputting” is contemplated to include methods of receiving, displaying, printing, storing for further retrieval or processing, making available for download, and/or transmitting the results generated by the equations, the system and methods discussed herein. The present invention further contemplates computer systems and computer readable media having programs, applications and/or output or communication devices for receiving, displaying, printing, further processing, and transmitting the results or information derived from the results, as well as for making the results available for retrieval by other users.
While the invention has been particularly shown and described with reference to preferred embodiments thereof, it will be understood by those skilled in the art that they have been presented by way of example only, and not limitation, and various changes in form and details can be made therein without departing from the spirit and scope of the invention. Thus, the breadth and scope of the present invention should not be limited by any of the above-described exemplary embodiments, but should be defined only in accordance with the following claims and their equivalents. Additionally, all references cited herein, including issued U.S. patents, or any other references, are each entirely incorporated by reference herein, including all data, tables, figures, and text presented in the cited references. Also, it is to be understood that the phraseology or terminology herein is for the purpose of description and not of limitation, such that the terminology or phraseology of the present specification is to be interpreted by the skilled artisan in light of the teachings and guidance presented herein, in combination with the knowledge of one of ordinary skill in the art.
The foregoing description of the specific embodiments will so fully reveal the general nature of the invention that others can, by applying knowledge within the skill of the art (including the contents of the references cited herein), readily modify and/or adapt for various applications such specific embodiments, without undue experimentation, without departing from the general concept of the present invention. Therefore, such adaptations and modifications are intended to be within the meaning and range of equivalents of the disclosed embodiments, based on the teaching and guidance presented herein.
The following is the ANOVA scripts used in MATLAB in the Examples herein:
This application is a non-provisional application claiming priority to U.S. Provisional Application No. 60/779,412, filed Mar. 7, 2006, which is incorporated herein by reference in its entirety.
Number | Name | Date | Kind |
---|---|---|---|
3652990 | Darwin | Mar 1972 | A |
4245258 | Holladay | Jan 1981 | A |
5121443 | Tomlinson | Jun 1992 | A |
5520881 | Koestler et al. | May 1996 | A |
5676811 | Makino et al. | Oct 1997 | A |
5772863 | Shoemaker et al. | Jun 1998 | A |
6166647 | Wong | Dec 2000 | A |
6218687 | Ziegler | Apr 2001 | B1 |
6266704 | Reed et al. | Jul 2001 | B1 |
6277256 | Garzon et al. | Aug 2001 | B1 |
6289328 | Shaffer | Sep 2001 | B2 |
6375816 | Jach et al. | Apr 2002 | B1 |
6524460 | Hachtel et al. | Feb 2003 | B1 |
6535562 | Mohseni et al. | Mar 2003 | B1 |
6539356 | Matsui et al. | Mar 2003 | B1 |
6578770 | Rosen | Jun 2003 | B1 |
6767442 | Scheer et al. | Jul 2004 | B1 |
6785401 | Walker et al. | Aug 2004 | B2 |
7034701 | Rose-Pehrsson et al. | Apr 2006 | B1 |
7316899 | McDevitt et al. | Jan 2008 | B2 |
7445699 | Scheer et al. | Nov 2008 | B2 |
7447609 | Guralnik et al. | Nov 2008 | B2 |
7505841 | Sun et al. | Mar 2009 | B2 |
7672782 | Mead et al. | Mar 2010 | B2 |
20020164046 | Walker et al. | Nov 2002 | A1 |
20030136676 | Miwa et al. | Jul 2003 | A1 |
20030161520 | Yamano et al. | Aug 2003 | A1 |
20040128279 | Matsuura et al. | Jul 2004 | A1 |
20040140213 | Kanters et al. | Jul 2004 | A1 |
20050025342 | Lee et al. | Feb 2005 | A1 |
20050285941 | Haigh et al. | Dec 2005 | A1 |
20060006997 | Rose-Pehrsson et al. | Jan 2006 | A1 |
20060159315 | Renz | Jul 2006 | A1 |
20080030352 | Shaw | Feb 2008 | A1 |
20100121796 | Staines et al. | May 2010 | A1 |
Number | Date | Country |
---|---|---|
WO 2005002313 | Jan 2005 | WO |
Number | Date | Country | |
---|---|---|---|
60779412 | Mar 2006 | US |