Embodiments described herein generally relate to a topology-augmented system for AI-model mismatch, and in an embodiment, but not by way of limitation, a system and method for estimating uncertainty in an artificial neural network.
Many users of artificial neural networks require an accurate neural network uncertainty estimation prior to integration of the artificial neural network into field or production systems. It is particularly important to have an accurate uncertainty when utilizing artificial neural networks and artificial intelligence in military defense systems. That is, determining whether the prediction of the artificial neural network can be trusted.
In the drawings, which are not necessarily drawn to scale, like numerals may describe similar components in different views. Like numerals having different letter suffixes may represent different instances of similar components. Some embodiments are illustrated by way of example, and not limitation, in the figures of the accompanying drawings.
A new method of determining the uncertainty of an artificial neural network (ANN) is disclosed, and it can be referred to as topology-augmented uncertainty. The new method for classification integrates ANN classification with out-of-distribution detection. In short, an out of distribution detection module is integrated with a deep learning ANN. It is edge-deployable, that is, an embodiment can be installed at locations wherein the real data are collected, such as in a military environment.
Specifically, in embodiment, given input data and a trained ANN, the system determines whether the data are out-of-distribution relative to the training data set and then determines if the ANN can reliably make a prediction. One application area is in a Synthetic Aperture Rader (SAR) based Automatic Target Recognition (ATR) of military vehicles with deep networks. The deployment of such a system in practice requires out-of-distribution detection capabilities to determine when the ATR cannot be trusted. There are also many other applications for this and other embodiments of this disclosure.
The ANN 100 is first trained with training data 120. After training, input data 130 for which a prediction is desired are input in the ANN 100. As indicated in
From the persistence diagram, the topological uncertainty of the layer can be calculated as defined in
Referring now specifically to
Then, at 220, a latent uncertainty of the ANN is determined. The latent uncertainty includes a function of a centroid for a latent representation of a class, a standard deviation of the latent representation of the class, and a latent representation of input data into the layer (222).
At 230, the uncertainty of the ANN is estimated as a function of the topological uncertainty and the latent uncertainty. More specifically, as indicated at 232, the uncertainty of the ANN is in an out of distribution condition when the topological uncertainty is greater than a first threshold and the latent uncertainty is greater than a second threshold. If the uncertainty is in an out of distribution condition, additional means are used to obtain the output classification (234). If the uncertainty is in an in distribution condition, then it can be concluded that the input data has been classified with an acceptable uncertainty (236).
The benefit of a combination of the topological uncertainty and the latent uncertainty of an ANN is illustrated in
Example computing platform 400 includes at least one processor 402 (e.g., a central processing unit (CPU), a graphics processing unit (GPU) or both, processor cores, compute nodes, etc.), a main memory 401 and a static memory 406, which communicate with each other via a link 408 (e.g., bus). The computing platform 400 may further include a video display unit 410, input devices 417 (e.g., a keyboard, camera, microphone), and a user interface (UI) navigation device 411 (e.g., mouse, touchscreen). The computing platform 400 may additionally include a storage device 416 (e.g., a drive unit), a signal generation device 418 (e.g., a speaker), a sensor 424, and a network interface device 420 coupled to a network 426.
The storage device 416 includes a non-transitory machine-readable medium 422 on which is stored one or more sets of data structures and instructions 423 (e.g., software) embodying or utilized by any one or more of the methodologies or functions described herein. The instructions 423 may also reside, completely or at least partially, within the main memory 401, static memory 406, and/or within the processor 402 during execution thereof by the computing platform 400, with the main memory 401, static memory 406, and the processor 402 also constituting machine-readable media.
While the machine-readable medium 422 is illustrated in an example embodiment to be a single medium, the term “machine-readable medium” may include a single medium or multiple media (e.g., a centralized or distributed database, and/or associated caches and servers) that store the one or more instructions 423. The term “machine-readable medium” shall also be taken to include any tangible medium that is capable of storing, encoding or carrying instructions for execution by the machine and that cause the machine to perform any one or more of the methodologies of the present disclosure or that is capable of storing, encoding or carrying data structures utilized by or associated with such instructions. The term “machine-readable medium” shall accordingly be taken to include, but not be limited to, solid-state memories, and optical and magnetic media. Specific examples of machine-readable media include non-volatile memory, including but not limited to, by way of example, semiconductor memory devices (e.g., electrically programmable read-only memory (EPROM), electrically erasable programmable read-only memory (EEPROM)) and flash memory devices; magnetic disks such as internal hard disks and removable disks; magneto-optical disks; and CD-ROM and DVD-ROM disks.
The above detailed description includes references to the accompanying drawings, which form a part of the detailed description. The drawings show, by way of illustration, specific embodiments that may be practiced. These embodiments are also referred to herein as “examples.” Such examples may include elements in addition to those shown or described. However, also contemplated are examples that include the elements shown or described. Moreover, also contemplated are examples using any combination or permutation of those elements shown or described (or one or more aspects thereof), either with respect to a particular example (or one or more aspects thereof), or with respect to other examples (or one or more aspects thereof) shown or described herein.
Publications, patents, and patent documents referred to in this document are incorporated by reference herein in their entirety, as though individually incorporated by reference. In the event of inconsistent usages between this document and those documents so incorporated by reference, the usage in the incorporated reference(s) are supplementary to that of this document; for irreconcilable inconsistencies, the usage in this document controls.
In this document, the terms “a” or “an” are used, as is common in patent documents, to include one or more than one, independent of any other instances or usages of “at least one” or “one or more.” In this document, the term “or” is used to refer to a nonexclusive or, such that “A or B” includes “A but not B,” “B but not A,” and “A and B,” unless otherwise indicated. In the appended claims, the terms “including” and “in which” are used as the plain-English equivalents of the respective terms “comprising” and “wherein.” Also, in the following claims, the terms “including” and “comprising” are open-ended, that is, a system, device, article, or process that includes elements in addition to those listed after such a term in a claim are still deemed to fall within the scope of that claim. Moreover, in the following claims, the terms “first,” “second,” and “third,” etc. are used merely as labels, and are not intended to suggest a numerical order for their objects.
The above description is intended to be illustrative, and not restrictive. For example, the above-described examples (or one or more aspects thereof) may be used in combination with others. Other embodiments may be used, such as by one of ordinary skill in the art upon reviewing the above description. The Abstract is to allow the reader to quickly ascertain the nature of the technical disclosure. It is submitted with the understanding that it will not be used to interpret or limit the scope or meaning of the claims. Also, in the above Detailed Description, various features may be grouped together to streamline the disclosure. However, the claims may not set forth every feature disclosed herein as embodiments may feature a subset of said features. Further, embodiments may include fewer features than those disclosed in a particular example. Thus, the following claims are hereby incorporated into the Detailed Description, with a claim standing on its own as a separate embodiment. The scope of the embodiments disclosed herein is to be determined with reference to the appended claims, along with the full scope of equivalents to which such claims are entitled.
Example No. 1 is a process for estimating an uncertainty of an artificial neural network comprising determining a topological uncertainty of the artificial neural network by forming a bipartite graph between input and output nodes in a layer of the artificial neural network and generating a persistence diagram as a function of the bipartite graph; determining a latent uncertainty of the artificial neural network; and estimating the uncertainty of the artificial neural network as a function of the topological uncertainty and the latent uncertainty.
Example No. 2 includes all the features of Example No. 1, and optionally includes a process wherein the bipartite graph comprises a weight matrix for the layer and data input into the layer.
Example No. 3 includes all the features of Example Nos. 1-2, and optionally includes a process wherein the latent uncertainty comprises a function of a centroid for a latent representation of a class, a standard deviation of the latent representation of the class, and a latent representation of input data into the layer.
Example No. 4 includes all the features of Example Nos. 1-3, and optionally includes a process wherein the uncertainty of the artificial neural network comprises an out of distribution condition when the topological uncertainty is greater than a first threshold and the latent uncertainty is greater than a second threshold.
Example No. 5 includes all the features of Example Nos. 1-4, and optionally includes a process wherein the computing of the persistence diagram comprises:
Example No. 6 includes all the features of Example Nos. 1-5, and optionally includes a process wherein the estimating the uncertainty of the artificial neural network comprises an out of distribution condition, and comprising using additional means to verify the out of distribution condition.
Example No. 7 includes all the features of Example Nos. 1-6, and optionally includes a process wherein the estimating the uncertainty comprises an in distribution condition and a classification of data input with an acceptable uncertainty.
Example No. 8 is a machine-readable medium comprising instructions that when executed by a processor executes a process comprising determining a topological uncertainty of the artificial neural network by forming a bipartite graph between input and output nodes in a layer of the artificial neural network and generating a persistence diagram as a function of the bipartite graph; determining a latent uncertainty of the artificial neural network; and estimating the uncertainty of the artificial neural network as a function of the topological uncertainty and the latent uncertainty.
Example No. 9 includes all the features of Example No. 8, and optionally includes a machine-readable medium wherein the bipartite graph comprises a weight matrix for the layer and data input into the layer.
Example No. 10 includes all the features of Example Nos. 8-9, and optionally includes a machine-readable medium wherein the latent uncertainty comprises a function of a centroid for a latent representation of a class, a standard deviation of the latent representation of the class, and a latent representation of input data into the layer.
Example No. 11 includes all the features of Example Nos. 8-10, and optionally includes a machine-readable medium wherein the uncertainty of the artificial neural network comprises an out of distribution condition when the topological uncertainty is greater than a first threshold and the latent uncertainty is greater than a second threshold.
Example No. 12 includes all the features of Example Nos. 8-11, and optionally includes a machine-readable medium wherein the computing of the persistence diagram comprises:
Example No. 13 includes all the features of Example Nos. 8-12, and optionally includes a machine-readable medium wherein the estimating the uncertainty of the artificial neural network comprises an out of distribution condition, and comprising using additional means to verify the out of distribution condition.
Example No. 14 includes all the features of Example Nos. 8-13, and optionally includes a machine-readable medium wherein the estimating the uncertainty comprises an in distribution condition and a classification of data input with an acceptable uncertainty.
Example No. 15 is a system including a computer processor and a memory coupled to the computer processor; wherein the computer processor and the memory are operable for determining a topological uncertainty of an artificial neural network by forming a bipartite graph between input and output nodes in a layer of the artificial neural network and generating a persistence diagram as a function of the bipartite graph; determining a latent uncertainty of the artificial neural network; and estimating the uncertainty of the artificial neural network as a function of the topological uncertainty and the latent uncertainty.
Example No. 16 includes all the features of Example No. 15, and optionally includes a system wherein the latent uncertainty comprises a function of a centroid for a latent representation of a class, a standard deviation of the latent representation of the class, and a latent representation of input data into the layer.
Example No. 17 includes all the features of Example Nos. 15-16, and optionally includes a system wherein the uncertainty of the artificial neural network comprises an out of distribution condition when the topological uncertainty is greater than a first threshold and the latent uncertainty is greater than a second threshold.
Example No. 18 includes all the features of Example Nos. 15-17, and optionally includes a system wherein the computing of the persistence diagram comprises:
Example No. 19 includes all the features of Example Nos. 15-18, and optionally includes a system wherein the estimating the uncertainty of the artificial neural network comprises an out of distribution condition, and comprising using additional means to verify the out of distribution condition.
Example No. 20 includes all the features of Example Nos. 15-19, and optionally includes a system wherein the estimating the uncertainty comprises an in distribution condition and a classification of data input with an acceptable uncertainty.
Number | Date | Country | |
---|---|---|---|
20240135150 A1 | Apr 2024 | US |