COPYRIGHT® 2016-2017 Alitheon, Inc. A portion of the disclosure of this patent document contains material which is subject to copyright protection. The copyright owner has no objection to the facsimile reproduction by anyone of the patent document or the patent disclosure, as it appears in the Patent and Trademark Office patent file or records, but otherwise reserves all copyright rights whatsoever. 37 C.F.R. § 1.71(d)(2017).
The present disclosure relates to the field of centralized databases storing digital fingerprints of objects that enable enhanced security, rapid searching, and high reliability. Other aspects include methods and apparatus to identify, track, and authenticate physical objects utilizing a suitable database.
Many different approaches are known to uniquely identify and authenticate objects, including labeling and tagging strategies using serial numbers, bar codes, holographic labels, RFID tags, and hidden patterns using security inks or special fibers. All these methods can be duplicated, and many add substantial costs to the production of the objects sought to be protected. Physical labels and tags are also themselves at risk of being lost, stolen, or counterfeited.
The following is a summary of the present disclosure in order to provide a basic understanding of some features and context. This summary is not intended to identify key/critical elements of the disclosure or to delineate the scope of the disclosure. Its sole purpose is to present some concepts of the present disclosure in a simplified form as a prelude to a more detailed description that is presented later.
There are many known approaches to establishing or reestablishing the authenticity of an object, including secure supply chains, expert assessment, and counterfeit detection. With the growth, however, of digital production (for example on a 3-D printer) and the ease of copying the digital files that drive such production, current production methods have become less secure. What is lacking in current methods, and is provided by the present disclosure, is the integration of the authentication and production processes into a single secure process. In embodiments, a digital fingerprint processor may be configured to: identify an activation of at least one of the one or more machines to attempt to produce or manufacture at least one of physical product or physical manufacture; responsive to completion of one or more operations associated with the activation by the one or more machines, acquire digital image data of a portion (e.g., an entire portion or some portion) of a physical object on or inside the one or more machines; analyze the digital image data to form a digital fingerprint of the physical object, wherein the digital fingerprint is responsive to structure of the physical object; and store the digital fingerprint in a database record of the database system.
Additional aspects and advantages of this disclosure will be apparent from the following detailed description of preferred embodiments, which proceeds with reference to the accompanying drawings.
In order to describe the manner in which the above-recited and other advantages and features of the present disclosure can be obtained, a more particular description follows by reference to the specific embodiments thereof which are illustrated in the appended drawings. Understanding that these drawings depict only typical embodiments of the disclosure and are not therefore to be considered to be limiting of its scope, the disclosure will be described and explained with additional specificity and detail through the use of the accompanying drawings in which:
Reference will now be made in detail to embodiments of the inventive concept, examples of which are illustrated in the accompanying drawings. The accompanying drawings are not necessarily drawn to scale. In the following detailed description, numerous specific details are set forth to enable a thorough understanding of the inventive concept. It should be understood, however, that persons having ordinary skill in the art may practice the inventive concept without these specific details. In other instances, well-known methods, procedures, components, circuits, and networks have not been described in detail so as not to unnecessarily obscure aspects of the embodiments.
It will be understood that, although the terms first, second, etc. may be used herein to describe various elements, these elements should not be limited by these terms. These terms are only used to distinguish one element from another. For example, a first machine could be termed a second machine, and, similarly, a second machine could be termed a first machine, without departing from the scope of the inventive concept.
The terminology used in the description of the inventive concept herein is for the purpose of describing particular embodiments only and is not intended to be limiting of the inventive concept. As used in the description of the inventive concept and the appended claims, the singular forms “a,” “an,” and “the” are intended to include the plural forms as well, unless the context clearly indicates otherwise. It will also be understood that the term “and/or” as used herein refers to and encompasses any and all possible combinations of one or more of the associated listed objects. It will be further understood that the terms “comprises” and/or “comprising,” when used in this specification, specify the presence of stated features, integers, steps, operations, elements, and/or components, but do not preclude the presence or addition of one or more other features, integers, steps, operations, elements, components, and/or groups thereof.
The methods described in the present disclosure enable the identification of an object without the need for attaching or associating physical tags or identifying materials with the object. A system does this by creating a unique digital signature for the object, which is referred to as a digital fingerprint. Digital fingerprinting utilizes the structure of the object, including random and deliberate features created during the manufacturing or use of the object, to generate a unique digital signature for that object similar to the way in which a human fingerprint references the friction ridges on a finger. Also, like a human fingerprint, the digital fingerprint can be stored and retrieved to identify objects when they are encountered at a later date.
Eliminating the need to add tags or any physical modifications to the object offers a number of advantages to manufacturers, distributors, sellers, and owners of goods. Forgoing the addition of physical identifiers reduces the cost of manufacturing and is more secure than physical togging. Moreover, physical identifiers can be lost, modified, stolen, duplicated, or counterfeited whereas digital fingerprints cannot.
Unlike prior art approaches that simply utilize a comparison of pixels, a system in accordance with the present disclosure utilizes the extraction of features to identify and authenticate objects. Feature extraction enables users to take a large amount of information and reduce it to a smaller set of data points that can be processed more efficiently. For example, a large digital image that contains tens of thousands of pixels may be reduced to a few locations of interest that can be used to identify an object. This reduced set of data is called a digital fingerprint. This digital fingerprint contains a set of fingerprint features or locations of interest which are stored as feature vectors. Feature vectors make image processing more efficient and reduce storage requirements as the entire image need not be stored in the database, only the feature vectors need to be stored. Examples of feature extraction algorithms include but are not limited to edge detection, corner detection, blob detection, wavelet features, Gabor, gradient and steerable output filter histograms, scale-invariant feature transformation, active contours, shape contexts, and parameterized shapes.
While the most common applications of the system may be in the authentication of manufactured goods and documents, the system is designed to be applicable to any object that can be identified, characterized, quality tested, or authenticated with a digital fingerprint. These include but are not limited to mail pieces, parcels, art, coins, currency, precious metals, gems, jewelry, apparel, mechanical parts, consumer goods, integrated circuits, firearms, pharmaceuticals, and food and beverages. Here the term “system” is used in a broad sense, including the methods of the present disclosure as well as apparatus arranged to implement such methods.
In this application, the term “scan” is used in the broadest sense, referring to any and all means for capturing an image or set of images, which may be in digital form or transformed into digital form. Images may, for example, be two dimensional, three dimensional, or in the form of a video. Thus a “scan” may refer to an image (or digital data that defines an image) captured by a scanner, a camera, a specially adapted sensor or sensor array (such as a CCD array), a microscope, a smartphone camera, a video camera, an x-ray machine, a sonar, an ultrasound machine, a microphone (or other instruments for converting sound waves into electrical energy variations), etc. Broadly, any device that can sense and capture either electromagnetic radiation or mechanical wave that has traveled through an object or reflected off an object or any other means to capture surface or internal structure of an object is a candidate to create a “scan” of an object. Various means to extract “fingerprints” or features from an object may be used for example, through sound, physical structure, chemical composition, or many others. The remainder of this application will use terms like “image” but when doing so, the broader uses of this technology should be implied. In other words, alternative means to extract “fingerprints” or features from an object should be considered equivalents within the scope of this disclosure.
In this application, different forms of the words “authenticate” and “authentication” will be used broadly to describe both authentication and attempts to authenticate which comprise creating a digital fingerprint of the object. Therefore, “authentication” is not limited to specifically describing successful matching of inducted objects or generally describing the outcome of attempted authentications. As one example, a counterfeit object may be described as “authenticated” even if the “authentication” fails to return a matching result. In another example, in cases where unknown objects are “authenticated” without resulting in a match and the authentication attempt is entered into a database for subsequent reference the action described as “authentication” or “attempted authentication” may also, post facto, also be properly described as an “induction”. An authentication of an object may refer to the authentication of an entire object or of a portion of an object.
Because digital fingerprinting works with many different types of objects, it may be useful to define what regions of digital images of objects are to be used for the extraction of features for authentication purposes. The chosen regions may vary for different classes of objects. In some embodiments, a chosen region may be the image of the entire object; in other embodiments chosen regions may be one or more sub-regions of the image of the object.
For instance, in the case of a photograph, a digital image of the entire photograph may be chosen for feature extraction. Each photograph is different and there may be unique feature information anywhere in a photograph. In such a case, the authentication region may be the entire photograph.
In some embodiments, multiple regions may be used for fingerprinting. In some examples, there may be several regions where significant variations take place among different similar objects that need to be distinguished while, in the same objects, there may be regions of little significance. In other examples, a template may be used (see
In one embodiment, an object, such as a bank note, may be deemed authenticated if a few small arbitrary regions scattered across the surface are fingerprinted, possibly combined with one or more recognitions of, for example, the contents of a region signifying the value of the bank note or one containing the bank note serial number. In such examples, the fingerprints of any region (along with sufficient additional information to determine the bank note value and its purported identity) may be considered sufficient to establish the authenticity of the bill. In some embodiments, multiple fingerprinted regions may be referenced in cases where one or more region may be absent from an object (through, for example, tearing) when, for example, a bank note is presented for authentication. In other embodiments, however, all regions of an object may need to be authenticated to ensure an object is both authentic and has not been altered.
In one embodiment, a passport may provide an example of feature extractions from multiple authentication regions, see
In some embodiments, an ability to define and store optimal authentication regions for classes of objects may offer benefits to a user. In some embodiments, it may be preferable to scan limited regions of objects rather than to scan entire objects. For instance, in the case of an article of designer clothing, scanning a clothing label may be preferable to scanning an entire garment. Further, defining such regions may enable detection of partial alteration of an object.
Once an authentication region is defined, specific applications may be created for different markets or classes of objects that may assist users in locating and scanning an optimal authentication region. In some embodiments, for example when utilizing a mobile device, a location box and crosshairs may automatically appear in the viewfinder of a smartphone camera application, to help the user center the camera on an authentication region, and automatically lock onto a region and complete a scan when the device is focused on an appropriate area. It should be noted that, although some examples suggested above are two-dimensional objects (passport, bank note), the present disclosure is fully applicable to three-dimensional objects as well. As previously noted, scanning may be of any kind, including 2-D, 3-D, stereoscopic, HD, etc. and is not limited to the use of visible light or to the use of light at all (as previously noted, sonar and ultrasound are, for example, appropriate scanning technologies).
In some embodiments, objects may have permanent labels or other identifying information attached to them. In addition to the objects themselves, these attachments may also be referenced as features for digital fingerprinting, particularly where the label or other identifying information becomes a permanent part of the object. In one example, a permanent label may be used as an authentication region for the object to which it is affixed. In another example, a label may be used in conjunction with the object itself to create a fingerprint of multiple authentication regions referencing both a label and an object to which the label is affixed.
In one example, wine may be put into a glass bottle and a label affixed to the bottle. Since it is possible that a label may be removed and re-applied elsewhere merely using the label itself as an authentication region may not be sufficient. In this case the authentication region may be defined so as to include both a label and a substrate it is attached to—in this example some portion of a label and some portion of a glass bottle. This “label and substrate” approach may be useful in defining authentication regions for many types of objects, such as various types of goods and associated packaging. In other instances, authentication may reveal changes in the relative positions of some authentication regions such as in cases where a label has been moved from its original position, which may be an indication of tampering or counterfeiting. If an object has “tamper-proof” packaging, this may also be included in the authentication region.
In some embodiments, multiple authentication regions may be chosen from which to extract unique features. In a preferred embodiment, multiple authentication regions may be selected to enable the separate authentication of one or more components or portions of an object. For example, in one embodiment, features may be extracted from two different parts of a firearm. Both features may match the original firearm but since it is possible that both parts may have been removed from the original firearm and affixed to a weapon of different quality, it may also be useful to determine whether the relative positions of the parts have changed. In other words, it may be helpful to determine that the distance (or other characteristics) between Part A's authentication region and Part B's authentication region remains consistent with the original feature extraction. If the positions of Parts A and B are found to be consistent to the relative locations of the original authentication regions, the firearm may be authenticated. Specifications of this type may be stored with or as part of a digital fingerprint of an object.
Once a digital fingerprint of an object is acquired, a characterization of the object and corresponding fingerprint may be recorded in a database. For example, in some embodiments, a fingerprint may comprise one or more feature vectors. In some applications, the database may be secure. In some embodiments, a unique identifier or ID may be assigned to an object, for example in cases where an ID may be a convenient index in an application. However, an ID is not required since a digital fingerprint itself may serve as a key for searching a database. In other words, by identifying an object by the unique features and characteristics of the object itself, arbitrary identifiers, labels, tags, etc. are rendered unnecessary.
A database of digital fingerprints may form the basis of a system to identify or track an object in a supply chain, distribution network, sales channel, or any other series of locations, ownerships, or possessions. An identification system based on digital fingerprinting has unique advantages and capabilities that are not available with identification systems based on currently available methods.
Current identifiers such as holograms, bar codes, and serial numbers may all be duplicated (or otherwise counterfeited) with varying degrees of effort. Because such identifiers are vulnerable to diversion and counterfeiting, a counterfeit object with a duplicated identifier—and counterfeit objects with otherwise counterfeited identifiers—may enter a supply chain or distribution network. The counterfeit identifiers may subsequently be registered in a conventional identification or track-and-trace system. All current identification systems rely on determining whether the identifier (label, hologram, RFID tag) is legitimate, not whether the object itself is legitimate.
Due to this weakness, identification and tracking systems based on conventional approaches like bar codes or serial numbers cannot prevent counterfeit objects from entering a supply chain or prevent the resulting corruption of the system database. A counterfeit object may be mistakenly identified as genuine, and generate a false audit trail as it is identified through the supply chain. Two or more objects with the same ID (one genuine, one or more counterfeit) may exist at the same time without detection. Without physically examining the objects it may be impossible to tell which object is genuine and, depending on the quality and origin of the counterfeits, even physical examination of an object may be insufficient to determine its authenticity.
Once a counterfeit object has been detected in a conventional system, false audit trails must be removed from the system database to restore integrity. This may be difficult depending on the structure of the database and the complexity of the identification data. In some cases, the objects may not have any further contact with a system (for instance if they are purchased by a consumer), and the record will never be identified as false, leaving the database permanently corrupted.
In some embodiments of the present disclosure, an object may be scanned and identified at initial manufacture. Alternatively, an object may be scanned and identified at any subsequent time or location for entry into a database. Preferably, the scanning and identification is carried out when an object is either in the possession of its manufacturer or has been transferred by secure means to the point of scanning so that its legitimacy at the point of identification may be adequately established. However, this is not necessary and the adequate legitimacy of an object may be established through various other means both prior to or following induction.
In an embodiment, the system subsequently identifies the scanned and identified object every time the object is scanned again, typically at discrete steps in manufacturing, distribution, and sale, and as part of a data collaboration scheme.
Examples of scanning locations might include scanning at the point of manufacture, when packaged, when placed in inventory, when shipped, and at a retail point of sale (e.g. upon arrival and again when sold), as illustrated in the tracking process 300 of
As previously described, a “scan” may refer to an image (or to digital data that defines an image) captured by a broad range of capturing devices. In an embodiment, a scan of an object may capture both features of the object and features of an identifier that has been attached to the object. Feature vectors extracted from authentication regions located on an attached identifier are based on the substances of which the identifier is physically comprised rather than the information (typically alphanumeric) that is intended to be communicated by the identifier. For instance, in the case of a wine bottle, features may be captured from the bottle and from a label affixed to the bottle. If the label includes a standard UPC bar code, the paper of the label and the ink pattern of the bar code may be used to extract a feature vector without reading the alphanumeric information reflected by the bar code. An identifier, such as a UPC bar code print consisting of lines and numbers, has no greater significance in the creation and use of a feature vector than a set of randomly printed lines and numbers.
A server 412 may be provisioned to provide identification and/or tracking data analysis and reporting. The server 412 has access to a database 420 which may be used to store digital fingerprints and related data. The server can query or search the database 420 for digital fingerprint search and matching. The database 420 is preferably coupled to the cloud server 406 in some embodiments. A mobile user device 410 such as a smartphone, tablet, laptop computer, or dedicated device may be configured for communications with the server 412 to request and receive a reply or authentication report for an object of interest. This architecture is simplified, and in any event, is merely illustrative and not intended to be limiting.
In some embodiments, sensors may be attached to the object, and sensor data can flow back to the database in either a continuous fashion (near real time), or in discrete data transfer events. For example, data transfer may occur when an authentication event occurs. For instance, if there is a GPS chip attached to the object, data flow can start when the object is first registered in the system, and continue to flow as the object changes location. Continuous or intermittent data updates may also be buffered in local memory in a sensor attached to an object, and then downloaded the next time the object is scanned and authenticated. This example provides an itinerary record of where the object has traveled.
As an example of the potential uses of sensor data, many products like food and beverages can degrade with exposure to certain environmental factors during storage and shipment. Examples of sensor data could include temperature, humidity, light exposure, altitude, oxygen level, or other factors, as well as location such as GPS data.
The next step, block 510, is to query the database for the next valid or expected location. A unique itinerary may not be known, but a set of valid or expected locations may be known. The next actual location of the object (as determined by imaging and matching digital fingerprints) may be compared to the expected location(s) returned by the database, block 512. If that comparison indicates a departure from the expected or authorized route, decision 520, the result may be reported to a user, block 522. (A report that the object is on track may be reported as well.) Other options may be implemented such as a quantity check, block 524. The process returns or terminates at block 526.
Most existing identification systems are designed only to be accessed by manufacturers or their authorized distributors, and some require specialized scanners or equipment. However, consumers also have a vested interest in determining whether the objects they are buying are authentic. Other parties may also have an interest in determining the authenticity of an object, for example in response to offered or future incentives for object data collection. In some embodiments, the present system is designed to enable anyone along the supply, distribution, or sales chain, from manufacturer to the retail consumer or other members of the public, to access the system and determine whether the object is authentic. A specialized scanner is not required in all cases. For example, in one embodiment a mobile phone application designed for public use can be employed to scan an object, query the database, and determine if the object is authentic.
Finally, data collected by a digital fingerprinting system offers a variety of useful information to stakeholders along the supply, distribution, and sales chain. Reports can be generated on individual objects, or on sets of objects. These reports can include, but are not limited to, the locations of objects over time, audit trails, points of entry of counterfeit goods, and exposure to environmental variables over the course of an object's lifetime.
In some instances, an identifier such as a tag or a label may be considered a useful addition to an object, mainly for two reasons. The first main reason is that a human may need to reference it for information. For example, a tag or a label may inform a store clerk as to the particular style and size of a piece of clothing by a particular manufacturer, or it may tell a postal carrier the address to which a mail piece should be delivered. The second main reason has to do with the employment of machine-readable tags, for example when an Intelligent Mail Barcode on a mail piece (unreadable by humans) is used to route the mail piece by machine. The entire class of existing machine-readable tags can be replaced by the methods of the present disclosure. In some instances, tags or labels may still be needed for human information but they are unnecessary for the purpose of authentication or any machine-readability functions.
Because digital fingerprinting exploits natural features and often involves scanning an object under variable conditions, it is highly unlikely that two different scans will produce the exact same fingerprint. As a result, an ability to look up objects in the database when there is a near-miss is included. In one example, two feature vectors [0, 1, 5, 5, 6, 8] and [0, 1, 6, 5, 6, 8] are not identical but, given a certain difference metric, they may be close enough to say with a level of certainty that they are from a certain object that has been scanned before. This is particularly true if, otherwise, the nearest feature vector of a different object is, for example, [5, 2, 5, 8, 6, 4]. For example, a distance between vectors of n-dimensions may be calculated and used as one metric of similarity or “closeness of match” between two or more vectors. The distance to the next nearest candidate may also be considered.
In some embodiments, where we have an original document or other object fingerprinted, the digital fingerprinting techniques allow region-by-region matching, making apparent which (if any) regions have changed from the original. Thus, for example, a good overall match on a passport may be achieved but if none of the matches happen in the photograph it becomes apparent that the photograph has likely been changed. Further, if some individual or group, say a known terrorist group, has a certain pattern or regularity to altering passports-such as to change the photo, the date of birth, and one digit of the passport number-then this ability to find altered regions may also provide the ability to discern a pattern of changes that may be used to identify a “signature” modus operandi of a particular counterfeiter. Thus, aspects of digital fingerprinting may be applied not only to detect a forged or altered object, but may also be used to identify a counterfeiter or otherwise discover the source of an altered object.
This disclosure teaches an integrated, secure process that combines production or other creation of physical objects with secure induction of digital information, including a digital fingerprint of that object, into an authentication system for later use. That later use may include identification, authentication, tracking, or other purposes. The digital fingerprint may be associated with other data such as a serial number, manual, photograph, tracking and manufacturer information, or other information. The digital fingerprint and associated information may be stored in a secure database for later use.
The exact method of creating such a digital fingerprint is not important, in some embodiments. There are many satisfactory ways to extract a digital fingerprint of an object known in the literature, including SIFT, SURF, and others.
The secure induction process taught here does not depend on any particular database system, provided the database is so constructed that no one, including the creator of the object, can inappropriately enter authentication data into the system. For example, no one can take an object he did not create and induct it as though he did. Nor can anyone enter a non-existent object into the authentication process. Nor can anyone produce an object with the system that is not inducted (this last is to avoid “midnight runs” of counterfeit objects on a real assembly line).
Because the creation of the object and the induction of the digital fingerprint are part of a single process, and because the digital fingerprint does not rely on tags, labels, barcodes or other artificial identifiers of an object, the process taught here is better proofed against counterfeiting than existing manufacturing processes.
Manufacturing is becoming increasingly decentralized. The advent of 3-D printing and other rapid prototyping or small-lot production systems, and the ease with which the digital files behind many such systems can be copied, mandate an authentication process that at the same time ensures that the local manufacturer cannot corrupt the process (for example by issuing false certificates of authenticity) while securely tying the object to authenticating information generated when the provenance of the object is known.
If a manufacturer controls the production, induction, and ultimate authentication of the object, tying the creation and digital fingerprinting processes together in a way safe from, say, the manufacturer's own fraud may not be critical. When, however, parts are contracted out, and particularly when they can be made on relatively inexpensive equipment from digital files (such as on a 3-D printer), it becomes important to establish that the digital fingerprints to be used m authentication are captured as part of the manufacturing process and that an independent record linking the creation of the object with its authentication information be created in a way not corruptible by, say, counterfeiters. That secure integrated process is what this disclosure teaches.
Some embodiments include a single secure process for the automatic creation and induction (for later authentication) of a digital fingerprint of an object. In one embodiment, an object is created on a 3-D printer. The printer creates the object in accordance with its instructions and, while doing so, extracts one or more digital fingerprints of the object, where those digital fingerprints take advantage of the natural randomness of the printing process (e.g. “random” surface features) but may also reference other, deliberate features (e.g., artificial identifiers such as microscopic etchings). The digital fingerprints may be tied to other digital information about the object and the creation process. The digital fingerprints and associated data are stored securely n a database for later use for authentication. The overall system is secure from inappropriate tampering by the producer or anyone else, perhaps by encrypting the database to secure it from tampering during the manufacturing process or at a later time.
Among the objects associated with a digital fingerprint of a newly-manufactured object may be a template (for how the just-created components will later be assembled), a time and date stamp for the creation, a production device identifying number, product warnings, intended product recipients, sensor data related to, say, a controlled manufacturing environment, as well as user-supplied information such as product manuals and certificates of creation, provenance, or authenticity. The manufacturing process may automatically create a secure data record for later retrieval as part of object authentication. The ultimate use of the data associated with the object at its point of creation is not relevant.
In another embodiment, an artist may create a work of art such as an etching, digitally fingerprint the artwork, associate the digital fingerprint with an authentication certificate or other information associated with the work, and insert the resulting data in a secure database for later authentication by say, a purchaser.
In another embodiment, all components that constitute an object may be inducted during their individual production processes. Tat induction may include a list or partial list of all of the final product's inducted components and their intended relationships in the final product. For example, while a computer is assembled on an assembly line, each circuit board, drive, and chip is inducted prior to final assembly. An actor or stakeholder may later to authenticate all critical components of an object as well as the object itself to verify an object's authenticity. Some embodiments thus provide a much higher level of authenticity assurance, than information acquired from authenticating the outside of the completed object. This approach insures against a wide array of scenarios including sabotage, the swapping of legitimate components for illegitimate, and other fraudulent activities.
In a further embodiment, this method may be used in fraud prevention where goods are returned to original supplier or point of sale. A person seeking to commit fraud may intentionally replace the original with non-standard components. Under the teachings of this disclosure, if, at the point of manufacture or assembly sanctioned or branded components of an object are themselves inducted and associated with the final object, the object itself cannot be authenticated unless its associated components are as well.
The machine 4 to produce manufacture physical products or manufactures may be any machine such as one to utilize raw materials to produce physical objects (such as a 3D printer to layer a material to form a physical object, a printer to deposit ink on paper to form a document, or the like, or combinations thereof), or one to utilize assemble/arrange/connection/etc. one or more components to form a system (e.g., to arrange gears to form a gear train). The machine 4 may include a processor 11 and an interface 8 (e.g., a network interface) to communicate with the database system 19 before, during, and/or after production or manufacture for authenticated production.
The digital fingerprint processor 15 may be configured to identify an activation of the machine 4 to attempt to produce or manufacture at least one of the physical products or physical manufactures. In some examples, the digital fingerprint processor 15 may be in communication with a processor 11 installed in the machine. For instance, the digital fingerprint processor 15 may detect the activation of the machine by receiving a signal generated by the processor 11 and transmitted over the network to indicate the activation.
The digital fingerprint processor 15 may be configured to acquire digital image data of a portion of a physical object, which may be inside/on/etc. the machine 4. In some examples, the digital fingerprint processor 15 and/or the processor 11 may control a camera 12 that is integrated in the machine 12 or non-integrated (e.g., a separable component, perhaps a mobile device, mounted on the machine or otherwise positioned proximate to the machine).
The portion of the physical product may be a completed surface. For instance, if the machine 4 is a 3D printer, the camera 12 may capture an image of a completed surface (the 3D print job need not be complete at a time the image is captured so long as the imaged surface is complete). For example, the imaged surface may be edges of completed layers where additional layers are still to be deposited on top of the completed layers. In some examples such as some 3D printing examples, the image may be a surface that is to be an external surface of the completed 3D print job; however, in other examples the image may be of any surface that is to be accessible on completion of the manufacture or production (for instance an image of a surface of a gear on the interior of a watch).
The digital fingerprint processor 15 may be configured to analyze the digital image data to forma digital fingerprint of the physical object. The digital fingerprint may be responsive to structure of the physical object. The digital fingerprint processor 15 may store the digital fingerprint in a database record 17 of the data structure 16, which may be to induct at least one of the physical product or physical manufacture 5 produced by the machine 4.
It should be appreciated that in some examples, the imaging may be, for instance, during or after production of the physical product or physical manufacture 5. In one example, the imaging may be prior to completion of the production of the physical product or manufacture 5 produced by the machine 4 (the image may be of a completed surface and/or a component to be used to complete production of the physical manufacture, e.g., a gear in the case of producing gear trains or even an unmarked portion of a piece of paper showing a structure of paper fibers in the case of producing documents). In some examples, the digital image data may be acquired responsive to completion of one or more operations associated with the activation by the machine 4. For instance, the machine 4 may be turned on, initialized, loaded finish some or all phases of fabrication, or the like, or combinations thereof, and the digital image data may be acquired responsive to the completion of one or more of these operations. In some examples, the digital image data may be acquired responsive to a three-dimension object being fabricated, e.g., partially fabricated (the digital image data may be of a completed portion of the three-dimension object, say a bottom portion of a sidewall of an uncompleted the three-dimension object where additional layers are yet to be added to the top). For instance, the digital image data may be of a “clean” piece of paper to be used for an initialized print job, or a portion of the piece of paper that has been printed (e.g., marked paper). The digital image may be acquired of a gear responsive to an assembly of a gear train, or the gear being loaded into the machine (e.g., yet-to-be-assembled into the gear train say positioned on a feeder such as a conveyor to be used to assemble the gear train).
In examples in which the imaging is prior to completion of the production, the digital fingerprint processor 15 may perform a validation check prior to completion of the production. Based on a result of the validation process, the digital fingerprint processor 15 may control the machine 4 to prevent the production from completing in the case of unauthorized production (e.g., to abort the job). For instance, the digital fingerprint processor 15 may signal the processor 11 or signal a power supply (not shown) to hinder or stop an attempt for a “midnight run”. In some examples, it may be possible for the processor 11 to perform a validation check responsive to an attempt to activation the machine. For instance, if the processor 11 is unable to communicate with the digital fingerprint processor 15 responsive to a user control to operate the machine 4, the process 11 may reject a command to operate the machine 4 and/or put the machine 4 into a low power state (in which say the processor 11 may operate but one or more other components of the machine may not receive operating power).
In block 802, the processor may acquire digital image data of a portion of a physical object proximate to the one or more machines (e.g., on the one or more machines, in the one or machines, or the like, or combinations thereof) responsive to completion of one or more operations associated with the activation by the one or more machines. In block 803, the processor may analyze the digital image data to form a digital fingerprint that is of the physical object and responsive to a structure of the physical object. In block 804, the processor may store the digital fingerprint in a database record of the database system.
It should be appreciated that some or all of process 800 may be performed for each physical object to be manufactured by the one or more machines (e.g., each gear produced by gear-producing machine), in some embodiments. Also, if the physical object is part of an assembly such as a gear-train, some or all of process 800 may be repeated for each component (e.g., each gear), and some or all of process 800 may be repeated for the assembly. In such a case, each digital fingerprint may be stored in its own record, which may be linked together.
Some systems or processes of authenticated production may include authorization check(s) combined with induction for authentication (authorization features may be used to prevent midnight runs in the first place); however, this is not required. Even if a midnight run is performed without creating records, such product may not may not be inducted into the authentication system and may not be authenticatable due to the lack of induction. Such non-authenticatable product may be detected at points of authentication in other parts of a supply chain.
In block 901, a processor of an authentication system (e.g., a processor to control operation of one or more machines, such as processor 11 of
In diamond 902, the processor may identify whether it is coupled to an authentication system, e.g., whether it can communicate with a digital fingerprint processor). If the processor is not coupled to the authentication system, the processor may prevent the one or more machines from producing or manufacturing the physical products or manufactures in block 903.
In block 904, the processor may identify information about the attempt (e.g., acquire digital image data of a portion of a physical object proximate to the one or more machines in responsive to completion of one or more operations associated with the activation by the one or more machines). In some examples, the one or more machines may include a component to limit and/or detect physical access to the physical object once loaded into the one or more machines as part of an initialization.
In block 905, the processor may transmit the identified information over the network to the authentication system (e.g., transmit the acquired digital image data and optionally other data, such as login credentials of an operator, time of day, intrusion information, or the like, or combinations thereof). In block 906, the processor may identify whether production or manufacture is authorized, e.g., may check whether a signal such as an authorization signal is received. In block 907, the processor may prevent the one or more machines from producing or manufacturing the physical products or manufactures, e.g., prevent completion of one or more additional operations associated with the attempt.
In block 910, if authorized, the processor may not prevent the one or more machines from producing or manufacturing based on the attempt (e.g., allow all operations by the one or more machines on the physical object to produce a single authentication-database-inducted one of the physical products or manufactures). In block 911, the processor may identify metadata about the manufacture or production for the completed attempt (such as time of completion), and may provide the metadata to the authentication system, which may add the metadata to the record.
In the illustrated example, a manufacturer may use authenticated production to manufacture a gear train 1020. In one example, the manufacturer may authenticate 1002 a physical object entering a physical space controlled by the manufacturer. e.g., the security “wall” 1007. This may be one of the authentication requests 1060. For instance, the manufacturer may receive a physical object, may communicate with digital fingerprint processor 1042 of a database system over a secure communications link 1013, e.g., may provide an image of the physical object to the digital fingerprint processor 1042. The digital fingerprint processor 1042 may perform any authenticated process described herein based on the image to match the image to a record existing in the secure database server 1044, and may provide a result responsive to a match.
The result may verify that the physical object is the gear 1006. In some examples, the gear 1006 may originate from a secure component manufacturer having its own security “wall” 1008. The secure component manufacturer may have produced the inducted component 1006 using authenticated production including sending and/or receiving information over secure communications link 1012 and integrated induction image capture 1009. For example, the secure component manufacturer may have performed any of the processes described with respect to
Referring again to the gear train manufacturer, the manufacturer may, following authentication, utilize the gear 1006 in assembly 1014 of gear train 1020 (e.g., may produce a system using gear 1006 and/or some other inducted components). Also, the gear train manufacturer may induct 1022 the gear train 1020 into the database system as part of a production process. The database system may store metadata in the record for the gear train 1020, such as information about an authorized user 1056 indicated by authentication credentials supplied in connection with the induction request, or other metadata for inventory notifications 1054 such as information about a time or other characteristic of the authenticated production of the gear train 1020.
For instance, the gear train manufacturer may perform any of the processes described with respect to
The database system may be physically located in a different physical space than the physical space of security “wall” 1007. For instance, the database system may be located in a security “wall” 1040. Personnel given authorization to enter the gear train plant may not necessarily be given authorization to enter a data center of the database system.
The third example shows a counterfeit component coming from a manufacturer and being added to the object. This is very similar to the previous example except for the external origin of the counterfeit and the possibility that, under the current system, the recipient may not know the component is counterfeit.
Finally, on the far right, we show the customer purchasing a counterfeit object unintentionally.
Nothing in our technology is designed to counter the problem of known knock-offs or known thefts, willingly purchased.
The very large, and rapidly growing, trade in counterfeits is proof of how difficult it is to guard against any of these illicit activities. Counterfeits get around current anti-counterfeiting methods either by being intentional internal fraud (the “homemade” component problem) or by mimicking legitimate objects well enough to fool the recipient.
The technology taught in our patent documents, when employed in a system such as the one outlined in
The authentication system taught in our patent documents can be applied to a supply chain to reduce or eliminate the risks of counterfeit objects entering into or legitimate objects being diverted from, the chain.
The arrows 1201 show the stakeholder receiving components for his production. To ensure that the received components are legitimate, the stakeholder authenticates them as described below.
The stakeholder then produces his objects and inducts them into the database along with other required information so that they may later be authenticated by downstream stakeholders. He then ships them (arrows 1213) to his customers.
Authentication has several stages. Although seen from the point of view of any object in the system, induction precedes authentication, when seen from the perspective of an individual stakeholder, the opposite is the case. It is from that latter perspective that this diagram is discussed.
Not all objects need be authenticated. The database has previously been populated by an entity with a primary stake in the final object being legitimate, his may be the final recipient (in the case of contracted parts) or the brand holder (in the case of high-end consumer goods), or anyone else as appropriate. All objects required by the completed object that are sufficiently important to be authenticated are entered along with the identity of their authorized manufacturers.
Authentication comprises the following steps:
Undetected Diversion of a Legitimate Object
Illicit introduction of a counterfeit. If a part requires authentication and recipient attempts to authenticate it, a counterfeit part will not pass authentication and thus not enter the supply chain. In the event that the recipient fails to authenticate the object, his own products will fail of authentication by downstream stakeholders. Only an end-user who either intentionally receives a counterfeit object or who doesn't care is not protected.
Illicit substitution. If a legitimate manufacturer decides to cut corners and either purchase or himself produce a substitute for apart that requires authentication (what we have been calling “essential”), he cannot authenticate that part (since it did not come from a legitimate manufacturer who had previously inducted it and since he himself cannot induct it). As a result, the objects he produces using that part will not pass authentication by downstream stakeholders.
Diversion. Each object that requires authentication has a database entry associated with its manufacturer and with its intended recipient. If the intended recipient never authenticates the object, then it can be presumed to have gone astray (either between the manufacturer and intended recipient or have been put to illicit use by the intended recipient).
In other words, an object is authentic when it is inducted into the database by its manufacturer. This can only happen when the manufacturer is authorized to manufacture the object and all of the object's (essential) components are authentic.
Once all the essential components have been authenticated (there may be components that do not require authentication), the stakeholder assembles them into his products. He then inducts his products into the data store. No induction is done by end-users or by middlemen whose products are not required to be authenticated.
Induction comprises the following steps:
Example 1 is a method of capturing an image as an integral part of the manufacturing process, of extracting a digital fingerprint from that image, of storing the digital fingerprint in a secure database for later use in authentication. The digital fingerprint data stored in the database may include contextual and metadata associated with the object. The object, through its digital fingerprint, may itself serve as a physical token to gain access to the additional associated digital data in the database. This could be done by imaging the object, extracting a digital fingerprint of the object, and using the extracted digital fingerprint to access the associated data. The “manufactured object” may be a document and the “manufacturing process” may include creating and authenticating a document on a printer also fitted for extracting a digital fingerprint of the created document. In this embodiment, the pages of a will might be digitally fingerprinted as the pages are printed, for later use in authentication of the originals.
Example A-1 is an apparatus for authenticated production. The apparatus may include one or more machines to produce or manufacture physical products or physical manufactures; a database system in communication with the one or more machines, the database system including a digital fingerprint processor configured to: identify an activation of at least one of the one or more machines to attempt to produce or manufacture at least one of the physical products or physical manufactures; responsive to completion of one or more operations associated with the activation by the one or more machines, acquire digital image data of a portion of a physical object proximate to the one or more machines; analyze the digital image data to form a digital fingerprint of the physical object, wherein the digital fingerprint is responsive to structure of the physical object; and store the digital fingerprint in a database record of the database system.
Example A-2 includes the subject matter of example A-1 (and/or any other example herein), wherein the digital fingerprint processor is further configured to: identify the completed production or manufacture of the at least one of the physical products or manufactures; and store information about a time or other characteristic of the completed production or manufacture in the database record.
Example A-3 includes the subject matter of any of examples A-2 through A-3 (and/or any other example herein), wherein the digital fingerprint comprises information about the physical object, and wherein the digital fingerprint processor is further configured to: obtain at least one of additional information about the physical object, information about the attempt, information about a state of the at least one machine, or information an operator of the at least one machine; and store the at least one of the additional information, the information about the attempt, the information about the state of the at least one machine, or the information about the operation of the at least one machine.
Example A-4 includes the subject matter of any of examples A-2 through A-3 (and/or any other example herein), wherein the additional information about the physical object comprises contextual and/or metadata associated with the physical object.
Example A-5 includes the subject matter of any of examples A-2 through A-4 (and/or any other example herein), wherein the additional information comprises a serial number or other value present in the digital image data.
Example A-6 includes the subject matter of any of examples A-2 through A-5 (and/or any other example herein), wherein the at least one machine comprises a fabrication device and the portion of the physical object comprises a completed surface.
Example A-7 includes the subject matter of any of examples A-2 through A-6 (and/or any other example herein), wherein the portion of the physical object comprises a surface to be accessible on completion of at least one of the physical products or physical manufactures.
Example A-8 includes the subject matter of any of examples A-2 through A-7 (and/or any other example herein), wherein the at least one of the physical products or physical manufactures comprises a system including a first component and a second component, and wherein said physical object comprises the first component.
Example A-9 includes the subject matter of any of examples A-2 through A-8 (and/or any other example herein), wherein the component of the plurality of components comprises a first component, and wherein the digital fingerprint processor is further configured to: acquire additional digital image data of a portion of the second component; analyze the additional digital image data to form a digital fingerprint of the second component, wherein the digital fingerprint is responsive to structure of the second component; and store the digital fingerprint of the second component in the same or another database record of the database system.
Example A-10 includes the subject matter of any of examples A-2 through A-9 (and/or any other example herein), wherein the digital fingerprint processor is further configured to transmit a control signal to the at least one machine to prevent the attempt to produce or manufacture the at least one of the physical products or physical manufactures from completing based on content of the digital image data.
Example A-11 is method of authenticated production using one or more machines to produce or manufacture physical products or manufactures. The method may include identifying an activation of at least one machine of the one or more machines to attempt to produce or manufacture at least one of the physical products or physical manufactures; responsive to completion of one or more operations associated with the activation by the one or more machines, acquiring digital image data of a portion of a physical object on or inside the one or more machines; analyzing the image data to form a digital fingerprint of the physical object, wherein the digital fingerprint is responsive to structure of the physical object; and storing the digital fingerprint in a database record of a database system.
Example A-12 includes the subject matter of example A-11 (and/or any other example herein), wherein the physical object comprises a component of the at least one of the physical products or physical manufactures.
Example A-13 includes the subject matter of any of examples A-11 through A-12 (and/or any other example herein), further comprising completing or manufacturing the at least one of the physical products or physical manufactures following the acquisition of the digital image data.
Example A-14 includes the subject matter of any of examples A-11 through A-13 (and/or any other example herein), wherein the record comprises a new record created responsive to forming the digital fingerprint of the physical object.
Example A-15 includes the subject matter of any of examples A-11 through A-14 (and/or any other example herein), further comprising preventing the at least one machine from producing or manufacturing the least one of the physical products or physical manufactures responsive to a result of the analysis of the digital image data.
Example A-16 includes the subject matter of any of examples A-11 through A-15 (and/or any other example herein), wherein the physical products or manufactures comprises printed documents.
Example A-17 includes the subject matter of any of examples A-11 through A-16 (and/or any other example herein), further comprising: subsequent to storing the digital fingerprint in the database record, identifying information corresponding to a target physical object, wherein identifying the first information corresponding to the target physical object includes acquiring second digital image data of a portion of the target physical object extracting features from the second digital image data to form a second digital fingerprint querying the database system to seek a matching record based on the second digital fingerprint; and in the case that a matching record is returned responsive to the querying, updating the matching record with an indication that the second digital fingerprint was matched to it.
Example A-18 includes the subject matter of any of examples A-11 through A-17 (and/or any other example herein), wherein the at least one machine comprises a fabrication device and the portion of the physical object comprises a completed surface.
Example A-19 includes the subject matter of any of examples A-11 through A-18 (and/or any other example herein), wherein the fabrication device comprises at least one of a three-dimension printer or a document printer.
Example A-20 is an apparatus for authenticated production. The apparatus may include one or more machines to produce or manufacture physical products or physical manufactures, the one or more machines including a processor configured to: identify an activation of at least one of the one or more machines to attempt to produce or manufacture at least one of the physical products or physical manufactures; and ascertain whether to permit the one or more machines to produce or manufacture the at least one of the physical products or physical manufactures; and a database system in communication with the first processor, the database system including a second digital fingerprint processor configured to: identify information about the attempt; signal the first processor if the attempt is not authorized based on the identified information about the attempt; responsive to completion of one or more operations associated with the activation by the one or more machines, analyze digital image data of a portion of a physical object on or inside the one or more machines to form a digital fingerprint of the physical object, wherein the digital fingerprint is responsive to structure of the physical object; and store the digital fingerprint in a database record of the database system.
Example A-21 is a method of authenticated production using one or more machines to produce or manufacture systems, the method comprising: identifying an activation of at least one machine of the one or more machines to attempt to produce or manufacture at least one of the systems using one or more parts; wherein the one or more parts are inducted into a database system, and wherein the database system includes a first database record including one or more digital fingerprints of the one or more parts, respectively; responsive to completion of one or more operations associated with the activation by the one or more machines, acquiring digital image data of a portion of a physical object associated with the system, the physical object having the one or more parts coupled thereto; analyzing the image data to form a digital fingerprint of the physical object, wherein the digital fingerprint is responsive to structure of the physical object; and performing at least one of storing the digital fingerprint in the first database record or storing the digital fingering in a second database record that is of the database system and linked to the first database record.
Example A-22 includes the subject matter of example A-21 (and/or any other example herein), further comprising: identifying an activation of the one or more machines to attempt to produce or manufacture at least one of the one or more parts; responsive to completion of one or more operations associated with the activation to attempt to produce or manufacture the at least one of the one or more parts, acquiring digital image data of a portion of a physical object associated with the at least one of the one or more parts; analyzing the image data to form a digital fingerprint of the physical object associated with the at least one of the one or more parts, wherein the digital fingerprint is responsive to structure of the physical object associated with the at least one of the one or more parts; and storing the digital fingerprint in the first database record of the database system.
Example A-23 includes the subject matter of any of examples A-21 through A-22 (and/or any other example herein), wherein the system comprises a computer, and the one or more parts comprise at least one of a circuit board, a drive, or a chip.
Example A-25 includes the subject matter of any of examples A-21 through A-24 (and/or any other example herein), further comprising: storing first metadata associated with the production or manufacture of the one or more parts in the first record; and storing second metadata associated with the production or manufacture of the at least one of the systems in at least one of the first database record or the second database record that is linked to the first database record.
Example A-26 includes the subject matter of any of examples A-21 through A-25 (and/or any other example herein), wherein the first metadata identifies a relative physical location of each one or more parts relative to the at least one system for authenticating the at least one system based an arrangement of the one or more parts.
Example A-27 is one or more memory storing instructions to be executed by one or more processors, wherein the instructions when executed perform operations corresponding to the steps of any of the examples A-11 through A-19 and A-21 through A-26 (and/or any other example herein).
Most of the equipment discussed above comprises hardware and associated software. For example, the typical electronic device is likely to include one or more processors and software executable on those processors to carry out the operations described. We use the term software herein in its commonly understood sense to refer to programs or routines (subroutines, objects, plug-ins, etc.), as well as data, usable by a machine or processor. As is well known, computer programs generally comprise instructions that are stored in machine-readable or computer-readable storage media Some embodiments of the present invention may include executable programs or instructions that are stored in machine-readable or computer-readable storage media, such as a digital memory. We do not imply that a “computer” in the conventional sense is required in any particular embodiment. For example, various processors, embedded or otherwise, may be used in equipment such as the components described herein.
Memory for storing software again is well known. In some embodiments, memory associated with a given processor may be stored in the same physical device as the processor (“on-board” memory); for example, RAM or FLASH memory disposed within an integrated circuit microprocessor or the like. In other examples, the memory comprises an independent device, such as an external disk drive, storage array, or portable FLASH key fob. In such cases, the memory becomes “associated” with the digital processor when the two are operatively coupled together, or in communication with each other, for example by an I/O port, network connection, etc. such that the processor can read a file stored on the memory. Associated memory may be “read only” by design (ROM) or by virtue of permission settings, or not. Other examples include but are not limited to WORM, EPROM, EEPROM, FLASH, etc. Those technologies often are implemented in solid state semiconductor devices. Other memories may comprise moving parts, such as a conventional rotating disk drive. All such memories are “machine readable” or “computer-readable” and may be used to store executable instructions for implementing the functions described herein.
A “software product” refers to a memory device in which a series of executable instructions are stored in a machine-readable form so that a suitable machine or processor, with appropriate access to the software product, can execute the instructions to carry out a process implemented by the instructions. Software products are sometimes used to distribute software. Any type of machine-readable memory, including without limitation those summarized above, may be used to make a software product. That said, it is also known that software can be distributed via electronic transmission (“download”), in which case there typically will be a corresponding software product at the transmitting end of the transmission, or the receiving end, or both.
Having described and illustrated the principles of the invention in a preferred embodiment thereof, it should be apparent that the invention may be modified in arrangement and detail without departing from such principles. We claim all modifications and variations coming within the spirit and scope of the following claims.
Number | Name | Date | Kind |
---|---|---|---|
4218674 | Brosow et al. | Aug 1980 | A |
4423415 | Goldman | Dec 1983 | A |
4677435 | Causse et al. | Jun 1987 | A |
4700400 | Ross | Oct 1987 | A |
4883971 | Jensen | Nov 1989 | A |
4921107 | Hofer | May 1990 | A |
5031223 | Rosenbaum et al. | Jul 1991 | A |
5079714 | Manduley et al. | Jan 1992 | A |
5393939 | Nasuta et al. | Feb 1995 | A |
5422821 | Allen et al. | Jun 1995 | A |
5514863 | Williams | May 1996 | A |
5518122 | Tilles et al. | May 1996 | A |
5521984 | Denenberg | May 1996 | A |
5703783 | Allen et al. | Dec 1997 | A |
5719939 | Tel | Feb 1998 | A |
5734568 | Borgendale et al. | Mar 1998 | A |
5745590 | Pollard | Apr 1998 | A |
5883971 | Bolle et al. | Mar 1999 | A |
5923848 | Goodhand et al. | Jul 1999 | A |
5974150 | Kaish et al. | Oct 1999 | A |
6205261 | Goldberg | Mar 2001 | B1 |
6246794 | Kagehiro et al. | Jun 2001 | B1 |
6292709 | Uhl et al. | Sep 2001 | B1 |
6327373 | Yura | Dec 2001 | B1 |
6343327 | Daniels et al. | Jan 2002 | B2 |
6360001 | Berger et al. | Mar 2002 | B1 |
6370259 | Hobson et al. | Apr 2002 | B1 |
6400805 | Brown et al. | Jun 2002 | B1 |
6424728 | Ammar | Jul 2002 | B1 |
6434601 | Rollins | Aug 2002 | B1 |
6470091 | Koga et al. | Oct 2002 | B2 |
6539098 | Baker et al. | Mar 2003 | B1 |
6549892 | Sansone | Apr 2003 | B1 |
6597809 | Ross et al. | Jul 2003 | B1 |
6643648 | Ross et al. | Nov 2003 | B1 |
6697500 | Woolston et al. | Feb 2004 | B2 |
6741724 | Bruce et al. | May 2004 | B1 |
6768810 | Emanuelsson et al. | Jul 2004 | B2 |
6778703 | Zlotnick | Aug 2004 | B1 |
6805926 | Cote et al. | Oct 2004 | B2 |
6816602 | Coffelt et al. | Nov 2004 | B2 |
6829369 | Poulin et al. | Dec 2004 | B2 |
6961466 | Imagawa et al. | Nov 2005 | B2 |
6985925 | Ogawa | Jan 2006 | B2 |
6985926 | Ferlauto et al. | Jan 2006 | B1 |
7016532 | Boncyk et al. | Mar 2006 | B2 |
7031519 | Elmenhurst | Apr 2006 | B2 |
7096152 | Ong | Aug 2006 | B1 |
7120302 | Billester | Oct 2006 | B1 |
7121458 | Want et al. | Oct 2006 | B2 |
7152047 | Nagel | Dec 2006 | B1 |
7171049 | Snapp | Jan 2007 | B2 |
7204415 | Payne et al. | Apr 2007 | B2 |
7212949 | Bachrach | May 2007 | B2 |
7333987 | Ross et al. | Feb 2008 | B2 |
7343623 | Ross | Mar 2008 | B2 |
7356162 | Caillon | Apr 2008 | B2 |
7379603 | Ross et al. | May 2008 | B2 |
7436979 | Bruce et al. | Oct 2008 | B2 |
7477780 | Boncyk et al. | Jan 2009 | B2 |
7518080 | Amato | Apr 2009 | B2 |
7602938 | Prokoski | Oct 2009 | B2 |
7674995 | Desprez et al. | Mar 2010 | B2 |
7676433 | Ross et al. | Mar 2010 | B1 |
7680306 | Boutant et al. | Mar 2010 | B2 |
7720256 | Desprez et al. | May 2010 | B2 |
7726457 | Maier et al. | Jun 2010 | B2 |
7726548 | Delavergne | Jun 2010 | B2 |
7748029 | Ross | Jun 2010 | B2 |
7822263 | Prokoski | Oct 2010 | B1 |
7834289 | Orbke et al. | Nov 2010 | B2 |
7853792 | Cowburn | Dec 2010 | B2 |
8022832 | Vogt et al. | Sep 2011 | B2 |
8032927 | Ross | Oct 2011 | B2 |
8108309 | Tan | Jan 2012 | B2 |
8180174 | Di et al. | May 2012 | B2 |
8180667 | Baluja et al. | May 2012 | B1 |
8194938 | Wechsler et al. | Jun 2012 | B2 |
8316418 | Ross | Nov 2012 | B2 |
8374020 | Katti | Feb 2013 | B2 |
8374399 | Talwerdi | Feb 2013 | B1 |
8374920 | Hedges et al. | Feb 2013 | B2 |
8391583 | Mennie et al. | Mar 2013 | B1 |
8428772 | Miette et al. | Apr 2013 | B2 |
8437530 | Mennie et al. | May 2013 | B1 |
8457354 | Kolar et al. | Jun 2013 | B1 |
8477992 | Paul et al. | Jul 2013 | B2 |
8520888 | Spitzig et al. | Aug 2013 | B2 |
8526743 | Campbell et al. | Sep 2013 | B1 |
8774455 | Elmenhurst et al. | Jul 2014 | B2 |
8959029 | Jones et al. | Feb 2015 | B2 |
9031329 | Farid et al. | May 2015 | B1 |
9058543 | Campbell et al. | Jun 2015 | B2 |
9152862 | Ross et al. | Oct 2015 | B2 |
9170654 | Boncyk et al. | Oct 2015 | B1 |
9224196 | Duerksen et al. | Dec 2015 | B2 |
9234843 | Sopori et al. | Jan 2016 | B2 |
9245133 | Durst et al. | Jan 2016 | B1 |
9350552 | Elmenhorst et al. | May 2016 | B2 |
9350714 | Freeman et al. | May 2016 | B2 |
9361507 | Hoyos et al. | Jun 2016 | B1 |
9361596 | Ross et al. | Jun 2016 | B2 |
9424461 | Yuan et al. | Aug 2016 | B1 |
9443298 | Ross et al. | Sep 2016 | B2 |
9558463 | Ross et al. | Jan 2017 | B2 |
9582714 | Ross et al. | Feb 2017 | B2 |
9646206 | Ross et al. | May 2017 | B2 |
9665800 | Kuffner | May 2017 | B1 |
9741724 | Seshadri et al. | Aug 2017 | B2 |
10037537 | Withrow et al. | Jul 2018 | B2 |
10043073 | Ross et al. | Aug 2018 | B2 |
10192140 | Ross et al. | Jan 2019 | B2 |
10199886 | Li et al. | Feb 2019 | B2 |
10346852 | Ross et al. | Jul 2019 | B2 |
10505726 | Andon et al. | Dec 2019 | B1 |
10540664 | Ross et al. | Jan 2020 | B2 |
10572883 | Ross et al. | Feb 2020 | B2 |
10614302 | Withrow et al. | Apr 2020 | B2 |
10621594 | Land et al. | Apr 2020 | B2 |
10740767 | Withrow | Aug 2020 | B2 |
10936838 | Wong | Mar 2021 | B1 |
20010010334 | Park et al. | Aug 2001 | A1 |
20010054031 | Lee et al. | Dec 2001 | A1 |
20020015515 | Lichtermann et al. | Feb 2002 | A1 |
20020073049 | Dutta | Jun 2002 | A1 |
20020134836 | Cash et al. | Sep 2002 | A1 |
20020168090 | Bruce et al. | Nov 2002 | A1 |
20030015395 | Hallowell et al. | Jan 2003 | A1 |
20030046103 | Amato et al. | Mar 2003 | A1 |
20030091724 | Mizoguchi | May 2003 | A1 |
20030120677 | Vernon | Jun 2003 | A1 |
20030138128 | Rhoads | Jul 2003 | A1 |
20030179931 | Sun | Sep 2003 | A1 |
20030182018 | Snapp | Sep 2003 | A1 |
20030208298 | Edmonds | Nov 2003 | A1 |
20030219145 | Smith | Nov 2003 | A1 |
20040027630 | Lizotte | Feb 2004 | A1 |
20040101174 | Sato et al. | May 2004 | A1 |
20040112962 | Farrall et al. | Jun 2004 | A1 |
20040218791 | Jiang et al. | Nov 2004 | A1 |
20040218801 | Houle et al. | Nov 2004 | A1 |
20040250085 | Tattan et al. | Dec 2004 | A1 |
20050007776 | Monk et al. | Jan 2005 | A1 |
20050038756 | Nagel | Feb 2005 | A1 |
20050065719 | Khan et al. | Mar 2005 | A1 |
20050086256 | Owens et al. | Apr 2005 | A1 |
20050111618 | Sommer et al. | May 2005 | A1 |
20050119786 | Kadaba | Jun 2005 | A1 |
20050125360 | Tidwell | Jun 2005 | A1 |
20050131576 | De et al. | Jun 2005 | A1 |
20050137882 | Cameron et al. | Jun 2005 | A1 |
20050160271 | Brundage et al. | Jul 2005 | A9 |
20050169529 | Owechko et al. | Aug 2005 | A1 |
20050188213 | Ku | Aug 2005 | A1 |
20050204144 | Mizutani | Sep 2005 | A1 |
20050251285 | Boyce et al. | Nov 2005 | A1 |
20050257064 | Boutant et al. | Nov 2005 | A1 |
20050289061 | Kulakowski et al. | Dec 2005 | A1 |
20060010503 | Inoue et al. | Jan 2006 | A1 |
20060083414 | Neumann et al. | Apr 2006 | A1 |
20060109520 | Gossaye et al. | May 2006 | A1 |
20060131518 | Ross et al. | Jun 2006 | A1 |
20060177104 | Prokoski | Aug 2006 | A1 |
20060253406 | Caillon | Nov 2006 | A1 |
20070056041 | Goodman | Mar 2007 | A1 |
20070071291 | Yumoto et al. | Mar 2007 | A1 |
20070085710 | Bousquet et al. | Apr 2007 | A1 |
20070094155 | Dearing | Apr 2007 | A1 |
20070211651 | Ahmed et al. | Sep 2007 | A1 |
20070211964 | Agam et al. | Sep 2007 | A1 |
20070230656 | Lowes et al. | Oct 2007 | A1 |
20070263267 | Ditt | Nov 2007 | A1 |
20070269043 | Launay et al. | Nov 2007 | A1 |
20070282900 | Owens et al. | Dec 2007 | A1 |
20080005578 | Shafir | Jan 2008 | A1 |
20080008377 | Andel | Jan 2008 | A1 |
20080011841 | Self et al. | Jan 2008 | A1 |
20080013804 | Moon et al. | Jan 2008 | A1 |
20080016355 | Beun et al. | Jan 2008 | A1 |
20080128496 | Bertranou et al. | Jun 2008 | A1 |
20080130947 | Ross et al. | Jun 2008 | A1 |
20080219503 | Di et al. | Sep 2008 | A1 |
20080250483 | Lee | Oct 2008 | A1 |
20080255758 | Graham et al. | Oct 2008 | A1 |
20080272585 | Conard et al. | Nov 2008 | A1 |
20080290005 | Bennett et al. | Nov 2008 | A1 |
20080294474 | Furka | Nov 2008 | A1 |
20090028379 | Belanger et al. | Jan 2009 | A1 |
20090057207 | Orbke et al. | Mar 2009 | A1 |
20090106042 | Maytal et al. | Apr 2009 | A1 |
20090134222 | Ikeda | May 2009 | A1 |
20090154778 | Lei et al. | Jun 2009 | A1 |
20090157733 | Kim et al. | Jun 2009 | A1 |
20090223099 | Versteeg | Sep 2009 | A1 |
20090232361 | Miller | Sep 2009 | A1 |
20090245652 | Bastos | Oct 2009 | A1 |
20090271029 | Doutre | Oct 2009 | A1 |
20090287498 | Choi | Nov 2009 | A2 |
20090307005 | Omartin et al. | Dec 2009 | A1 |
20100027834 | Spitzig et al. | Feb 2010 | A1 |
20100054551 | Decoux | Mar 2010 | A1 |
20100070527 | Chen | Mar 2010 | A1 |
20100104200 | Baras et al. | Apr 2010 | A1 |
20100157064 | Cheng et al. | Jun 2010 | A1 |
20100163612 | Caillon | Jul 2010 | A1 |
20100166303 | Rahimi | Jul 2010 | A1 |
20100174406 | Miette et al. | Jul 2010 | A1 |
20100286815 | Zimmermann | Nov 2010 | A1 |
20110026831 | Perronnin et al. | Feb 2011 | A1 |
20110064279 | Uno | Mar 2011 | A1 |
20110081043 | Sabol et al. | Apr 2011 | A1 |
20110091068 | Stuck et al. | Apr 2011 | A1 |
20110161117 | Busque et al. | Jun 2011 | A1 |
20110188709 | Gupta et al. | Aug 2011 | A1 |
20110194780 | Li et al. | Aug 2011 | A1 |
20110235920 | Iwamoto et al. | Sep 2011 | A1 |
20110267192 | Goldman et al. | Nov 2011 | A1 |
20120042171 | White et al. | Feb 2012 | A1 |
20120089639 | Wang | Apr 2012 | A1 |
20120130868 | Loeken | May 2012 | A1 |
20120177281 | Frew | Jul 2012 | A1 |
20120185393 | Atsmon et al. | Jul 2012 | A1 |
20120199651 | Glazer | Aug 2012 | A1 |
20120242481 | Gernandt et al. | Sep 2012 | A1 |
20120243797 | Di Venuto Dayer et al. | Sep 2012 | A1 |
20120250945 | Peng et al. | Oct 2012 | A1 |
20130110719 | Carter et al. | May 2013 | A1 |
20130162394 | Etchegoyen | Jun 2013 | A1 |
20130212027 | Sharma | Aug 2013 | A1 |
20130214164 | Zhang et al. | Aug 2013 | A1 |
20130273968 | Rhoads et al. | Oct 2013 | A1 |
20130277425 | Sharma et al. | Oct 2013 | A1 |
20130284803 | Wood et al. | Oct 2013 | A1 |
20140032322 | Schwieger et al. | Jan 2014 | A1 |
20140140570 | Ross et al. | May 2014 | A1 |
20140140571 | Elmenhurst et al. | May 2014 | A1 |
20140184843 | Campbell et al. | Jul 2014 | A1 |
20140201094 | Herrington et al. | Jul 2014 | A1 |
20140270341 | Elmenhurst et al. | Sep 2014 | A1 |
20140314283 | Harding | Oct 2014 | A1 |
20140380446 | Niu et al. | Dec 2014 | A1 |
20150058142 | Lenahan et al. | Feb 2015 | A1 |
20150067346 | Ross | Mar 2015 | A1 |
20150078629 | Gottemukkula et al. | Mar 2015 | A1 |
20150086068 | Mulhearn et al. | Mar 2015 | A1 |
20150110364 | Niinuma et al. | Apr 2015 | A1 |
20150117701 | Ross | Apr 2015 | A1 |
20150127430 | Hammer | May 2015 | A1 |
20150248587 | Oami et al. | Sep 2015 | A1 |
20150294189 | Benhimane et al. | Oct 2015 | A1 |
20150309502 | Breitgand | Oct 2015 | A1 |
20150371087 | Ross et al. | Dec 2015 | A1 |
20160034913 | Zavarehi et al. | Feb 2016 | A1 |
20160034914 | Gonen et al. | Feb 2016 | A1 |
20160055651 | Oami | Feb 2016 | A1 |
20160057138 | Hoyos et al. | Feb 2016 | A1 |
20160072626 | Kouladjie | Mar 2016 | A1 |
20160117631 | Mccloskey et al. | Apr 2016 | A1 |
20160162734 | Ross et al. | Jun 2016 | A1 |
20160180485 | Avila et al. | Jun 2016 | A1 |
20160180546 | Kim et al. | Jun 2016 | A1 |
20160189510 | Hutz | Jun 2016 | A1 |
20160203387 | Lee et al. | Jul 2016 | A1 |
20160300234 | Moss-Pultz et al. | Oct 2016 | A1 |
20160335520 | Ross et al. | Nov 2016 | A1 |
20170004444 | Krasko et al. | Jan 2017 | A1 |
20170032285 | Sharma et al. | Feb 2017 | A1 |
20170076132 | Sezan et al. | Mar 2017 | A1 |
20170132458 | Short et al. | May 2017 | A1 |
20170153069 | Huang et al. | Jun 2017 | A1 |
20170243230 | Ross et al. | Aug 2017 | A1 |
20170243231 | Withrow et al. | Aug 2017 | A1 |
20170243232 | Ross et al. | Aug 2017 | A1 |
20170243233 | Land et al. | Aug 2017 | A1 |
20170249491 | Macintosh et al. | Aug 2017 | A1 |
20170251143 | Peruch et al. | Aug 2017 | A1 |
20170253069 | Kerkar et al. | Sep 2017 | A1 |
20170295301 | Liu | Oct 2017 | A1 |
20170300905 | Withrow et al. | Oct 2017 | A1 |
20170344823 | Withrow et al. | Nov 2017 | A1 |
20170344824 | Martin | Nov 2017 | A1 |
20170372327 | Withrow | Dec 2017 | A1 |
20180000359 | Watanabe | Jan 2018 | A1 |
20180012008 | Withrow et al. | Jan 2018 | A1 |
20180018627 | Ross et al. | Jan 2018 | A1 |
20180018838 | Fankhauser et al. | Jan 2018 | A1 |
20180024074 | Ranieri et al. | Jan 2018 | A1 |
20180024178 | House et al. | Jan 2018 | A1 |
20180047128 | Ross et al. | Feb 2018 | A1 |
20180053312 | Ross et al. | Feb 2018 | A1 |
20180121643 | Talwerdi et al. | May 2018 | A1 |
20180144211 | Ross et al. | May 2018 | A1 |
20180315058 | Withrow et al. | Nov 2018 | A1 |
20180349694 | Ross et al. | Dec 2018 | A1 |
20190026581 | Leizerson | Jan 2019 | A1 |
20190034518 | Liu | Jan 2019 | A1 |
20190034694 | Ross | Jan 2019 | A1 |
20190102873 | Wang et al. | Apr 2019 | A1 |
20190102973 | Oyama et al. | Apr 2019 | A1 |
20190130082 | Alameh et al. | May 2019 | A1 |
20190228174 | Withrow et al. | Jul 2019 | A1 |
20190266373 | Hirokawa | Aug 2019 | A1 |
20190279017 | Graham et al. | Sep 2019 | A1 |
20190287118 | Ross et al. | Sep 2019 | A1 |
20190342102 | Hao et al. | Nov 2019 | A1 |
20190362186 | Irshad et al. | Nov 2019 | A1 |
20200153822 | Land et al. | May 2020 | A1 |
20200226366 | Withrow et al. | Jul 2020 | A1 |
20200233901 | Crowley et al. | Jul 2020 | A1 |
20200250395 | Ross et al. | Aug 2020 | A1 |
20200257791 | Shannon et al. | Aug 2020 | A1 |
20200334689 | Withrow | Oct 2020 | A1 |
20200349379 | Ross | Nov 2020 | A1 |
20200356772 | Withrow et al. | Nov 2020 | A1 |
Number | Date | Country |
---|---|---|
102006005927 | Aug 2007 | DE |
0439669 | Aug 1991 | EP |
0759596 | Feb 1997 | EP |
1016548 | Jul 2000 | EP |
1016549 | Jul 2000 | EP |
1719070 | Apr 2009 | EP |
2107506 | Oct 2009 | EP |
2166493 | Mar 2010 | EP |
2195621 | Nov 2013 | EP |
2866193 | Apr 2015 | EP |
2257909 | May 2015 | EP |
2869240 | May 2015 | EP |
2869241 | May 2015 | EP |
3208744 | Aug 2017 | EP |
3249581 | Nov 2017 | EP |
3267384 | Jan 2018 | EP |
3270342 | Jan 2018 | EP |
3435287 | Jan 2019 | EP |
3514715 | Jul 2019 | EP |
2097979 | Nov 1982 | GB |
2482127 | Jan 2012 | GB |
61234481 | Oct 1986 | JP |
H07192112 | Jul 1995 | JP |
2007213148 | Aug 2007 | JP |
2010146158 | Jul 2010 | JP |
20120009654 | Feb 2012 | KR |
2005086616 | Sep 2005 | WO |
2006038114 | Apr 2006 | WO |
2007028799 | Mar 2007 | WO |
2007031176 | Mar 2007 | WO |
2007071788 | Jun 2007 | WO |
2007090437 | Aug 2007 | WO |
2007144598 | Dec 2007 | WO |
2009030853 | Mar 2009 | WO |
2009089126 | Jul 2009 | WO |
2009115611 | Sep 2009 | WO |
2010018464 | Feb 2010 | WO |
2010018646 | Feb 2010 | WO |
2012145842 | Nov 2012 | WO |
2013051019 | Apr 2013 | WO |
2013126221 | Aug 2013 | WO |
2013173408 | Nov 2013 | WO |
2015004434 | Jan 2015 | WO |
2016081831 | May 2016 | WO |
Entry |
---|
Anonymous, “Intrinsic Characteristics for Authentication” & “Alp Vision Advances Security Through Digital Technology,” Authentication News Volume 12, (No. 9) pp. 2, 7 and 8, dated Sep. 2006, 3 pages total. |
Beekhof et al., “Secure Surface Identification Codes,” Proceeding of the SPIE 6819: Security Forensics, Steganography, and Watermarking of Multimedia Contents X:68190D, 2008. (12 pages). |
Boa et al., “Local Feature based Multiple Object Instance Identification using Scale and Rotation Invariant Implicit Shape Model,” 12th Asian Conference on Computer Vision, Singapore, Singapore, Nov. 1-5, 2014, pp. 600-614. |
Buchanan et al., “Fingerprinting documents and packaging,” Nature 436 (7050): 475, 2005. |
Cavoukian et al. “Biometric Encryption: Technology for Strong Authentication, Security and Privacy,” 2008, WE, Intl. Fed. Iot Info Processing, vol. 261; Policies and Research in Identity Management; pp. 57-77. |
Clifton Smith; “Fireball: A Forensic Ballistic Imaging System”; Security Science, Edith Cowan University; IEEE; 1997. |
Di Paola et al., “An Autonomous Mobile Robotic System for Surveillance of Indoor Environments,” International Journal of Advanced Robotic Systems 7(1): 19-26, 2010. |
Drew, M. S., et al., “Sharpening from Shadows: Sensor Transforms for Removing Shadows using a Single Image,” Color and Imaging Conference, vol. 5, Society for Imaging Science and Technology, 2009, pp. 267-271. |
Ebay, “eBay Launches Must-Have !Phone App Red Laser 3.0” published Nov. 18, 2011; https://www.ebayinc.com/stories/news/ ebay-launches-must-have-iphone-app-redlaser-30/, downloaded Mar. 21, 2019, 7 pages). |
Entropy.com Website History, Wayback Machine; https://web.archive.org/web/20 I 60330060808/https://www.entrupy.com/; Mar. 30, 2016 (Year: 2016), 2 pages. |
Extended European Search Report, dated Oct. 26, 2017, for European Application No. 17179894.5-1871, 8 pages. |
Farid, “Digital Image Forensics”, Dartmouth CS 89/189, Sprint 2013; 199 pages. |
Fischler et al., “Random Sample Consensus: A Paradigm for Model Fitting with Applications to Image Analysis and Automated Cartography,” Communication of the ACM 24(6); 381-395, 1981. |
Huang, et al., “An Online Ballistics Imaging System for Firearm Identification”; 2010 2nd International Conference on Signal Processing Systems (ICSPS). |
Kartik et al., “Security System with Face Recognition, SMS Alert and Embedded Network Video Monitoring Terminal,” International Journal of Security, Privacy and Trust Management 2(5):9-19, 2013. |
Leng, et al., “A Novel Binarization Algorithms for Ballistics Imaging Systems”; 2010 3rd International Congress on mage and Signal Processing (CISP2010). |
Li, “Image Processing for the Positive Identification of Forensic Ballistics Specimens,” Proceedings of the 6th International Conference of Information Fusion, Caims, Australia, Jul. 8-11, 2003, pp. 1494-1498. |
Li; “Firearm Identification System Based on Ballistics Image Processing”; 2008 Congress on Image and Signal Processing. |
Maddern et al., “Illumination Invariant Imaging: Applications in Robust Vision-based Localization, Mapping and Classification for Autonomous Vehicles,” IEEE International Conference on Robotics and Automation, Hong Kong, May 31-Jun. 7, 2014, 8 pages. |
Matsumoto et al., “Nano-artifact metrics based on random collapse of resist,” Scientific Reports 4:6142,2014 (5 pages). |
Mistry et al., “Comparison of Feature Detection and Matching Approaches: SIFT and SURF,” Global Research and Development Journal for Engineering, vol. 2, Issue 4, Mar. 2017, 8 pages. |
NCOA Link at http:/ /ribbs.usps.gov/ncoalink/ncoalink_print.htm; dated May 27, 2009; 3 pages. |
Online NCOALink Processing Acknowledgement Form (PAF) Released by Lorton Data; http:/ /us.generation-nt.com/online-ncoalink-processing-acknowledgement-form-paf-released- by-press-1567191 .html; release dated Jun. 2, 2009; 1 page. |
Rublee et al., “ORB: an efficient alternative to SIFT or SURF,” IEEE International Conference on Computer Vision, Barcelona, Spain, Nov. 6-13, 2011, 8 pages. |
Schneider et al., “A Robust Content Based Digital Signature for Image Authentication,” Proceeding of the International Conference on Image Processing Lausanne, Switzerland, Sep. 19, 1996, pp. 227-230. |
Schwabe Williamson & Wyatt, PC—Listing of Related Cases; dated Sep. 16, 2017; 2 pages. |
Sharma et al., “The Fake vs Real Goods Problem: Microscopy and Machine Learning to the Rescue,” KDD 2017 Applied Data Science Paper, Aug. 13-17, 2017, Halifax, NS, Canada, 9 pages. |
Shi et al., “Smart Cameras: Fundamentals and Classification,” Chapter 2, Belbachir (ed.), Smart Cameras, Springer, New York, New York, USA 2010, pp. 19-34. |
Shields, “How To Shop Savvy With Red Laser,” published online on Mar. 22, 2010; https ://i phone appstomn .net/reviews/lifesty le/how-to-shop-savvy-with-redlaser/, downloaded Mar. 22, 2010, 8 pages). |
Takahashi et al., “Mass-produced Parts Traceability System Based on Automated Scanning of Fingerprint of Things,” 15th IAPR International Conference on Machine Vision Applications, Nagoya, Japan, May 8-12, 2017, 5 pages. |
United States Postal Service Publication 28 “Postal Addressing Standards”, dated Jul. 2008; text plus Appendix A only; 55 pages. |
United States Postal Service, “NCOALink Systems”, http://www.usps.com/ncsc/addressservices/moveupdate/changeaddress.htm, website accessed Jun. 23, 2010, 2 pages. |
Veena et al., “Automatic Theft Security System (Smart Surveillance Camera),” Computer Science & Information Technology 3:75-87, 2013. |
Woods, “Counterfeit-spotting truth machine launches out of Dumbo,” published online on Feb. 11, 2016, downloaded from http://technically/brooklyn/2016/02/11/entrupy-counterfeit-scanner/ on Mar. 20, 2019, 3 pages. |
Extended European Search Report, dated Aug. 18, 2021, for European Application No. 21164207.9—17 pages. |
Extended European Search Report, dated Aug. 18, 2021, for European Application No. 21164207.9, 13 pages. |
Extended European Search Report, dated Aug. 19, 2021, for European Application No. 21164353.1, 9 pages. |
Non-Final Office Action Issued in U.S. Appl. No. 16/553,943, dated Sep. 1, 2021, 13 pages. |
Non-Final Office Action Issued in U.S. Appl. No. 16/827,701, dated Aug. 17, 2021, 19 pages. |
Non-Final Office Action Issued in U.S. Appl. No. 16/866,468, dated Sep. 9, 2021, 24 pages. |
Non-Final Office Action Issued in U.S. Appl. No. 16/872,202, dated Sep. 24, 2021, 24 pages. |
Hensler, J., et al., “Hybrid Face Recognition Based on Real-time Multi-camera Stereo-Matching”, ICIAP: International Conference on Image Analysis and Processing, 17th International Conference, Naples, Italy, Sep. 9-13, 2013, 10 pages. |
Scott, Von Duhn , et al., “Three-View Surveillance Video Based Face Modeling For Recognition”, Biometrics Symposium, 2007, IEEE, PI, Sep. 30, 2007, 6 pages XP031202430. |
Extended European Search Report Application No. 21153877.2, dated: Jun. 15, 2021, 8 pages. |
Extended European Search Report, dated Jun. 18, 2021, for European Application No. 21153355.9, 8 pages. |
Huang et al., “A Novel Binarization Algorithm for Ballistic Imaging Systems,” 3rd International Congress on Image and Signal Processing, Yantai, China, Oct. 16-18, 2010, pp. 1287-1291. |
Farid, Ahmed , et al., “Integrated fingerprint verification method using a composite signature-based watermarking technique”, Optical Engineering, The Catholic University of America, (Year: 2007), 6 pages. |
Jain, Anil K, et al., “Biometric Cryptosystems: Issues and Challenges”, Proceedings of the IEEE, IEEE, New York, US, vol. 92, No. 6, Jun. 1, 2004, XP011112757, pp. 948-960. |
Zaeri, Naser , “Minutiae-based Fingerprint Extraction and Recognition, 2020 (year 2010)”, 47 pages. |
Number | Date | Country | |
---|---|---|---|
20210141886 A1 | May 2021 | US |
Number | Date | Country | |
---|---|---|---|
62358506 | Jul 2016 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 15641157 | Jul 2017 | US |
Child | 17153231 | US |