Large volumes of digital documents are generated every moment in the information era. Digitalized documents make it much easier for different people to coordinate and share information. Publishers are producing more content in digital forms; eBooks are becoming increasingly popular among readers; enterprises are becoming paper-less in their daily operations. Digitalizing documents also creates the threat where valuable documents can be easily copied and subject to unauthorized use. The problem is particular serious for enterprises, where their documents are often proprietary and copyright sensitive. Unauthorized copying and using of those documents can cause serious economic damages to the document owners.
Document fingerprinting is the technique to extract features from large sets of documents (referred to as the corpus), so that copying, including partial copies of documents, can be identified with high confidence using those features. A document fingerprint system can be applied to search similar documents inside a large corpus, detect plagiarism of existing documents, detect leakage of sensitive documents, etc. Corpus is all the writings or works of a particular kind or on a particular subject, such as the complete works of an author. Throughout this disclosure, the terms “corpus” and “library” may be used interchangeably depending on the context.
An inverted index (also referred to as postings file or inverted file) is an index data structure storing a mapping from content, such as words or numbers, to its locations in a database file, or in a document or a set of documents. The inverted index data structure is a central component of a typical full text search system. A goal of a search engine implementation is to optimize the speed of the query: find the documents where word X occurs. Once a forward index is developed, which stores lists of words per document, it is next inverted to develop an inverted index. Querying the forward index would require sequential iteration through each document and to each word to verify a matching document. The time, memory, and processing resources to perform such a query are not always technically realistic. Instead of listing the words per document in the forward index, the inverted index data structure is developed which lists the documents per word.
In general, in one aspect, the invention relates to a method for comparing documents. The method includes extracting, by a computer processor, a plurality of extracted elements from a first formatted document, wherein each of the plurality of extracted elements corresponds to a text element of the first formatted document, extracting, by the computer processor, a first plurality of text fingerprints from a sequence of the plurality of extracted elements to form a first text feature of the first formatted document, comparing, by the computer processor, the first text feature and a second formatted document to generate a comparison result, and determining, in response to the comparison result meeting a pre-determined criterion, that each of the first formatted document and the second formatted document contains a common text content.
Other aspects and advantages of the invention will be apparent from the following description and the appended claims.
Specific embodiments of the invention will now be described in detail with reference to the accompanying figures. Like elements in the various figures are denoted by like reference numerals for consistency.
In the following detailed description of embodiments of the invention, numerous specific details are set forth in order to provide a more thorough understanding of the invention. In other instances, well-known features have not been described in detail to avoid obscuring the invention.
Embodiments of the invention provide a method and a system for fingerprinting formatted documents with mostly textual content. Examples of formatted textual documents include paper documents as well as electronic documents, such as those in the MS Word, PowerPoint, PDF, or other suitable electronic formats. In one or more embodiments, the document fingerprinting addresses the threat where malicious users uses a PrintScreen function to copy a displayed document, uses a camera to take a picture of a displayed or printed document, or uses a photo copier to copy a printed document. As is known to those skilled in the art, the PrintScreen function is activated by a computer user command to copy whatever is currently on the display screen to a data buffer. The copied content in the data buffer can then be printed as a hardcopy or converted to an electronic file. While the camera and the photo copier may use optical scanning technique, the PrintScreen function may be considered as an electronic scanning technique where the displayed content on the computer screen is copied by scanning a display buffer driving the computer screen.
In addition, the malicious user may also uses an optical character recognition (OCR) tool to capture an OCR result of these unauthorized copies. The malicious user then illegally distributes the document content via the copied images and/or the OCR results to other unauthorized third parties.
In one or more embodiments, two types of features of text information in a document are used to extract the unique and reliable fingerprints of the document. Those two features are referred to as text content feature and text geometric feature. The text content feature captures the information of what are the words used in a document and how those words are used to compose phrases and sentences; the text geometric feature captures the information of how the words are physically arranged on each page into lines of the document. In one or more embodiments, whether using the content feature or the geometric feature, a set of n-grams are extracted from the text information feature as the fingerprints of a document. In one or more embodiments, when building the fingerprint of a document via its text content feature, a “gram” is a word in the document; when building fingerprint via its text geometric feature, a “gram” is an integer number representing the length of a word normalized by the length of the line where that word belongs to. For example, the word length and line length may be measured by number of pixels when the document is shown on screen.
In one or more embodiments, the features of all documents in a corpus are extracted, these extracted features are used to determine whether a document A is a copy of another document B in the corpus. Specifically, the text content and geometric features of document A are extracted and then matched to the features of documents in the corpus. In one or more embodiments, searching for a matched document in a corpus is based on performing lookup in the inverted indices that map features to documents in the corpus. Accordingly, the inverted index associated with the documents in the corpus is used to search for another document B in the corpus that is very similar to document A. In one or more embodiments, a document searching technique via inverted index known to those skilled in the art may be used.
In one or more embodiments, two documents A and B are compared by comparing their extracted features to determine if they share a common portion (i.e., a common text content).
A shown in
Further as shown in
Further as shown in
In one or more embodiments, the text analyzer (111) is configured to extract a set of extracted elements (e.g., extracted elements A (150a)) from a first image (e.g., image A (101a)) of a first formatted document (e.g., formatted document A (101)). In particular, each element of the extracted elements A (150a) corresponds to a text element (e.g., text element A (101c)) of the formatted document A (101). For example, the text element A (101c) may be an alphanumeric character, a syllable, a word, a phrase, etc. in the formatted document A (101).
In one or more embodiments, the text analyzer (111) includes a text content analyzer (112) that is configured to extract words (151a) from the image A (101a). In one or more embodiments, the text analyzer (111) includes a text geometric analyzer (113) that is configured to extract word lengths (151b) from the image A (101a). In one or more embodiments, the text geometric analyzer (113) is further configured to extract a line length (not shown) and normalize each of the word lengths (151b) based on the corresponding line length where the particular word length belongs.
Although not explicitly shown, the extracted elements B (150b) may be extracted using the text analyzer (111) in a similar manner and contains similar words and/or word lengths as those in the extracted elements A (150a).
In one or more embodiments, the text analyzer (111) uses an optical character recognition (OCR) module (e.g., a software algorithm or a hardware unit that is not explicitly shown) to extract the words (151a) and/or word lengths (151b) from the image A (101a). The accuracy of the text analyzer (111) to perform such extractions may be modeled by an OCR error rate model. In one or more embodiments, the OCR error rate model describes the text content OCR error rate for accurately extracting a consecutive segment of the words (151a), such as a n-word segment (i.e., n-gram of words). In one or more embodiments, the OCR error rate model describes the text geometry error rate for extracting a consecutive segment of the word lengths (151b), such as an n-word-length segment (i.e., n-gram of word lengths). For example, the error rate may be a function of the size of the segment (i.e., number of words or word lengths in the n-gram). Typically, the longer the n-gram of words, the higher the error rate in accurately extracting all of the words in the n-gram from the image A (101a).
In one or more embodiments, the fingerprint extractor (121) is configured to extract a set of text fingerprints (e.g., text content fingerprint (161a), text geometric fingerprint (162a), etc.) from a sequence of the extracted elements A (150a) (e.g., a sequence in the words (151a), a sequence in the word lengths (151b)) to form the text feature A (160a) (i.e., text content feature (161) and/or text geometric feature (162)) of the image A (101a).
In one or more embodiments, the fingerprint extractor (121) includes a text content fingerprint extractor (122) that is configured to extract a segment of consecutive words from the words (151a) to form the text content fingerprint (161a). In one or more embodiments, the fingerprint extractor (121) includes a text geometric fingerprint extractor (123) that is configured to extract a segment of consecutive word lengths from the word lengths (151b) to form the text geometric fingerprint (162a). In one or more embodiments, the text geometric fingerprint (162a) is based on normalized word lengths by at least dividing each of the word lengths (151b) by a respective line length where the particular word length belongs. In one or more embodiments, the normalized word lengths are generated using further scaling and rounding operations. An example word length calculation is shown in the equation (5) described later in this disclosure. Using normalized word length based on the respective line length reduces document comparison error due to document image size differences. Additional details of such extractions are described in reference to FIGS. 2 and 3A-3E below.
In one or more embodiments, text content fingerprint (161a) and/or text geometric fingerprint (162a) include n-grams. In one or more embodiments, the fingerprint extractor (121) is further configured to identify the aforementioned OCR error rate model of the text analyzer (111). Generally, the text content OCR error rate is higher than the text geometry error rate for same size segments (e.g., corresponding to same number of words). As noted above, the OCR error rate may be a function of the particular length of the segment of the words (151a) and/or the word lengths (151b) to form the respective fingerprint. Accordingly, the fingerprint extractor (121) determines a length of the n-gram to optimize the document comparison accuracy by balancing the OCR error rate (based on the OCR error rate model) and a document matching error rate (based on a document matching error rate model that will be described later). In general, the optimal n for document comparison may be different depending on whether the comparison is based on the text content feature (161) or the text geometric feature (162).
Although not explicitly shown, the text feature B (160b) may contain similar text content fingerprint and/or text geometric fingerprint as those in the text feature A (160a). In one or more embodiments, the text feature B (160b) is extracted from the image B (102a) in a similar manner as how the text feature A (160a) is extracted from the image A (101a). In one or more embodiments, the text feature B (160b) is extracted directly from the formatted document B (102) when it is in a text format.
In one or more embodiments, the comparison module (131) is configured to compare the text feature A (160a) extracted from the image A (101a) and the formatted document B (102) in the formatted document library (110) to generate a comparison result. For example, the comparison result may be generated during search for a document in the formatted document library (110) that is most similar to the formatted document A (101). In one or more embodiments, the comparison and/or the search is based on inverted index search. Specifically, an inverted index data structure of the formatted document library (110) is formed from text features (e.g., text feature B (160b)) of all documents (e.g., the formatted document B (102)) in the formatted document library (110). In one or more embodiments, the inverted index data structure includes an inverted index for each text fingerprint extracted from any document in the formatted document library (110). The inverted index maps a particular text fingerprint to any document containing the particular text fingerprint and is tagged with a tally of how many times the particular text fingerprint occurring in the mapped document. For example, the text feature B (160b) may include a set of n-grams and a tally for each n-gram occurring in the image B (102a) or occurring directly in the formatted document B (102). This tally is included in the inverted index data structure to tag the inverted index mapping each n-gram in the text feature B (160b). In one or more embodiments, the comparison module (131) is configured to calculate, for each n-gram in the text feature A (160a), a frequency of occurrence for that n-gram occurring in the image B (102) based on inverted index look up of the formatted document library (110) using the inverted index data structure, in particular the portion of the inverted index data structure contributed by the text feature B (160b). In other words, the document comparison result is based at least on the frequency of how often each n-gram in the text feature A (160a) occurring in the image B (102) that is calculated using the text feature B (160b) and the fingerprint (e.g., a tally of a n-gram) tallies therein.
In one or more embodiments, the comparison module (131) is further configured to determine, in response to the comparison result meeting a pre-determined criterion, that the formatted document B (102) is the most similar document in the formatted document library (110) as compared to the formatted document A (101) (or the image A (101a)). Accordingly, each of the formatted document A (101) (or the image A (101a)) and the formatted document B (102) is determined to contain a common text content. In one scenario, the formatted document A (101) and the formatted document B (102) are the same document. In another scenario, the formatted document A (101) may be a superset or a subset of the formatted document B (102). The accuracy of such determination may be modeled by a document matching error rate model that describes the document matching error rate as a function of various parameters, such as the size of the n-gram (i.e., fingerprint) in the text feature A (160a) and text feature B (160b). As noted above, this fingerprint comparison error rate model may be used by the fingerprint extractor (121) to determine a length of the n-gram that optimizes the document comparison accuracy by balancing the OCR error rate and the document matching error rate.
In one or more embodiments, the comparison module (131) is further configured to divide the n-grams in the text feature A (160a) into subsets, such that the inverted index search is performed by concurrently performing subsets searches. For example, the concurrent subsets searches may include at least a first comparison subset search between a first subset of n-grams in the text feature A (160a) and the formatted document B (102) using the inverted index data structure, and a second comparison between a second subset of n-grams in the text feature A (160a) and the formatted document B (102) using the inverted index data structure. In one or more embodiments, the inverted index data structure may also be divided into subsets for performing the comparisons in an even more concurrent manner.
Additional details of calculating the frequency of occurrence, generating the comparison result from the frequency of occurrence, and making the determination from the comparison result are described in reference to FIGS. 2 and 3A-3E below.
In one or more embodiments, the formatted document comparison system (180) is based on text content comparison where the text geometric analyzer (113), the text geometric fingerprint extractor (123), the word lengths (151b), and the text geometric feature (162) are not included. In one or more embodiments, the formatted document comparison system (180) is based on text geometric comparison where the text content analyzer (112), the text content fingerprint extractor (122), the words (151a), and the text content feature (161) are not included.
In one or more embodiments, the aforementioned elements of the formatted document comparison system (180) may include hardware, software, programmable logic, or combinations thereof. For example, one or more of the text analyzer (111), fingerprint extractor (121), and comparison module (131) may include software instructions executing on a hardware processor (not shown) of the formatted document comparison system (180).
The flowchart depicted in
Initially in Step 201, a set of extracted elements are extracted from an image (referred to as the first image) of a formatted document (referred to as the first formatted document). In one or more embodiments, the first image is a photo image or a scanned image of the first formatted document. For example, the first image may be a result of photographing or scanning a displayed copy or a printed copy of an electronic document. In another example, the first image may be a result of photographing or scanning a paper document. In one or more embodiments, the first image is a result of un-authorized photographing, scanning, or otherwise copying the displayed/printed version of the electronic document or the paper document itself. Note that the electronic document may itself be in an image format, in which case the first image may be the original image of the electronic document or a copy thereof. As noted above, depending on the context, the term “document image” may refer to a photo/scanned image of a formatted document, the original image of a formatted document if it is already in an image format, or a further copy of such original image. Further, the terms “document,” “document image,” and “document in image format” may be used interchangeably depending on the context.
In one or more embodiments, the extracted elements are extracted using optical character recognition (OCR) technique and each of the extracted elements corresponds to a text element of the first formatted document. For example, the text element may be an alphanumeric character, a word, a phrase, a paragraph, etc. In one or more embodiments, each extracted element is an OCR extracted word corresponding to a word (i.e., a text element) in the first formatted document. In one or more embodiments, each extracted element is the length of an OCR extracted word (referred to as an OCR extracted word length) corresponding to a word (i.e., a text element) in the first formatted document. One skilled in the art will appreciate that there is a non-zero error rate associated with any practical OCR technique. Further, using any practical OCR technique, the reliability or accuracy of the OCR extracted word, as compared to the corresponding word in the first formatted document, is typically lower than the reliability or accuracy of the OCR extracted word length. In other words, an OCR tool typically have errors in the extracted n-gram and the longer the n-gram is, the less probability the OCR tool can correctly extract all characters in the n-gram. However, even when the n-gram may contain some incorrectly extracted character(s), the lengths of words in the n-gram often are still correct.
In Step 202, a set (referred to as the first set) of text fingerprints are extracted from a sequence of set of the extracted elements to form a first text feature of the first image. In the embodiments where each extracted element is an OCR extracted word, the first text feature is a text content feature and each text fingerprint in the text content feature is a text content fingerprint. In one or more embodiments, the text content fingerprint is based on a segment of consecutive words (i.e., a sequence) from all the OCR extracted words of the first image. In the embodiments where each extracted element is the length of an OCR extracted word, the first text feature is a text geometric feature and each text fingerprint in the text geometric feature is a text geometric fingerprint. In one or more embodiments, the text geometric fingerprint is based on a segment of consecutive word lengths (i.e., a sequence) from all the OCR extracted word lengths of the first image. Examples of the text content feature including multiple text content fingerprints and the text geometric feature including multiple text geometric fingerprints are described below as well as in reference to
In Step 203, the first text feature of the first image and a second formatted document are compared to generate a comparison result. In one or more embodiments, the second formatted document is from a document library containing multiple documents from which a search is performed for the most similar one to the first image and/or the first formatted document. For example, the comparison result may be generated during the search for this most similar document. In one or more embodiments, the comparison and/or the search is based on inverted index search. Specifically, an inverted index data structure of the document library is formed from text features (e.g., a second text feature of the second formatted document) of all documents (e.g., the second formatted document) in the document library. In one or more embodiments, the inverted index data structure includes an inverted index for each text fingerprint extracted from any document in the document library. The inverted index maps a particular text fingerprint to any document containing the particular text fingerprint and is tagged with a tally of how many times the particular text fingerprint occurring in the mapped document. For example, the second text feature may include a set of n-grams and a tally for each n-gram occurring in the second formatted document or a second image thereof. This tally is included in the inverted index data structure to tag the inverted index mapping each n-gram in the second text feature. In one or more embodiments, the comparison result is generated by calculating, for each n-gram in the first text feature, a frequency of occurrence for that n-gram occurring in the second image based on inverted index look up of the document library using the inverted index data structure, in particular the portion of the inverted index data structure contributed by the second text feature. In other words, the document comparison result is based at least on the frequency of how often each n-gram in the first text feature occurring in the second image that is calculated using the second text feature and the fingerprint (e.g., a tally of an n-gram) tallies therein.
In one or more embodiments, the second text feature may be previously generated directly from the second formatted document that is an electronic document. In one or more embodiments, the second text feature may be previously generated from a second image of the second formatted document.
In one or more embodiments, the comparison generates a text feature similarity score between the first text feature and the second text feature. For example, the text feature similarity score may be based on a frequency of occurrence of each text fingerprint of the first text feature also occurring in the second formatted document. In one or more embodiments, the frequency of occurrence is calculated based on the inverted index search described above. Accordingly, the comparison result is based at least on the frequency of occurrence.
In Step 204, in response to the comparison result meeting a pre-determined criterion, it is determined that the second formatted document is the most similar document in the document library as compared to the first formatted document. Accordingly, each of the first formatted document and the second formatted document is determined to contain common text content. For example, the first formatted document and the second formatted documents may be the same. In another example, the first formatted document may be a subset or a superset of the second formatted document.
In Step 205, a determination is made as to whether the first formatted document and the second formatted document are to be compared again based on a different type of extracted elements. For example, the Steps 201 through 204 may have been performed in an initial iteration using OCR extracted words as the extracted elements. In this example, the determination may be whether the comparison is to be performed again using OCR extracted word lengths (e.g., normalized word lengths that are normalized based on the line length where the particular word length belongs) as the extracted elements. In one or more embodiments, the determination may be based on an error rate of the OCR word extraction. For example, the error rate may be determined by matching all OCR extracted words to a dictionary of valid words.
If the match rate exceeds a pre-determined threshold (e.g., 99.9%, 90%, 80%, etc.), the determination is no and the method ends. Said in other words, the match rate exceeding the pre-determined threshold indicates that the OCR word extraction is sufficiently accurate for an acceptable document comparison result. Accordingly, no more comparison needs to be performed further.
If the match rate is less than the pre-determined threshold (e.g., 99.9%, 90%, 80%, etc.), the determination is yes and the method returns to Step 201. Said in other words, the match rate being less than the pre-determined threshold indicates that the OCR word extraction is not accurate enough for an acceptable document comparison result. Accordingly, further comparison needs to be performed. For example, the Steps 201 through 204 may be performed in a second iteration using OCR extracted word lengths as the extracted elements. In one or more embodiments, text geometric features are only extracted in response to determining that the text content match rate to the valid word library is less than the pre-determined threshold.
An example of performing Steps 201 through 204 is described below using either OCR extracted words or the OCR extracted word lengths as the extracted elements.
In one or more embodiments, Cdoc represents the set of documents (i.e., the corpus) from which their fingerprints are extracted. In one or more embodiments, given a document
In one or more embodiments, the text feature is the text content feature. As noted above, the text content feature extracted from a document d is the sequence of words in the document, represented by Wd=(w1d, w2d, . . . , wxd), where wid is the ith word in document d. After extracting Wd, a set of n-grams is computed from Wd where an n-gram has n words appearing consecutively in Wd. For example, the set of 2-grams for Wd=(s1d, w2d, w3d, w4d) is {(w1d, w2d)(w2d, w3d)(w3d, w4d)}. Each n-gram here is referred to as a text content fingerprint of document d. The list of text content fingerprints of document d, represented by Fd=(f1d, f2d, . . . , fx−n+1d), is the text content feature of document d. In one or more embodiments, an inverted index data structure (also referred to simply as the reverse index) of inverted indices is built for mapping those text content fingerprints (i.e., n-grams of words) of d. In one or more embodiments, to reduce the number of unique fingerprints saved in the inverted index data structure, the sequence of words Wd are preprocessed by removing pre-determined “stop words” and stemming the remaining words before computing the text content fingerprints. After processing all documents in corpus Cdoc, an inverted index of Cdoc is generated for each text content fingerprint (i.e., n-gram of words) of all documents to map the particular text content fingerprint to document(s) containing the particular text content fingerprint in corpus Cdoc.
To determine whether a document
where
TF(fi
DF(fi
In one or more embodiments, TF(fi
In one or more embodiments, document
Fingerprinting documents via their textual content features works well if extracting text from document
Generally, extracting word length in the document image is more reliable than recognizing the actual text. In one or more embodiments, the text geometric feature is used as a higher level document feature to pinpoint a document image as a copy or a partial copy of another document within the corpus. For example, after segmenting the document image into text lines and words, the text geometric feature may be generated based on the length of each word in the document image, as measure by the number of pixels from the left side of a word to the right side of the word. In one or more embodiments, when the OCR error rate in recognizing actual text from the document image is below a pre-determined threshold, the text content feature is not used for document comparison. Instead, the text geometric feature is used by configuring the OCR to recognize geometric regions (e.g., text lines and text words) and associated geometric sizes in the document image.
In one or more embodiments, the word length in a document image is normalized based on the length of the text line it belongs to. Given a formatted document d and each word in word sequence Wd=(w1d, w2d, . . . wxd), the normalized length of word widεWd, lid is computed as:
where WordLength (wid) represents the physical length of word wid on the document image, LineLength (wid) represents the physical length of the text line where word wid belongs to on the document image, M is a user defined or otherwise pre-determined integer, and function Rnd(x) rounds x to its closest integer.
The sequence of normalized length of all words in document d is represented by (l1d, l2d, . . . , lxd). An n-gram computation is then applied to this integer sequence. In one or more embodiments, a text geometric fingerprint fid is an n-gram (lid, di+1d, . . . , li+n−1d). The list of text geometric fingerprints computed from document d, denoted by Fd=(f1d, f2d, . . . , fkd) is the text geometric feature of document d. Feature Fd is then added into a text geometric inverted index of Cdoc that maps text geometric fingerprints (i.e., n-grams of normalized word lengths) to documents containing those text geometric fingerprints for all documents in corpus Cdoc.
To determine whether a document
where σ is a predefined small number referred to as fuzzy range, and is the set of integers.
In one or more embodiments, a text geometric similarity score is used to identify the document in Cdoc that is most similar to
In one or more embodiments, the document
In one or more embodiments, the text fingerprints are based on n-grams extracted using OCR techniques. On one side assuming all n-gram fingerprints are accurately extracted from the document, document comparison accuracy becomes better as the length of n-gram increases. On the other side, the larger the parameter n is, the more likely an n-gram may be different from its original value due to possible OCR errors. When computing the text content similarity between document d and its copy
Considering both these conflicting effects of the length of the n-gram on the document comparison accuracy versus the n-gram extraction accuracy, a compromised n-gram length may be determined to optimize the overall compounded accuracy based on an OCR error rate model. The OCR error rate model represents error rate of an optical character recognition (OCR) module used to extract the extracted elements from the document image.
The OCR error rate is defined as the probability that a character in the extracted text is different from the actual character. In one or more embodiments, a simplified error rate model assumes that the probability of each character to be in is the same. In one or more embodiments, a more sophisticated error rate model may define variable probability of a particular character to be in error depending on which character it is, the position of the particular character in the document, the surrounding characters next to the particular document, and other relevant parameters.
In one or more embodiments, text feature comparison is performed in a concurrent manner. Specifically, the dividing the set of text fingerprints is divided into subsets for performing concurrent comparisons. In other words, the comparison of the first text feature and the second text feature is partitioned into multiple concurrent comparisons. For example, a first comparison between a first subset of the divided subsets and the second text feature, and a second comparison between a second subset of the divided subsets and the second text feature are performed concurrently.
Let F
Embodiments of the invention may be implemented on virtually any type of computer regardless of the platform being used. For example, as shown in
Further, those skilled in the art will appreciate that one or more elements of the aforementioned computer system (400) may be located at a remote location and connected to the other elements over a network. Further, embodiments of the invention may be implemented on a distributed system having a plurality of nodes, where each portion of the invention (e.g., various modules of
While the invention has been described with respect to a limited number of embodiments, those skilled in the art, having benefit of this disclosure, will appreciate that other embodiments can be devised which do not depart from the scope of the invention as disclosed herein. Accordingly, the scope of the invention should be limited only by the attached claims.
Number | Name | Date | Kind |
---|---|---|---|
6978419 | Kantrowitz | Dec 2005 | B1 |
7734627 | Tong | Jun 2010 | B1 |
7984029 | Alspector et al. | Jul 2011 | B2 |
8266121 | Roitblat et al. | Sep 2012 | B2 |
8620907 | Risvik et al. | Dec 2013 | B2 |
20020114515 | Hotta et al. | Aug 2002 | A1 |
20050060643 | Glass et al. | Mar 2005 | A1 |
20060029296 | King et al. | Feb 2006 | A1 |
20080044016 | Henzinger | Feb 2008 | A1 |
20110153653 | King et al. | Jun 2011 | A1 |
20110295856 | Roitblat et al. | Dec 2011 | A1 |
20110311140 | Urbach et al. | Dec 2011 | A1 |
20120290597 | Henzinger | Nov 2012 | A1 |
20130173563 | Alspector et al. | Jul 2013 | A1 |
20130212090 | Sperling et al. | Aug 2013 | A1 |
20140075566 | Farkash et al. | Mar 2014 | A1 |