The present disclosure relates generally to processing database information and, more specifically, to data classification methods and apparatus for use in fusing multiple databases into a single database.
Successful planning, development, deployment and marketing of products and services depend heavily on having access to relevant, high quality market research data. Companies have long recognized that improving the manner in which marketing data is collected, processed, and analyzed often results in more effective delivery of the right products and services to consumers and increased revenues. Recently, companies have sought to more effectively target marketing efforts toward specific groups or individuals having certain combinations of demographic characteristics and psychographic profiles. Such highly targeted marketing efforts may provide a company a significant competitive advantage, particularly for highly competitive markets in which increased revenues are obtained primarily as a result of increased market share.
Market researchers have long dealt with the practical tradeoff between the desire to develop database information that enables companies to develop and deploy highly targeted marketing plans and the desire to develop database information that is more versatile in its application or utility. For example, a database developed from a respondent panel or survey that has been narrowly tailored to provide information related to the television viewing behaviors of a particular regional population having a particular demographic profile may be of little, if any, use when attempting to determine the fast food consumption habits of another population having that same demographic profile.
In response to the practical difficulties (e.g., the cost) associated with assembling market research panels or surveys covering multiple types of consumption activities, behaviors, preferences, etc., market researchers have employed database fusion techniques to efficiently merge or fuse database information from multiple research panels or surveys (typically two at a time) into a single database representing a single virtual population group or respondent-level panel. It is well known that the fusion of two datasets or databases into one dataset or database may enable the development of a database that reveals correlations between the consumption activities, preferences, etc. associated with two datasets or databases in a manner that the individual datasets could not. In other words, existing market research databases can be combined or fused in different ways to generate new datasets or databases that reveal respondent behaviors and/or relationships not previously revealed by the independent databases, without having to physically develop and pay for an expensive multi-purpose respondent panel or survey.
Typically, the fusion of databases or datasets involves a statistical analysis to identify a mathematical function that can be used to predict respondent usage patterns. In general, the mathematical function produced as a result of the statistical analysis is used to guide or facilitate the process of matching observations or records in the datasets or databases to be fused. In some cases, known distance function techniques are used to measure the similarities between observations or records. In other cases, the statistical analysis may process usage data using regression modeling techniques to identify those variables that are common to the databases or datasets to be fused and best suited to match observations or records.
To simplify and/or enhance a data fusion process, it is often desirable to group or segment database observations or records, each of which typically corresponds to a particular person, respondent, or household, according to a plurality of classes, or groups representing different types or levels of consumption behavior (e.g., non-consumers, low consumers, medium consumers, high consumers, etc.) By classifying, grouping, or segmenting the data to be fused, a simplified or separate fusion process can be carried out for each segment. The smaller size of the segments (in comparison to the dataset(s) the segments compose) enables the fusion process to be performed more quickly and efficiently. In addition, the data classification, grouping, or segmentation can produce better results that, for example, enable more accurate prediction of consumption behaviors.
While known fusion techniques typically rely on the use of distance functions or regression models to predict consumption behavior, the resulting predictions are not well-suited to classify or group the records or observations within datasets to be fused into discrete classes or groups. For instance, as noted above, it may be desirable to segment, classify, or group the observations or records within the datasets into classes or groups such as non-consumers, high consumers, medium consumers, and low consumers. However, regression models and distance functions are specifically adapted to predict information (e.g., usage or consumption information) that is inherently continuous in nature (e.g., dollars spent) rather than discrete such as, for example, usage classifications or groups.
In general, the data classification methods, apparatus, and articles of manufacture described herein may be advantageously applied to enhance or improve the operation and/or results of data fusion processes. Specifically, the methods, apparatus, and articles of manufacture described herein analyze the nodes of a data classification tree to selectively form an arrangement or groups of nodes to represent respective classes, or groups associated with one or more databases or datasets. The groups of nodes are then used to classify each of the records within the database(s) or dataset(s) to be fused and the classified records may then be fused according to the classifications.
Before providing a detailed discussion of the above-outlined data classification methods, apparatus, and articles of manufacture, a general discussion of data classification trees is provided. In general, data classification trees or classification tree models are one well-known manner of classifying, grouping, or otherwise segmenting data. Typically, the classifications, groups, segments, etc. represent discrete usage classes or groups and, thus, typically represent a particular type of consumer behavior. For instance, it may be desirable to classify, group, or segment consumer data records or observations into high, medium, low, and non-consumption groups, or classes. Classifying, grouping, or segmenting consumer data records in this manner may help to provide a more intuitive understanding of the profile of a population, which may facilitate the selection of effective advertising types and time slots, product developments and, more generally, can facilitate targeted marketing efforts toward certain members of the population.
Classification trees or models are typically composed of a set of rules by which data (e.g., records or observations, each of which may correspond to a person, a household, etc.) within one or more datasets or databases can be classified, grouped, segmented, etc. into a plurality of discrete classes, groups, or segments. A set of descriptive variables (e.g., demographic information, psychographic information, etc.) may be logically interrelated in a hierarchical (e.g., tree-like) data structure according to the values of the variables. The lowest hierarchical level of the classification tree terminates in a set of nodes (i.e., terminal nodes), which are commonly referred to as leaves. Each terminal node or leaf is typically associated with or corresponds to one group, class, or segment and more than one of the terminal nodes may correspond to a given group, class, or segment. Each leaf or terminal node may then reflect a particular type of behavior (e.g., consumption behavior) that is known to be consistent with the class, group, or segment associated with the node.
In practice, the terminal nodes of a classification tree are not perfectly predictive but, instead, are probabilistic in nature and, thus, the class or group associated with each of the nodes is a predicted class or group. In this manner, each terminal node or leaf may be thought of as a likelihood, probability, or prediction that a data record or observation processed by the classification tree model or hierarchy is actually a member of the class or group with which it is associated according to the classification tree or model.
Typically, the class, group, or segment (e.g., type of consumption behavior) that occurs most frequently or that is most likely to occur for the data records associated with each terminal node determines the predicted class or group for that terminal node. Thus, when a dataset or plurality of data records are grouped or classified according to the classification tree, each terminal node may be associated with a frequency distribution of all or some of the classes or groups predicted by the tree. In other words, for each terminal node or leaf, there is at least some probability that data actually associated with each of the possible classes or groups will occur (i.e., will be classified or misclassified) at that node. However, the probability with which each of the possible classes occurs at a given node will typically vary from class to class as well as node to node. Again, as noted above, the class or group that is most likely to occur at a given terminal node typically determines the class or group assigned to that node and, thus, determines the classification or grouping of any data that is associated with that node as a result of being processed by the classification tree.
As can be appreciated from the foregoing, because each terminal node is assigned to correspond to only one class or group (i.e., each terminal node is used to predict only one class or group) and because the probabilistic nature of the nodes results in data actually associated with multiple classes being associated with each node, misclassification of data records can occur. For example, a classification tree may process a data record actually associated with a non-consumer and associate that record with a terminal node that is assigned to predict high volume consumers (e.g., classify a respondent that is actually a non-consumer as a high consumer). Thus, each terminal node typically has an associated error rate (i.e., a likelihood of misprediction or misclassification) that can be quantified and which is typically acceptable in view of the likelihood or probability that data records associated with that node by the classification tree are classified or grouped correctly.
In one example, the above-outlined data classification methods, apparatus, and articles of manufacture identify classification tree information such as, for example, the variables, the hierarchical relationships between the variables, the terminal nodes or nodes, the classes or groups to which the nodes originally correspond, etc. The classes or groups may correspond to usage or consumption classes or any other behavioral characteristic associated with a population of individuals, households, etc. However, as described in greater detail below, the classes or groups to which the nodes originally correspond may be changed or modified based on an analysis of the ability of the nodes to effectively predict the actual class or group to which data records belong. The classification tree information may be stored in a suitable data structure and/or retrieved from a volatile and/or non-volatile memory.
Analysis of the nodes of the classification tree begins by assigning each of the nodes a plurality of values, where each of the plurality of values is indicative of a relationship between that node and only one of the classes or groups associated with the nodes. The plurality of values associated with each of the nodes may be representative of a frequency distribution of the classes at that node (e.g., the frequency of occurrence, probability of occurrence, etc. for each of the classes at that node). In particular, each of the plurality of values may be an index value or ratio associated with the likelihood that a data record actually associated with a particular one of the classes or groups will be associated with the node corresponding to the value and, thus, predicted or classified (correctly or incorrectly) to have behavior(s) similar or identical to those associated with the class or group assigned to the node. The index values or ratios may be determined based on a comparison (e.g., a division) of the frequency of occurrence of a particular class at a node and the frequency of occurrence of the particular class within the population as a whole.
Continuing with the example, the node values may then be compared and each node may be assigned to one of the classes or groups based on the comparison. In particular, a class or group code may be assigned or otherwise associated with each node, where each class or group code corresponds to only one of the possible classes or groups associated with the nodes of the classification tree. Thus, each of the nodes is assigned to only one of the classes or groups and, in some cases, at least two of the nodes have the same class or group code. As a result, each of the nodes may become associated with a modified predicted class or group relative to the class or group originally associated with that node prior to the analysis of the classification tree nodes.
In general, the nodes may be assigned to the classes or groups using an iterative process (e.g., a process that may make multiple passes through the node value information) in which relatively larger node values preferably determine the class or group to which each of the nodes is assigned. In some cases, a previously assigned node may be re-assigned to another class or group for which that node has a larger or otherwise better value.
After the nodes have been assigned to the classes or groups, the data records or observations from one or more datasets or databases may be classified based on the assignment of the nodes to the classes or groups. More specifically, each data record or observation from the dataset or database may be analyzed using the classification tree to determine a terminal node that corresponds to the data record or observation. In other words, each data record or observation is analyzed using the classification tree and becomes associated with a particular one of the terminal nodes of the classification tree. The classes or groups to which the terminal nodes have been assigned as a result of the foregoing analysis are then associated with (e.g., by adding class or group identifying information to) the corresponding data records or observation to classify those data records or observations.
Following the classification of the data records or observations according to the classification tree and the modified predicted classes as set forth above, the classified data records or observations may be fused using fusion methods and apparatus such as, for example, the methods and apparatus disclosed in co-pending international patent application number PCT/US03/33392, the disclosure of which is hereby incorporated in its entirety.
Now turning to
In the example classification tree 100 of
A known method of using the example classification tree 100 to classify a respondent or other data record begins by retrieving the record from a dataset or database. Then, a value representative of household size is extracted from the record. If the extracted value representative of the household size is equal to one, then the record is associated with the terminal node N6, which corresponds to non-users, and the record (and, thus, the respondent associated with the record) is classified as a non-user.
On the other hand, if the household size is greater than or equal to two, then a value representative of the age of the respondent associated with the record is extracted from the record. If the age of the respondent is 23-29 or 31-47, then a value representative of the geographic region in which the respondent's household is located is extracted from the record. If the geographic region is the northeast, then the record is associated with the terminal node N2, which corresponds to high users, and the record (and, thus, the respondent associated with the record) is classified as a high user (i.e., a high consumption user). On the other hand, if the geographic region is not the northeast, then the record is associated with the terminal node N1, which corresponds to medium users, and the record (and, thus, the respondent associated with the record) is classified as a medium user.
If the age of the respondent is not 23-29 or 31-47, then a value representative of the number of children in the respondent's household is extracted from the record. If there are no children in the respondent's household, then the record is associated with the terminal node N5, which corresponds to low users, and the respondent record (and, thus, the respondent associated with the record) is classified as a low user.
On the other hand, if there are one or more children in the respondent's household, then the value representative of the geographic region in which the respondent's household is located is extracted from the record. If the respondent's household is located in the west, then the respondent is associated with the terminal node N3, which corresponds to low users, and the record (and, thus, the respondent associated with the record) is classified as a low user. If the respondent's household is not located in the west, then the respondent record is associated with the terminal node N4, which corresponds to medium users, and the record (and, thus, the respondent associated with the record) is classified as a medium user.
Thus, the above-described known manner of using a classification tree to classify data records analyzes the values associated with certain variables within each data record to determine which terminal node and, thus, which usage class, should be associated with that data record. Of course, as noted above, each of the terminal nodes (e.g., the nodes N1, N2, N3, N4, N5, and N6) is probabilistic in nature and, thus, predictive of the actual classification of the data record. As a result, the example classification tree 100 and known analysis method described above may properly classify some records, while other records are misclassified. However, typically, the degree to which a classification tree misclassifies can be quantified and such a tree may be refined to reduce the probability of misclassification to an acceptable level.
Now turning in detail to the example method 200 of
The example method 200 then analyzes the terminal nodes of the classification tree (block 204). In general, the node analysis process 204 may generate or assign a plurality of values to each of the nodes (e.g., the nodes N1, N2, N3, N4, N5, and N6 of the example classification tree 100 of
In addition to generating a value corresponding to each of the possible classes, or groups for each node, the node analysis process 204 generates a value representative of the proportion of an overall population that will be associated with each of the nodes. For example, as described in greater detail in connection with
After performance of the node analysis process (block 204), the nodes are grouped or arranged using the node values generated at block 204 according to class or group codes, where each class or group code corresponds to only one class, or group (e.g., a usage or consumption class) (block 206). As described in greater detail in connection with
In contrast to the node grouping or arrangement process (block 206) described herein, known classification tree methods typically classify a node as the class that most frequently occurs at that node, regardless of the characteristics of the overall population, regardless of the relative importance or size of the node in comparison to the overall population and other nodes, and regardless of the overall effectiveness of the classification tree at accurately predicting or classifying data records or observations.
After performance of the node grouping or arrangement process (block 206), the node grouping or arrangement (e.g., the example node grouping of
Finally, the data records, including their assigned class codes, may be fused using the class codes to guide or otherwise facilitate the fusion process (block 210). For example, in the case of a fold-over fusion (i.e., where an original dataset or database is split into two datasets or databases), the assigned class codes may be inserted as matching variables. Additionally or alternatively, the assigned class or group codes may be used to segment one or more datasets or databases to enhance the overall performance of a fusion process on the one or more datasets or databases.
In any event, after the values are generated for the selected node (block 304), the example method 300 determines if there are more nodes to process (block 306). If there are more nodes for which values need to be generated, then control returns to the block 302. On the other hand, if there are no more nodes to process, then the example method 300 ends or returns control to, for example, the example method 200 of
Now turning in detail to
While the example method 600 is described as moving from right to left across the columns of class or group data depicted in
In any event, in this example, the high usage class or group is first selected at block 602 and the example method 600 then selects an available node (initially all nodes are available) that is unassigned to the current group or class (i.e., the high usage class) and which has the largest value (e.g., index value) (block 604). Thus, in this example, the node N6 is selected at block 604 because it is available, unassigned to the high usage class, and has the highest index value (i.e., 160). Then, the example method 600 determines if the selected node (i.e., N6) is assigned to another group (i.e., one of the groups 1, 2, 3, or 4 shown in
The method 600 determines if a group can accept a selected node (block 608) by comparing the total sum-of-weights assigned to the group (including the currently selected node) to the total weight the group or class has within the overall population. For example, continuing with the example, the currently selected node N6 contributes a weight of 10% of the total population, there are no other nodes currently assigned the high usage group (i.e., group 4) so the total sum-of-weights would be 10% if the node N6 were added to the group 4, and the total weight of group 4 (i.e., the high usage class) is 20% of the total population. Accordingly, the example method 600 determines at block 608 that group 4, which corresponds to the high usage class or group, can accept the node N6 and assigns the node N6 to group 4 (block 610).
After assigning the node N6 to group 4, the example method 600 determines if group 4 is full (block 612). A group is full if the total weight of the nodes (i.e., the sum-of-weights) currently assigned to the group is equal to or exceeds the total weight of the class or group within the overall population. In the case of group 4, group 4 is full if the sum of weights assigned to group 4 equals or exceeds 20%. Continuing with the example, after the node N6 is assigned to group 4 at block 610, the example method 600 determines at block 612 that group 4 is not full and then determines if there are any remaining available nodes (in the high usage column of the data of
The example method 600 then continues in the above-described manner to next assign the node N9 (which has an index value of 140 and a weight of 5%) and the node N2 (which has an index value of 120 and a weight of 5%) to group 4. After the node N2 is assigned to group 4, the method 600 determines at block 612 that group 4 is full because the sum-of-weights assigned to group 4 equals 20%, which is the proportion of the overall population that should fall within the high usage class or group. After determining that group 4 is full at block 612, the method 600 determines if there are any currently unassigned nodes (i.e., nodes which have not been assigned to any group or class) (block 616). At this point, the nodes N1, N3, N4, N5, N7, and N8 have not yet been assigned and, thus, the method 600 returns control to block 602 to select the next group.
In this example, after first processing the data in the high usage column of the data of
The method 600 then continues and assigns the node Ng to group 3, which results in group 3 being full at block 612 and selection of group 2 for processing when the method returns control to block 602. The method 600 continues processing for node assignments to group 2 in the above-described manner so that the nodes N1, N5, and N0 are assigned to group 2. The nodes N3 and N7 are not assigned to group 2 because the sum-of-weights resulting from the assignment of either of these nodes to group 2 as determined at block 608 would exceed the permissible total of 20%, which is the total weight of the low usage group or class within the overall population. Thus, the method determines at block 608 that group 2 cannot accept either of the nodes N3 and N7 and makes those nodes unavailable to group 2 at block 620.
In addition, the method 600 may employ a tie-breaking scheme in the event that multiple nodes having the same value associated with the current usage class or group. For example, when assigning nodes to group 2, the method 600 may recognize that either of the nodes N0 or N4 (both of which have an index value of 110) could be assigned to group 2. In response, the method 600 may simply select the lower of the node numbers, may select the node having a lowest value in comparison to the other nodes within a group or class in which none of the tied nodes have been assigned, or may employ any other tie-breaking scheme.
After assigning the nodes N3 and N7 to group 1, the method 600 determines at block 616 that the node N4 has not been assigned and the method returns control to block 602, which again selects group 4 for processing (i.e. the method 600 begins to iterate through the groups again). During the second pass of the method 600 through the groups, the method 600 attempts to again first select the node N2 for assignment to group 4 because it is indicated as available to group 4 and has the largest value among the nodes that are not currently assigned to group 4. However during this pass, the method 600 determines that the node N2 is assigned to group 3 (block 606) and that the value of the node N2 in group 4 (i.e., 120) is not greater than the value of the node N2 in group 3 (i.e., 155). As a result, the method passes control to block 620 and makes the node N2 unavailable to group 4. The method 600 then continues and assigns the node N4 to group 4. After the method determines at block 616 that all the nodes have been assigned, the method 600 ends or returns control to, for example, the example method of
Also, it should be recognized that any or all of the structure shown in the example system 800 of
The methods described herein (e.g., the example methods depicted in
Now turning in detail to
The processor 900 may, for example, be implemented using one or more Intel® microprocessors from the Pentium® family, the Itanium® family or the XScale® family. Of course, other processors from other families are also appropriate.
The processor 900 is in communication with a main memory including a volatile memory 904 and a non-volatile memory 906 via a bus 908. The volatile memory 904 may be implemented by Synchronous Dynamic Random Access Memory (SDRAM), Dynamic Random Access Memory (DRAM), RAMBUS Dynamic Random Access Memory (RDRAM) and/or any other type of random access memory device. The non-volatile memory 906 may be implemented by flash memory and/or any other desired type of memory device. Access to the memory 904 is typically controlled by a memory controller (not shown) in a conventional manner.
The system 902 also includes a conventional interface circuit 910. The interface circuit 910 may be implemented by any type of well-known interface standard, such as an Ethernet interface, a universal serial bus (USB), and/or a third generation input/output (3GIO) interface.
One or more input devices 912 are connected to the interface circuit 910. The input device(s) 912 permit a user to enter data and commands into the processor 900. The input device(s) can be implemented by, for example, a keyboard, a mouse, a touchscreen, a track-pad, a trackball, isopoint and/or a voice recognition system.
One or more output devices 914 are also connected to the interface circuit 910. The output devices 914 can be implemented, for example, by display devices (e.g., a liquid crystal display, a cathode ray tube display (CRT), a printer and/or speakers). The interface circuit 910, thus, typically includes a graphics driver card.
The interface circuit 910 also includes a communication device such as a modem or network interface card to facilitate exchange of data with external computers via a network 916 (e.g., an Ethernet connection, a digital subscriber line (DSL), a telephone line, coaxial cable, a cellular telephone system, etc.).
The system 902 also includes one or more mass storage devices 918 for storing software and data. Examples of such mass storage devices include floppy disk drives, hard drive disks, compact disk drives and digital versatile disk (DVD) drives.
Although certain methods and apparatus and articles of manufacture have been described herein, the scope of coverage of this patent is not limited thereto. To the contrary, this patent covers all methods, apparatus and articles of manufacture fairly falling within the scope of the appended claims either literally or under the doctrine of equivalents.
This patent arises from a continuation of U.S. patent application Ser. No. 12/847,856, filed Jul. 30, 2010, now U.S. Pat. No. 8,234,226, which is a continuation of U.S. patent application Ser. No. 12/389,162, filed Feb. 19, 2009, now U.S. Pat. No. 7,792,771, which is a continuation of U.S. patent application Ser. No. 11/692,480, filed Mar. 28, 2007, now U.S. Pat. No. 7,516,111, which is a continuation of PCT International Application Serial No. PCT/US2004/031965, entitled “Data Classification Methods and Apparatus for Use with Data Fusion,” filed on Sep. 28, 2004, which are hereby incorporated herein by reference in their entireties.
Number | Name | Date | Kind |
---|---|---|---|
4646145 | Percy et al. | Feb 1987 | A |
4987539 | Moore et al. | Jan 1991 | A |
5201010 | Deaton et al. | Apr 1993 | A |
5227874 | Von Kohorn | Jul 1993 | A |
5341142 | Reis et al. | Aug 1994 | A |
5412731 | Desper | May 1995 | A |
5497430 | Sadovnik et al. | Mar 1996 | A |
5504675 | Cragun et al. | Apr 1996 | A |
5550928 | Lu et al. | Aug 1996 | A |
5612527 | Ovadia | Mar 1997 | A |
5630127 | Moore et al. | May 1997 | A |
5687322 | Deaton et al. | Nov 1997 | A |
5771307 | Lu et al. | Jun 1998 | A |
5774868 | Cragun et al. | Jun 1998 | A |
5806032 | Sproat | Sep 1998 | A |
5848396 | Gerace | Dec 1998 | A |
5850470 | Kung et al. | Dec 1998 | A |
5873068 | Beaumont et al. | Feb 1999 | A |
5937392 | Alberts | Aug 1999 | A |
5948061 | Merriman et al. | Sep 1999 | A |
5963653 | McNary et al. | Oct 1999 | A |
5991735 | Gerace | Nov 1999 | A |
6029139 | Cunningham et al. | Feb 2000 | A |
6070147 | Harms et al. | May 2000 | A |
6081269 | Quarendon | Jun 2000 | A |
6091956 | Hollenberg | Jul 2000 | A |
6137911 | Zhilyaev | Oct 2000 | A |
6202053 | Christiansen et al. | Mar 2001 | B1 |
6246796 | Horikoshi et al. | Jun 2001 | B1 |
6298330 | Gardenswartz et al. | Oct 2001 | B1 |
6385329 | Sharma et al. | May 2002 | B1 |
6400996 | Hoffberg et al. | Jun 2002 | B1 |
6430539 | Lazarus et al. | Aug 2002 | B1 |
6434520 | Kanevsky et al. | Aug 2002 | B1 |
6490567 | Gregory | Dec 2002 | B1 |
6574350 | Rhoads et al. | Jun 2003 | B1 |
6611842 | Brown | Aug 2003 | B1 |
6618725 | Fukuda et al. | Sep 2003 | B1 |
6650779 | Vachtesvanos et al. | Nov 2003 | B2 |
6662195 | Langseth et al. | Dec 2003 | B1 |
6662215 | Moskowitz et al. | Dec 2003 | B1 |
6671715 | Langseth et al. | Dec 2003 | B1 |
6683966 | Tian et al. | Jan 2004 | B1 |
6714917 | Eldering et al. | Mar 2004 | B1 |
6728390 | Rhoads et al. | Apr 2004 | B2 |
6742003 | Heckerman et al. | May 2004 | B2 |
6744906 | Rhoads et al. | Jun 2004 | B2 |
6748426 | Shaffer et al. | Jun 2004 | B1 |
6785671 | Bailey et al. | Aug 2004 | B1 |
6804376 | Rhoads et al. | Oct 2004 | B2 |
6836773 | Tamayo et al. | Dec 2004 | B2 |
6839682 | Blume et al. | Jan 2005 | B1 |
6850626 | Rhoads et al. | Feb 2005 | B2 |
6944319 | Huang et al. | Sep 2005 | B1 |
7032072 | Quinn et al. | Apr 2006 | B1 |
7035855 | Kilger et al. | Apr 2006 | B1 |
7072841 | Pednault | Jul 2006 | B1 |
7089194 | Berstis et al. | Aug 2006 | B1 |
7154146 | Wang et al. | Dec 2006 | B2 |
7243075 | Shaffer et al. | Jul 2007 | B1 |
7295970 | Gorin et al. | Nov 2007 | B1 |
7324447 | Morford | Jan 2008 | B1 |
7376080 | Riddle et al. | May 2008 | B1 |
7420992 | Fang et al. | Sep 2008 | B1 |
7428526 | Miller et al. | Sep 2008 | B2 |
7496661 | Morford et al. | Feb 2009 | B1 |
7516111 | Samson et al. | Apr 2009 | B2 |
7577576 | Baechtiger | Aug 2009 | B2 |
7607147 | Lu et al. | Oct 2009 | B1 |
7644422 | Lu et al. | Jan 2010 | B2 |
7698345 | Samson et al. | Apr 2010 | B2 |
7792771 | Samson et al. | Sep 2010 | B2 |
7912734 | Kil | Mar 2011 | B2 |
8046797 | Bentolila et al. | Oct 2011 | B2 |
8234226 | Samson et al. | Jul 2012 | B2 |
8355910 | McMillan et al. | Jan 2013 | B2 |
8364678 | Miller et al. | Jan 2013 | B2 |
20010010046 | Muyres et al. | Jul 2001 | A1 |
20010028662 | Hunt et al. | Oct 2001 | A1 |
20010042016 | Muyres et al. | Nov 2001 | A1 |
20010056405 | Muyres et al. | Dec 2001 | A1 |
20020002488 | Muyres et al. | Jan 2002 | A1 |
20020004744 | Muyres et al. | Jan 2002 | A1 |
20020059578 | Nishiura | May 2002 | A1 |
20020116124 | Garin et al. | Aug 2002 | A1 |
20020138492 | Kil | Sep 2002 | A1 |
20020183076 | Pande et al. | Dec 2002 | A1 |
20020184077 | Miller et al. | Dec 2002 | A1 |
20030009467 | Perrizo | Jan 2003 | A1 |
20030061132 | Yu, Sr. et al. | Mar 2003 | A1 |
20030105660 | Walsh et al. | Jun 2003 | A1 |
20030182249 | Buczak | Sep 2003 | A1 |
20030204384 | Owechko et al. | Oct 2003 | A1 |
20030208488 | Perrizo | Nov 2003 | A1 |
20050039206 | Opdycke | Feb 2005 | A1 |
20050065955 | Babikov et al. | Mar 2005 | A1 |
20050096971 | Baechtiger | May 2005 | A1 |
20050240462 | Inman et al. | Oct 2005 | A1 |
20050240468 | Inman et al. | Oct 2005 | A1 |
20060287969 | Li | Dec 2006 | A1 |
20080089591 | Zhou et al. | Apr 2008 | A1 |
20080104032 | Sarkar | May 2008 | A1 |
20080319834 | Miller et al. | Dec 2008 | A1 |
Number | Date | Country |
---|---|---|
2110866 | Jun 1994 | CA |
0642060 | Mar 1995 | EP |
0822535 | Apr 1998 | EP |
9927668 | Jun 1999 | WO |
0111506 | Feb 2001 | WO |
2005050482 | Jun 2005 | WO |
Entry |
---|
Bernstein et al., “An Introduction to Map Matching for Personal Navigation Assistants,” Princeton University, Aug. 1996, 17 pages. |
Fang et al., “Design of a Wireless Assisted Pedestrian Dead Reckoning System—The NavMote Experience,” IEEE Transactions on Instrumentation and Measurement, vol. 54, Issue 6, Dec. 2005, 17 pages. |
Chibelushi et al., “Audio-Visual Person Recognition: an Evaluation of Data Fusion Strategies,” University of Wales Swansea, United Kingdom, European Conference on Security and Detection, Apr. 28-30, 1997, 6 pages. |
Zhao et al., “Localization Using Combing Sensors and Dead-Reckoning,” New Science Publishers, Inc., Exerpt Web Page from ACM Portal-Computing Literature Search Engine, 2004, 3 pages. |
Written Opinion of the International Searching Authority in connection with International Patent Application No. PCT/US04/31965, mailed Jul. 18, 2005, 4 pages. |
International Search Report in connection with International Patent Application No. PCT/US04/31965, mailed Jul. 18, 2005, 4 pages. |
Written Opinion of the International Preliminary Examining Authority in connection with International Patent Application No. PCT/US04/31965, mailed Feb. 2, 2007, 4 pages. |
International Preliminary Report on Patentability in connection with International Patent Application No. PCT/US04/31965, report completed Sep. 25, 2007, 3 pages. . |
Office Communication for U.S. Appl. No. 11/692,480, mailed May 12, 2008, 12 pages. |
Office Communication for U.S. Appl. No. 11/692,480, mailed Aug. 28, 2008, 13 pages. |
Notice of Allowance for U.S. Appl. No. 11/692,480, mailed Nov. 20, 2008, 9 pages. |
Office Communication for U.S. Appl. No. 12/389,162, mailed Jan. 4, 2010, 7 pages. |
Notice of Allowance for U.S. Appl. No. 12/389,162, mailed May 3, 2010, 9 pages. |
Integrating support vector machines in a hierarchical output space decomposition framework, Yangchi Chen; Crawford, M.M.; Ghosh, J.; Geoscience and Remote Sensing Symposium, 2004. IGARSS '04. Proceedings. 2004 IEEE International vol. 2 Digital Object Identifier: 10.11 09/IGARSS.2004.1368565 Publication Year: 2004 , pp. 949-952 vol. 2. |
Exploring Multiple Trees through DAG Representations, Graham, M.; Kennedy, J.; Visualization and Computer Graphics, IEEE Transactions on vol. 13, Issue: 6 Digital Object Identifier: 10.1109/TVCG.2007.70556 Publication Year: 2007 , pp. 1294-1301. |
An iterative growing and pruning algorithm for classification tree design, Gelfand, S.B.; Ravishankar, C.S.; Delp, E.J.; Pattern Analysis and Machine Intelligence, IEEE Transactions on vol. 13 , Issue: 2 Digital Object Identifier: 10.1109/34.67645 Publication Year: 1991 , pp. 163-174. |
Comparison of a decision tree and maximum likelihood classifiers: application to SAR image of tropical forest, Simard, M.; Saatchi, S.; DeGrandi, G.; Geoscience and Remote Sensing Symposium, 2000. Proceedings. IGARSS 2000. IEEE 2000 International vol. 5 Digital Object Identifier: 10.11 09/IGARSS.2000.858324 : 2000 , pp. 2129-2139. |
Non-final Office action issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/847,856 on Dec. 19, 2011 (12 pages). |
Notice of Allowance issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/847,856 on Mar. 28, 2012 (5 pages). |
Business Wire, “Claritas Announces Second in its Suite of Internet-Deliverable Products; MyBestCustomers Comes Less Than a Month After the Launch of MyBestProspects,” Aug. 1, 2000, 3 pages. |
Business Wire, Sagent Partners with Claritas to Deliver Real-time Marketing Information for More Effective Customer Analysis, Mar. 29, 2000, 5 pages. |
Zhang, Heping, “Classification Trees for Multiple Binary Responses,” Journal of American Statistical Association, Mar. 1998, 15 pages. |
Zhang et al., “An Implementation of ID3 that Creates Executable Rules,” 1990, 1 page. |
Zhang et al., “A Prolog System for Case-Based Classification,” 1990, 1 page. |
Perner, Lars, “The Psychology of Consumers, Consumer Behavior and Marketing,” Dec. 12, 2006, 33 pages. |
Ma et al., “On Reconstructing Species Trees from Gene Trees in Term of Duplications and Losses,” 1998, 11 pages. |
Claritas, “Understanding Psychographics,” Dec. 6, 2006, 4 pages. |
United States Patent and Trademark Office, “Notice of Allowance,” issued in connection with U.S. Appl. No. 09/872,457, mailed May 13, 2008, 4 pages. |
United States Patent and Trademark Office, “Advisory Action,” issued in connection with U.S. Appl. No. 09/872,457, mailed Aug. 3, 2007, 2 pages. |
United States Patent and Trademark Office, “Advisory Action,” issued in connection with U.S. Appl. No. 09/872,457, mailed Jul. 10, 2007, 3 pages. |
United States Patent and Trademark Office, “Final Office Action,” issued in connection with U.S. Appl. No. 09/872,457, mailed Apr. 17, 2007, 6 pages. |
United States Patent and Trademark Office, “Non-Final Office Action,” issued in connection with U.S. Appl. No. 09/872,457, mailed Jan. 3, 2007, 20 pages. |
United States Patent and Trademark Office, “Final Office Action,” issued in connection with U.S. Appl. No. 09/872,457, mailed Sep. 5, 2006, 11 pages. |
United States Patent and Trademark Office, “Non-Final Office Action,” issued in connection with U.S. Appl. No. 09/872,457, mailed Mar. 17, 2006, 10 pages. |
United States Patent and Trademark Office, “Advisory Action,” issued in connection with U.S. Appl. No. 09/872,457, mailed Feb. 2, 2006, 3 pages. |
United States Patent and Trademark Office, “Final Office Action,” issued in connection with U.S. Appl. No. 09/872,457, mailed Nov. 17, 2005, 11 pages. |
United States Patent and Trademark Office, “Non-Final Office Action,” issued in connection with U.S. Appl. No. 09/872,457, mailed Jun. 3, 2005, 14 pages. |
United States Patent and Trademark Office, “Notice of Allowance,” issued in connection with U.S. Appl. No. 12/202,805, mailed Oct. 15, 2012, 13 pages. |
United States Patent and Trademark Office, “Non-Final Office Action,” issued in connection with U.S. Appl. No. 12/202,805, mailed May 24, 2012, 12 pages. |
United States Patent and Trademark Office, “Final Office Action,” issued in connection with U.S. Appl. No. 12/2002,805, mailed Nov. 4, 2011, 16 pages. |
United States Patent and Trademark Office, “Non-Final Office Action,” issued in connection with U.S. Appl. No. 12/202,805, mailed Apr. 21, 2011, 23 pages. |
United States Patent and Trademark Office, “Final Office Action,” issued in connection with U.S. Appl. No. 10/829,405, mailed Dec. 18, 2012, 16 pages. |
United States Patent and Trademark Office, “Non-Final Office Action,” issued in connection with U.S. Appl. No. 10/829,405, mailed Mar. 27, 2012, 13 pages. |
United States Patent and Trademark Office, “Final Office Action,” issued in connection with U.S. Appl. No. 10/829,405, mailed Aug. 9, 2011, 15 pages. |
United States Patent and Trademark Office, “Non-Final Office Action,” issued in connection with U.S. Appl. No. 10/829,405, mailed Dec. 20, 2012, 19 pages. |
United States Patent and Trademark Office, “Final Office Action,” issued in connection with U.S. Appl. No. 10/829,405, mailed May 17, 2010, 16 pages. |
United States Patent and Trademark Office, “Non-Final Office Action,” issued in connection with U.S. Appl. No. 10/829,405, mailed Aug. 3, 2009, 15 pages. |
United States Patent and Trademark Office, “Non-Final Office Action,” issued in connection with U.S. Appl. No. 10/829,405, mailed Jan. 5, 2009, 12 pages. |
United States Patent and Trademark Office, “Non-Final Office Action,” issued in connection with U.S. Appl. No. 10/829,405, mailed Apr. 4, 2008, 8 pages. |
United States Patent and Trademark Office, “Final Office Action,” issued in connection with U.S. Appl. No. 11/119,235, mailed Jul. 17, 2009, 11 pages. |
United States Patent and Trademark Office, “Non-Final Office Action,” issued in connection with U.S. Appl. No. 11/119,235, mailed Dec. 19, 2008, 9 pages. |
Number | Date | Country | |
---|---|---|---|
20120278328 A1 | Nov 2012 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 12847856 | Jul 2010 | US |
Child | 13538933 | US | |
Parent | 12389162 | Feb 2009 | US |
Child | 12847856 | US | |
Parent | 11692480 | Mar 2007 | US |
Child | 12389162 | US | |
Parent | PCT/US2004/031965 | Sep 2004 | US |
Child | 11692480 | US |