The acceptance of fraudulent items results in significant loss to many industries and consumers. In order to determine the veracity of items, many institutions have employed fraud detection processes. Such processes may identify certain items as potentially fraudulent. Upon identification, these items require further processing in order to determine their authenticity. The further processing may be done by a computer program, or in some situations, manually by experts trained in detection fraud. Both situations are often demanding upon resources.
In order to better identify fraudulent items, many industries employ multiple fraud detection processes. However, each fraud detection process generally identifies different items as being potentially fraudulent, thereby increasing the number of potentially fraudulent items that must be further examined. One solution to this problem has been to disregard a significant number of items identified by the multiple fraud detection processes and accept the items as being authentic. However, under such an approach, many actual fraudulent items are accepted as authentic, thus resulting in loss to both industries and consumers. It is with respect to this general environment that embodiments of the present disclosure have been contemplated.
Embodiments of the present disclosure relate to computer implemented systems and methods for producing an ordered suspect list of potentially fraudulent items. In embodiments, one or more fraudulent detection processes are applied to a set of input items to determine the veracity of the input items. In embodiments, the input items may be personal or corporate checks. In other embodiments, the input items may be any type of item that must be authenticated. The one or more fraud detection process may analyze items to produce one or more suspect lists. In embodiments, the one or more suspect lists may be unordered.
Embodiments of the disclosed methods operate upon the one or more suspect lists to produce a single, ordered suspect list. In such embodiments, confidence values may be determined for items on the one or more suspect lists. Confidence values may be determined using historically confirmed fraudulent data and historically flagged suspect data. The confidence values may relate to the likelihood that the suspect item is indeed fraudulent. In other words, the confidence value may relate to the probability of loss for each suspect item if the suspect item is accepted as authentic. In other embodiments, the confidence values may relate to any type of characteristic used in determining veracity.
In embodiments, after a confidence value has been assigned to every item on the one or more suspect lists, a voting process is applied to the one or more lists to determine a new confidence value for each item on the one or more suspect lists. In embodiments, the resulting combined suspect list may be ordered by new confidence values determined using the voting process. In other embodiments, the voting process may rescale the confidence values assigned to each item on the one or more suspect lists. In further embodiments, the expected loss may be calculated for each item on the suspect list. In such embodiments, the combined suspect list may be ordered by the expected loss.
This summary is provided to introduce a selection of concepts in a simplified form that are further described below in the Detailed Description. This summary is not intended to identify key features or essential features of the claimed subject matter, nor is it intended to be used to limit the scope of the claimed subject matter.
Embodiments of the present invention may be more readily described by reference to the accompanying drawings in which like numbers refer to like items and in which:
This disclosure will now more fully describe exemplary embodiments with reference to the accompanying drawings, in which some of the possible embodiments are shown. Other aspects, however, may be embodied in many different forms and the inclusion of specific embodiments in the disclosure should not be construed as limiting such aspects to the embodiments set forth herein. Rather, the embodiments depicted in the drawings are included to provide a disclosure that is thorough and complete and which fully conveys the intended scope to those skilled in the art. When referring to the figures, like structures and elements shown throughout are indicated with like reference numerals.
Embodiments of the present disclosure relate to computer implemented methods for producing an ordered suspect list of potentially fraudulent items. In embodiments, one or more fraud detection processes analyze a set of input items to determine the veracity of the input items. A fraud detection process may be transactional-based, image-based, or may utilize any other process or method known in the art to determine the veracity of an item. In embodiments, the input items may be a personal or corporate check. In other embodiments, the input items may be any type of item that must be authenticated. The one or more fraudulent detection process may produce one or more suspect lists. The suspect lists may contain suspect items (e.g., items that the particular fraud detection process identifies as potentially fraudulent). The specific criteria by which each fraud detection process identifies suspect items will vary depending on the type of fraud detection process (e.g., transaction-based) and the thresholds used therein. In embodiments, the one or more fraud detection processes may produce ordered suspect lists. In such embodiments, the suspect lists may be ordered by confidence values associated with each item on the suspect list. The confidence values may relate to the likelihood that the suspect item is indeed fraudulent. In other embodiments, the confidence values may relate to any type of characteristic used in determining veracity. In other embodiments, the one or more fraud detection processes may not associate a confidence value with each item. In still further embodiments, the one or more fraud detection processes may produce unordered suspect lists.
Embodiments of the disclosed methods operate upon the one or more suspect lists to produce a single, ordered suspect list. In such embodiments, confidence values may be determined for suspect items on the one or more suspect lists. In one embodiment, historically confirmed fraudulent data and historically flagged suspect data may be used to calculate confidence values for the suspect items. In another embodiment, the historically confirmed fraudulent data and historically flagged suspect data may be used to calculate confidence values for sub-ranges of data based upon a value associated with the suspect item and a type of suspect situation. In such embodiments, the confidence value may be assigned and/or used to modify the confidence value of any suspect item that falls within a specific sub-range. In other embodiments, the fraud detection processes may have previously determined confidence values for each item on the suspect list. In such embodiments, the confidence values may be rescaled in order to produce more accurate confidence values.
In embodiments, after a confidence value has been assigned to every suspect item on the one or more suspect lists, a voting process is applied to the one or more lists to determine a new confidence value for each item on the one or more suspect lists. In embodiments, the resulting combined suspect list may be ordered by the new confidence values determined in the voting process. In further embodiments, the expected loss may be calculated for each item on the suspect list. In such embodiments, the combined list may be ordered by the expected loss.
Flow proceeds to step 104 where the items are analyzed using a fraud detection process. In embodiments, the fraud detection process may be an image-based fraud detection process. In such embodiments, the image-based fraud detection process may determine whether the item is potentially fraudulent by analyzing the item for specific suspect situations. For example, if the items relate to check drawn on an account, an image-based fraud detection process may analyze the item for suspect situations such as: signature mismatch, check stock mismatch, disagreement between courtesy and legal amounts, whether the check is a preauthorized draft, whether alterations have been made to the check, or any other suspect situations known to the art. In other embodiments, the fraud detection process may be a transactional-based process. In such embodiments, the transactional-based fraud detection process may analyze the item for suspect situations such as, using checks drawn on an account as an example, unusual check number, unusual velocity of transactions, unusual check amount, or any other suspect situation known to the art. In still other embodiments, other types of fraud detection processes may analyze items based upon other suspect situations and/or features specific to the type of fraud detection process. One of skill in the art will appreciate that embodiments of the systems and methods disclosed herein will operate regardless of the type of fraud detection process used and that any number of different suspect situations may be employed with the embodiments disclosed herein.
After the fraud detection process or processes analyze the received items, images of items, or data from the items, flow proceeds to step 106 where the fraud detection process produces the results of the analysis. In embodiments, the results may be outputted in the form of a suspect list. The suspect list may be a list of items that the fraud detection process determined were fraudulent or potentially fraudulent. Flow then proceeds to step 108, where a confidence value is determined for each item in the suspect list. In some embodiments, the fraud detection process may assign a confidence value to each item on the suspect list. In other embodiments, a confidence value may not be assigned to the items on the suspect list. In such embodiments, confidence values are determined using historical information. An example of step 108 is discussed further with respect to
Flow proceeds to step 204, where historically confirmed fraudulent data is retrieved. In embodiments, the historically confirmed fraudulent data contains a history of confirmed fraudulent items (e.g., past items that were subsequently confirmed to be fraudulent). In embodiments, the historically confirmed fraudulent items are associated with a suspect feature such as suspect situations identified by a detection process or processes, or types of fraud. Additionally, in embodiments, the historically confirmed fraudulent item may also be associated with a value related to the item. For example, if the fraudulent item was a check, the value associated with it may be the value of the check or the difference between the courtesy amount and legal amount. In other embodiments, the value may incorporate additional expenses related to resolving a customer claim, such as personnel expenses or lost account revenue. In further embodiments, the historically confirmed fraudulent data may be associated with a time, such as a time that the item was received, a time indicated by the item, etc. Table 1 provides an example historically confirmed fraudulent data for a transactional-based fraud detection system for checks. For ease of illustration, the historically confirmed fraudulent data contains only 10 entries, however one of skill in the art will readily appreciate that, in embodiments, the history file may be larger or smaller, or may be composed of multiple files.
Table 1 is an example of historically confirmed fraudulent data containing 10 entries for checks ‘A’-‘I’ processed over a particular time period. While Table 1 illustrates the historically confirmed fraudulent data as a table, in other embodiments the confirmed fraudulent data may take any form (e.g., lists, records, etc). Each check in the history file was previously confirmed fraudulent, for example, using manual verification, account-holder confirmation, computer verification, or any other type of verification. One of skill in the art will appreciate that there are a number of manual and/or computer aided methods to confirm a check is indeed fraudulent; however, the actual manner in which the check was confirmed fraudulent has no bearing upon embodiments of the present disclosure. For simplicity's sake, assume that the 10 entries for checks ‘A’-‘I’ comprise the entire historically confirmed fraudulent data.
As previously mentioned, the fraudulent item may be associated with one or more suspect situations. For example, the suspect feature associated with check ‘A’ is ‘Velocity of Transactions’ (i.e., unusual number of check written within a given time period). In embodiments, the fraudulent item may be associated with the suspect feature that was used by the fraud detection system to identify or flag the item as potentially fraudulent. For example, check ‘A’ was flagged as potentially fraudulent due to its unusual ‘Velocity of Transactions’. In embodiments, there can be several suspect reasons used to indicate a suspect item. In further embodiments, complex suspect features (e.g., a combination of suspect reasons) may be used to identify a suspect item. Additionally, each fraudulent item may be associated with a value. In the example provided in Table 1, each fraudulent item is a check; therefore the value associated with the fraudulent item is the dollar value represented by check. For example, the value of check ‘A’ is ‘$52.00.’ In other embodiments additional costs may be taken into account such as, for example, indirect losses accrued in determining whether the suspect item is indeed fraudulent or other costs associated with the item being fraudulent. In other embodiments, if the fraudulent item is an item other than a check, another type of value may be associated with the item. In embodiments, the historically confirmed fraudulent data may be analyzed to determine characteristics of the data (e.g., in the example of Table 1, six of the ten entries were identified by unusual ‘Velocity of Transactions’).
While Table 1 illustrates an example confirmed fraudulent data for a transaction-based fraud detection process, confirmed fraudulent data may include different types of data for different types of fraud detection process. For example, Table 2 illustrates an example confirmed fraudulent data file that may be used with an image-based fraud detection process for checks processed during a particular time period.
Table 2 contains the ‘CHECK’, ‘SUSPECT SITUATION’, and ‘AMOUNT’ data as was included in Table 1; but adds additional ‘CONFIDENCE VALUE’ data. As with Table 1, the same type of information can be derived from Table 2 (e.g., 40% of fraudulent items identified by ‘Alterations’, 40% identified by ‘Signature Mismatch, and 20% identified by ‘Check Amount’). In embodiments, the confidence values may be expressed using different scales depending on the suspect situation (e.g., the confidence value for ‘Signature Mismatch’ may be a value between 1 and 100). Confidence values may also be expressed as functions. Again, in embodiments this type of information may be processed and extracted by the method 200, or may be previously calculated by another method or application and retrieved by the method 200 at step 204. However, Table 2 illustrates an example of how history files may differ depending on the type of fraud detection process used to analyze the items of interest. For example, Table 2 incorporates unique ‘SUSPECT SITUATIONS’ (e.g., ‘Signature Verification’) as well as unique data (e.g., ‘CONFIDENCE VALUES’). In embodiments, there may be multiple suspect situations associated with multiple confidence values. While Table 2 has been described as including confidence values from an image-based fraud detection process, in other embodiments, the image-based fraud detection process may not include confidence values. One of skill in the art will appreciate that the type of confirmed fraudulent data may vary, as illustrated by Table 1 and Table 2, depending on the process used and information retrieved in flagging suspects and confirming fraudulent items. Furthermore, while Table 1 and Table 2 illustrate two separate files containing confirmed fraudulent data for two different fraud detection process, one of skill in the art will readily appreciate that confirmed fraudulent data for two or more processes may be stored in a single file or datastore. In the latter case a single suspect item may be associated with suspect features from multiple detectors.
In further embodiments, the historically confirmed fraudulent data may contain statistics related to the historical data rather than data about the individual items themselves. Table 3 is an example of a historically confirmed fraudulent data file that contains statistical information. As demonstrated, such a file may contain data related to a ‘TIME PERIOD’, ‘SUSPECT SITUATION’, and ‘NUMBER OF FRADULENT ITEMS’. In this example, the historically fraudulent data contains information related to the total number of fraudulent items identified by a specific suspect situation. For example, in ‘TIME PERIOD’ ‘1’, ‘444’ confirmed fraudulent items were identified by unusual ‘Velocity of Transactions’. One of skill in the art will appreciate that in other embodiments, the historically confirmed fraudulent data may not be broken down into time periods. For example, the historically confirmed fraudulent data may contain statistical data related to the total number of confirmed fraudulent items ever identified by a suspect situation. In other embodiments, the historically confirmed fraudulent data may comprise percentages rather than number (e.g., 22% of confirmed fraudulent items were identified by unusual ‘CHECK NUMBER’). In still other embodiments, the historically confirmed fraudulent data may contain data related to a value of the items instead of a suspect situation (e.g., values between 1-100, 100-500, etc.) One of skill in the art will appreciate that, in such embodiments, the historically confirmed fraudulent data may contain any type of statistical information and/or may be preexisting.
Flow proceeds to step 206 where the output is compared to historically flagged suspect data. In embodiments, historically flagged suspect data may include data related to the total number of suspects flagged by a fraud detection process, the number of items of each suspect type flagged (for example, the number of items flagged due to ‘Velocity of Transactions’, ‘Check Amount’, or a combination of both suspect situations in a transactional-based fraud detection system). In other embodiments, the historically flagged suspect data may include other per-image data (e.g., when using an image-based transactional process) or statistical data related to historical output lists, such as the time the item was received or processed. In embodiments, the flagged data includes similar data as that was included in the confirmed fraudulent data, however the data relates to all suspect items flagged by the fraud detection process, whereas the confirmed fraudulent data relates to the items that were later determined fraudulent. In embodiments, the historically flagged suspect data and the historically confirmed fraudulent data are limited to the same historical time frame. In other embodiments, the flagged data may be preexisting.
An example of historically flagged suspect data follows. Flagged data may include historical information about the results of a fraud detection process. As an example, a fraud detection process may have flagged 10,000 suspect items. Of these 10,000 suspect items, 3,500 items were flagged for an unusual ‘Velocity of Transactions’ and another 2,000 items were flagged for an unusual ‘Check Number.’ In some embodiments, flagged data may contain statistical data related to the output of the fraud detection process, e.g., 2,000 ‘Check Number’ items, 35% ‘Velocity of Transactions’, etc. One of skill in the art will appreciate that the flagged suspect data may contain any type of data related to suspect items identified by a fraud detection process.
In other embodiments, the historically flagged suspect data may also contain information related to a time period in which the items were flagged. For example, the time period might relate to the last six months, in which case any data included in the historically flagged suspect data relates to suspect items processed by the fraud detection process within the last six months. In other embodiments, the historically flagged suspect data may contain data related to all the suspect items ever processed by the fraud detection process. In such embodiments, a time of process may be associated with every suspect item. Using the time of process data, applications accessing the flagged data information may select flagged data about suspect items according to when they were processed. For example, an application or method receiving the flagged data may desire to analyze only the data related to suspect items processed within the last month. The application would be able to select only flagged data related to suspect items processed within the last month based upon the time of process data associated with the flagged data. In embodiments, the time period used for the historically flagged suspect data is the same time period for the historically confirmed fraudulent data.
After retrieving the historically confirmed fraudulent data and the historically flagged suspect data in steps 204 and 206, respectively, flow proceeds to step 208 where a confidence value is assigned for the output from the fraud detection process. In embodiments, the output from the fraud detection process may be a suspect list and a confidence value may be calculated for each suspect item in the suspect list. In another embodiment, the suspect list may be divided into ranges and/or categories (e.g., all items with a suspect situation that is ‘Velocity of Transactions’ and/or a value between 100 and 1,000) and a confidence value may be calculated for each range. In such an embodiment, every item located within the range and/or category is assigned the calculated confidence value for the range and/or category. In embodiments, the confidence value may be calculated as an estimate of the probability for the suspect items to be fraudulent or any monotonic function of that probability. In embodiments, the confidence value may be calculated using historically confirmed fraudulent data and/or historically flagged suspect data (e.g., by correlating past results). While embodiments of
Flow then proceeds to step 302, where, in embodiments, a number of sub-ranges are determined. For example, sub-ranges may be divided by suspect situation, e.g., ‘Velocity of Transactions’, ‘Signature Mismatch’, etc., or a combination thereof. In other embodiments, sub-ranges may also be determined by amount, such as a check amount if suspect items relate to check. For example, in such a situation sub-ranges may include ranges from 0-$100, $101-$500, etc., until all check values are placed in a sub-range. In embodiments, sub-ranges may be determined based upon the distribution of values (e.g., by sampling the data to determine a best distribution in which to divide the data). In embodiments where there is confidence value associated with suspect situations, the confidence value could be split into sub-ranges (0-0.05, 0.05-0.4, 0.4-0.6, etc).
In embodiments, suspect items may then be organized and grouped according to sub-range. For example, in an embodiment, sub-ranges may be determined by both category and amount. The suspect items may be grouped into categories and then each category may be divided by into sub-ranges by value. As an example, one grouping may be all suspect items with a suspect situation with an amount within a certain range (e.g., suspect situation is ‘Velocity of Transactions’ and amount is $100-$500). One of skill in the art will appreciate that sub-ranges may be determined in any manner known to the art.
The number of sub-ranges determined in step 304 relates to the accuracy of the confidence value assigned and/or calculated in step 306. For example, a large number of sub-ranges may be advantageous for purposes of statistical relevance. However, this must be balanced by the fact that a smaller number of sub-ranges leads to a more refined probability calculation. Thus, in embodiments, these two competing factors may be considered in determining the number of sub-ranges to divide the data into.
In other embodiments, the sub-ranges may be predetermined. This may be a result of the historical data. For example, referring to Table 3, the historical data is divided by ‘TIME PERIOD’ and ‘SUSPECT SITUATION’. In embodiments where the historical data is previously categorized (e.g., by ‘TIME PERIOD’ and ‘SUSPECT SITUATION’), the sub-ranges determined in step 304 may be limited by such categorization.
Flow proceeds to step 306, where the number of confirmed fraudulent items, from the historically confirmed fraudulent data (
In embodiments, upon calculating the confidence value for each item in the output from the fraud detection process, the items in the output may be ordered.
Flow proceeds to step 406, where the method produces a suspect list including confidence values. In embodiments, the suspect list produced in step 406 includes confidence values for each suspect item in the suspect list. Flow then proceeds to step 408 where the suspect list is ordered by confidence values. In one embodiment, the suspect list may be ordered from highest confidence value to lowest confidence value. In another embodiment, the confidence value may be ordered from lowest confidence value to highest confidence value. In further embodiments, the ordered suspect list is used to determine which suspect items require further analysis. For example, suspect items with a high confidence value may have a higher probability of being actually fraudulent. In this situation, suspect items with a high confidence value should be further analyzed. The ordered suspect list provides an efficient mechanism for determining which suspect items require further analysis. In other embodiments, a threshold may be determined. The threshold may be predetermined or may be determined at step 408 based upon the characteristics of the output and/or suspect values. In embodiments, multiple thresholds may be determined. For example, separate threshold may be determined for each suspect type. In embodiments where a threshold is used, suspect items may only be included on the ordered suspect list if the confidence value meets the threshold requirement.
In embodiments, the suspect list may be passed as input to another application. In this embodiment, the application may use the ordered suspect list to perform further analysis on suspect items included on the suspect list. In further another embodiment, the suspect list may be stored for later use. In yet another embodiment, the suspect list may provide guidance as to which suspect items require manual verification. While the ordering of suspect lists has been described with reference to confidence values, one of skill in the art will recognize that suspect lists may be ordered by any other method known to the art. For example, suspect items may be ordered according to probability of loss, expected amount of loss (described herein with respect to
While the embodiments disclosed thus far have focused on producing an ordered suspect list based upon a single fraud detection process, the results of multiple fraud detection processes may be combined into a single ordered list. Combining multiple suspect lists identified by multiple fraud detection processes into a single, ordered suspect list results in a more accurate listing of potentially fraudulent items.
In embodiments, the second output may be produced by a second fraud detection process operating on the same input as a first fraud detection process. In other embodiments, the second output list may be produced by operating on a subset of the input used by the first detection process. For example, the second fraud detection process may receive as input the suspect items identified by the first fraud detection process. In such embodiments, the first fraud detection process may operate using a relaxed threshold in order to produce a larger set of suspect items.
In yet another embodiment, a confidence value may be associated with each item of the first or second outputs. In other embodiments, confidence values may not be associated with each item of the first and second outputs. In further embodiments, a confidence value may be associated with items of one of the first or second outputs but not the other. One of skill in the art will recognize that embodiments of the present disclosure will operate regardless of whether the output from the fraud detection processes include previously determined confidence values.
Flow proceeds to operation 506 where confidence values are determined for the first suspect list. In embodiments, confidence values may be calculated as described with respect to
At step 508, confidence values are determined for the second suspect list. Again, in embodiments, confidence values may be calculated as described with respect to
Flow proceeds to operation 510 where a voting process is applied to the first and second suspect lists. In embodiments, the voting process may be used to adjust the confidence values associated with each list according to the separate determinations of the first and second input processes. In embodiments, if confidence values were previously determined for the fist and second suspect lists, operations 506 and 508 may be skipped, and confidence values for the suspect items may be determined using the voting process of operation 510.
In an embodiment with two suspect lists, the voting process may produce confidence values based on three separate situations. First, if the suspect item is on the first list but not the second list, voting process may simply leave the confidence value to the suspect item unchanged. For example, the voting process may apply the following formula in this situation to derive a confidence value:
F(C1, 0)=C1
where F( ) is a function which receives as inputs a confidence value of a suspect item from a first suspect list and a second suspect list and outputs a confidence value based upon the inputs and C1 is the confidence value of the suspect item from the first confidence list. In this embodiment, if the suspect item only appears on one suspect list, the confidence value assigned to the suspect item remains unchanged. In other embodiments, voting process may apply a penalty to the confidence value because the suspect item only appeared on one suspect list. In other embodiments, the voting process may increase the confidence value of the suspect item when it only appears on a single list. One of skill in the art will appreciate that there are numerous ways of calculating a confidence value in this situation, all of which may be practice with the embodiments disclosed herein.
Second, if the suspect item is on the second list but not the first list, voting process, in embodiments, may simply leave the confidence value to the suspect item unchanged. For example, the voting process may apply the following formula in this situation to derive a confidence value:
F(0, C2)=C2
where F( ) is a function which receives as inputs a confidence value of a suspect item from a first suspect list and a second suspect list and outputs a confidence value based upon the inputs and C2 is the confidence value of the suspect item from the second confidence list. Again, if the suspect item only appears on one suspect list, the confidence value assigned to the suspect item remains unchanged. As previously mentioned, in other embodiments, voting process may apply a penalty to the confidence value because the suspect item only appeared on one suspect list. In other embodiments, the voting process may increase the confidence value of the suspect item when it only appears on a single list. One of skill in the art will appreciate that there are numerous ways of calculating a confidence value in this situation, all of which may be practice with the embodiments disclosed herein.
Third, in embodiments, if the suspect item is on both the first and second suspect lists, the voting process may determine confidence values by adding confidence value from the first list to the second list and then subtracting the product of the first and second confidence values from the result. For example, the voting process may apply the following formula to calculate confidence values:
F(C1,C2)=(C1+C2)−(C1×C2)
where F( ) is a function which receives as inputs a confidence value of a suspect item from a first suspect list and a second suspect list and outputs a confidence value based upon the inputs, C1 is the confidence value of the suspect item from the first confidence list, and C2 is the confidence value of the suspect item from the second confidence list. While specific embodiments of a voting process have been disclosed herein, one of skill in the art will recognize that any voting process known to the art may be employed with the embodiments of the present disclosure.
In embodiments, the confidence values resulting from the voting process employed at step 510 may be rescaled. In one embodiment, the confidence values may be rescaled using the method described with respect to step 306 (
After the confidence values are determined and/or adjusted at step 510, flow proceeds to step 512 where a single suspect list is produced. In embodiments, the single suspect list is produced by combining the first and second outputs produced by the first and second fraud detection processes. In further embodiments, the single suspect list is ordered by confidence value. In still further embodiments, a confidence value may be determined for the output as a whole, for groups of items associated with a specific suspect type, or separately for each item of the output. As previously described with reference to
In embodiments, the suspect list may be passed as input to another application. In this embodiment, the application may use the ordered suspect list to perform further analysis on suspect items included in the suspect list. In further another embodiment, the suspect list may be stored for later use. In yet another embodiment, the suspect list may provide guidance as to which suspect items require manual verification. While the ordering of suspect lists has been described with reference to confidence values, one of skill in the art will recognize that suspect lists may be ordered by any other method known to the art. For example, suspect items may be ordered according to probability of loss, expected amount of loss (described herein with respect to
Flow proceeds to operation 604 where the value of each suspect item is determined. In embodiments, the value of the suspect item may be determined with reference to historically confirmed fraudulent data or historically flagged suspect data (
Flow proceeds to step 606 where the expected loss is calculated by multiplying the value associated with a suspect item by the confidence value associated with the item. In embodiments, this value may be used to order the suspect list. Ordering suspect list by expected loss provides a beneficial way of determining which items to give priority if further fraudulent analysis is required (e.g., analysis by a separate application or manually). For example, with respect to checks drawn on an account, a first item may have a 95% of being fraudulent with a value of $10. A second check may have only a 5% chance of being fraudulent but a value of $10,000. If these items are ordered by confidence value or probability of loss, the first check clearly has priority over the second. However, calculating the expected loss using the method disclosed in
With reference to
In its most basic configuration, computer system 700 comprises at least one processing unit or processor 704 and system memory 706. The most basic configuration of the computer system 700 is illustrated in
Additionally, computer system 700 may also have additional features/functionality. For example, computer system 700 includes additional storage media 708, such as removable and/or non-removable storage, including, but not limited to, magnetic or optical disks or tape. In some embodiments, software or executable code and any data used for the described system is permanently stored in storage media 708. Storage media 708 includes volatile and non-volatile, removable and non-removable media implemented in any method or technology for storage of information such as computer readable instructions, data structures, program modules, or other data. In embodiments, mammogram images and/or results of probability determination are stored in storage media 708.
System memory 706 and storage media 708 are examples of computer storage media. Computer storage media includes, but is not limited to, RAM, ROM, EEPROM, flash memory or other memory technology, CD-ROM, digital versatile disks (DVD) or other optical storage, magnetic cassettes, magnetic tape, magnetic disk storage, other magnetic storage devices, or any other medium which is used to store the desired information and which is accessed by computer system 700 and processor 704. Any such computer storage media may be part of computer system 700. In some embodiments, mammogram images and/or results of probability determination are stored in system memory 706. In embodiments, system memory 706 and/or storage media 708 stores data used to perform the methods or form the system(s) disclosed herein, such as historically confirmed fraudulent data, historically flagged suspect data, probability distributions, etc. In embodiments, system memory 706 would store information such as suspect lists 714 and application instructions 716. In embodiments, suspect lists 714 may contain suspect lists produced by a fraud detection process. Application instructions 716, in embodiments, stores the procedures necessary to perform the disclosed methods and systems. For example, application data 716 may include functions or processes for calculating probability of loss or confidence values, resealing confidence values, voting procedures, etc.
Computer system 700 may also contain communications connection(s) 710 that allow the device to communicate with other devices. Communication connection(s) 710 is an example of communication media. Communication media may embody a modulated data signal, such as a carrier wave or other transport mechanism and includes any information delivery media, which may embody computer readable instructions, data structures, program modules, or other data in a modulated data signal. The term “modulated data signal” means a signal that has one or more of its characteristics set or changed in such a manner as to encode information or a message in the data signal. By way of example, and not limitation, communication media includes wired media such as a wired network or direct-wired connection, and wireless media such as an acoustic, RF, infrared, and other wireless media. In an embodiment, mammogram images and or determinations of probability results may be transmitted over communications connection(s) 710.
In some embodiments, computer system 700 also includes input and output connections 712, and interfaces and peripheral devices, such as a graphical user interface. Input device(s) are also referred to as user interface selection devices and include, but are not limited to, a keyboard, a mouse, a pen, a voice input device, a touch input device, etc. Output device(s) are also referred to as displays and include, but are not limited to, cathode ray tube displays, plasma screen displays, liquid crystal screen displays, speakers, printers, etc. These devices, either individually or in combination, connected to input and output connections 712 are used to display the information as described herein. All these devices are well known in the art and need not be discussed at length here.
In some embodiments, the component described herein comprise such modules or instructions executable by computer system 700 that may be stored on computer storage medium and other tangible mediums and transmitted in communication media. Computer storage media includes volatile and non-volatile, removable and non-removable media implemented in any method or technology for storage of information such as computer readable instructions, data structures, program modules, or other data. Combinations of any of the above should also be included within the scope of readable media. In some embodiments, computer system 700 is part of a network that stores data in remote storage media for use by the computer system 700.
An illustration of an embodiment of the method and system at work will aid in fully understanding an embodiment of the present disclosure. The following illustration is intended to provide a description of an embodiment of the disclosed systems and methods and is not intended to be used to limit the scope of the claimed subject matter. In embodiments, a financial institution may use multiple fraud detection processes (e.g., a transactional-based process and an image-based process) in order to authenticate the checks that the institution receives. Each of these processes may produce a suspect list identifying checks that may be fraudulent. The multiple fraud detection processes may operate on the same input, or may operate on the output of a first fraud detection process. In embodiments where the fraud detection processes operate on the output of the first fraud detection process, the threshold of the first fraud detection process may be relaxed in order to produce more suspect items.
Once the multiple fraud detection processes have produced multiple suspect lists, embodiments of the present disclosure may operate on the multiple suspect lists. In embodiments, confidence values may be determined for each suspect list that does not include confidence values. Confidence values may be determined using methods described with regard to
Once confidence values have been determined for each suspect list, a voting process is performed on the suspect lists. In embodiments, the voting processes may be the process described with regard to
Once the confidence values for suspect items have been determined and/or refined using the voting process, the suspect items may be combined into a single suspect list. The suspect list may be ordered by confidence value. In other embodiments, the confidence value may be ordered by probability of loss. In further embodiments, the expected loss for each suspect item may be calculated. In embodiments, the expected loss may be calculated by multiplying the value associated with a suspect item by the probability of loss associated with the item. In such embodiments, the single suspect list may be ordered according to expected loss.
This disclosure described some embodiments of the present invention with reference to the accompanying drawings, in which only some of the possible embodiments were shown. Other aspects may, however, be embodied in many different forms and should not be construed as limited to the embodiments set forth herein. Rather, these embodiments were provided so that this disclosure was thorough and complete and fully conveyed the scope of the possible embodiments to those skilled in the art.
Although the embodiments have been described in language specific to structural features, methodological acts, and computer-readable media containing such acts, it is to be understood that the possible embodiments, as defined in the appended claims, are not necessarily limited to the specific structure, acts, or media described. One skilled in the art will recognize other embodiments or improvements that are within the scope and spirit of the present invention. Therefore, the specific structure, acts, or media are disclosed only as illustrative embodiments. The invention is defined by the appended claims.