The present invention relates generally to detecting, monitoring and addressing data compromises, and more specifically to detecting data compromises, as applied to credit card fraud and other types of fraud, in a faster and more efficient manner.
Currently, fraud detection is a complex and manually intensive process that involves culling through billions of transactions to identify activities. Fraud is difficult to detect. Fraud detection is not an exact science and oftentimes transactions from good customers are declined in the process, thereby negatively impacting customer relations. The timeliness of fraud detection is also a major concern. If fraudulent activities are not detected early enough, fraudsters can make a major impact and cause substantial losses to customers, merchants, financial institutions and other entities.
Network intrusions and database attacks are larger, more frequent and more sophisticated. Data compromises are costing an exemplary financial institution and other players from $3 million to $25 million, and usually more, in gross fraud per month.
Oftentimes, fraud is only detected after the compromise has cost card members, merchants, card associations, and/or financial institutions millions of dollars.
Other drawbacks may also be present.
Accordingly, one aspect of the invention is to address one or more of the drawbacks set forth above. According to an embodiment of the present invention, a method for detecting one or more data compromises, wherein the method is executed by a programmed computer processor which communicates with a user via a network, comprises the steps of detecting an activity indicative of a data compromise based at least in part on a compromise detection report involving at least one of merchant compromise report, region compromise report and acquirer compromise report, wherein the compromise detection report is generated by an automated detection algorithm; classifying the activity based on a combination of risk level, size of data compromise and fraud severity; and identifying a specific mitigation response to the activity through an interface.
According to an exemplary embodiment of the present invention, an automated computer implemented method and system for detecting one or more data compromises further comprises determining whether one or more accounts associated with the activity have been probed by a fraudster to determine if the one or more accounts are valid; wherein the merchant compromise report identities one or more common merchants where authorizations were swiped with one or more counterfeit cards; wherein an account fraud rate is calculated for each common merchant, wherein the account fraud rate represents a number of counterfeit fraud accounts as compared to a total of accounts transacted at a merchant; wherein the merchant compromise report identifies merchant identifier, sales card count, fraud card count, fraud loss and account fraud rate; wherein the region compromise report identifies one or more non-U.S. countries having an increase in account fraud rate for a predetermined time period; wherein one or more card transactions that have turned fraud during a predetermined time period occurred in the identified one or more countries; wherein the region compromise report identifies region risk group, one or more non-U.S. countries, fraud indicator and account fraud rate; wherein the acquirer compromise report identities one or more common acquirers with authorizations involving one or more fraud card numbers; wherein the acquirer compromise report identifies acquirer identifier, sales time period and fraud data.
According to an exemplary embodiment of the present invention, a computer implemented system for detecting one or more data compromises comprises a detection module for detecting an activity indicative of a data compromise based at least in part on a compromise detection report involving at least one of merchant compromise report, region compromise report and acquirer compromise report, wherein the compromise detection report is generated by an automated detection algorithm; a classification module for classifying the activity based on a combination of risk level, size of data compromise and fraud severity; and a mitigation module for identifying a specific mitigation response to the activity through an interface.
In order to facilitate a fuller understanding of the present inventions, reference is now made to the appended drawings. These drawings should not be construed as limiting the present inventions, but are intended to be exemplary only.
An embodiment of the present invention improves the efficiency of identifying potential fraudulent activities. An embodiment of the present invention is directed to detecting data compromises in a faster and more efficient manner. Another embodiment of the present invention also relates to detecting probing transactions, which generally relate to testing transactions used by fraudsters to determine whether stolen data is valid and accounts targeted for fraud can be used by the fraudster.
Any common point or location either physical or virtual from where cardholder information is stolen or exposed may be considered a “compromise.” For example, a data compromise generally affects card holder information, such as card number, encoded data on the magnetic stripe, security code on the card (e.g., three-digit code, four-digit code, etc.), expiration date, etc. The compromise may involve one or more various participants, such as merchant, merchant acquirer, card association, processor, third party entities, etc. The compromise may involve a common or related place or instance where data is taken or stolen by an entity.
The data compromise management process of an embodiment of the present invention may involve several steps including: (1) detection; (2) classification; (3) drilldown; (4) mitigation and (5) termination. Compromises may be detected from external sources (e.g., card associations, such as VISA, Mastercard, etc.), fraud policy (e.g., proactive compromise detection reports), case analysis team (e.g., manual account reviews) and/or other sources. Classification may involve identifying whether transactions are card-present or card-not-present, severity level, fraud loss type, etc. Drilldown may identify risk indicators, customer indicators, account indicators and/or other data. Drilldown may also involve checking for probes. Mitigation may involve declining or queuing strategy analysis, applying card watch status controls, reissuing with expiration date change only, and/or other appropriate action. Termination may involve monitoring losses, SAR filings, and/or other actions. The data compromise management process of an embodiment of the present invention may be applied to various scenarios and applications. For example, an exemplary process may detect merchant compromise, region or country compromise and acquirer compromise. An aspect of the invention may involve test cards that are inserted into the criminal market on purpose to provide information on how they move across a fraud system.
Another embodiment of the invention is directed to probe detection which may be a tool to find types of tests that are used by fraudsters to validate a compromised account. A probe may involve a testing of an account to verify that the account is valid. Probes generally involve low dollar transactions that are not posted to avoid alerting the consumer. Probes may be classified as pure probes and parasitic probes. Pure probes may involve the situation where fraudsters create a mechanism to process authorizations. Parasitic probes generally involve fraudsters abusing merchants to test probes. For example, a newspaper website may be a target for probes where fraudsters can test transactions on accounts. The probe detection aspect of an embodiment of the present invention may identify probes by recognizing low dollar transactions that are not posted, high fraud rates on an account level that occur after the probe, merchant identifiers and/or other factors and considerations.
Compromises are costing an exemplary financial institution from $3 million to $25 million in gross fraud per month. An embodiment of the present invention identifies compromises on accounts faster and more effectively thereby mitigating substantial losses caused by fraud. A goal of an embodiment of the present invention is to narrow the window from the time of data compromise to detection and mitigation of risk relative to the compromise.
While the detailed description is directed to an exemplary application involving credit card compromises, the various embodiments of the invention may be applied to other scenarios and applications involving other fraudulent activities or other activities involving detecting data compromises. Other applications may be applied in varying scope.
A system 100 of an embodiment of the present invention may include a Processor 110, which may be stand alone, hosted by an entity, such as a financial institution, service provider, bank, etc. For example, Processor 110 may be affiliated or associated with a financial institution, bank and/or other entity with fraud concerns. In an exemplary embodiment involving a financial institution such as 130, the financial institution may host or support the Processor. In this example, the application of the data compromise algorithm of an embodiment of the present invention may appear to be performed by financial institution, as a single consolidated unit, as shown by 132.
According to another example, Processor 110 may be separate and distinct from Financial Institution 130. For example, Financial Institution 130, or other entity, may communicate to Processor 110 via a network or other communication mechanism, as shown by 122.
Processor 110 may access databases and/or other sources of information to identify data compromises and/or other information for effectively identifying compromises, fraudulent and potentially fraudulent events. For example, Processor 110 may access and/or maintain Database 140 and/or other database 142. Database 140 may include data, such as account information, transaction activity, payment activity, and/or other relevant data for one or more accounts. While a single database is illustrated in the exemplary figure, the system may include multiple databases at the same location or separated through multiple locations. The databases may be further combined and/or separated. In addition, the databases may be supported by Financial Institution 130 or an independent service provider. For example, an independent service provider may support the one or more databases and/or other functionality at a remote location. Other architectures may be realized. The components of the exemplary system diagrams may be duplicated, combined, separated and/or otherwise modified, as desired by various applications of the embodiments of the present invention as well as different environments and platforms.
Data Compromise Management Processor 110 may communicate with various entities, including card associations 160, e.g., VISA, Mastercard, etc., Card Member(s) 162, Merchant(s) 164, Acquiring Bank(s) 166, Third Party Processor(s) 168, External Source(s) 170 and/or Other Sources 172. In addition, the entities may communicate with each other. An embodiment of the present invention may also communicate to the Authorities 150, including police, law enforcement, FBI, terrorism bureaus, government entities and/or other entities. In addition, suspicious activity report (SAR) filings may be facilitated through an embodiment of the present invention, as shown by 152. Communication may be provided by Communication Network 122, 124, 126 and/or other communication mechanism. In addition, Data Compromise Management Processor 110 may have access to other sources of data and/or data feeds that identify other metrics and/or information that may be relevant for identifying fraud activities in accordance with the various embodiments of the present invention. While a single illustrative block, module or component is shown, these illustrative blocks, modules or components may be multiplied for various applications or different application environments. In addition, the modules or components may be further combined into a consolidated unit. Other architectures may be realized. The modules and/or components may be further duplicated, combined and/or separated across multiple systems at local and/or remote locations.
According to another embodiment of the present invention, Data Compromise Management Processor 110 may host a website or other electronic interface, as shown by Interface 210, where users can access data as well as provide data. For example, a financial institution, merchant and/or other entity may access information through an interface to view data, submit requests, provide data and/or perform other actions.
Detection Module 220 may perform analysis in accordance with the various embodiments of the present invention to detect, monitor and/or investigate data compromises, including fraud activities. The detection aspect of an embodiment of the present invention may find one or more points in the market indicative of the data compromise. The information gathered may be data driven, information driven and/or a combination. Data driven may represent data that is currently available, such as an entity's own internal data. For example, a merchant may identify potentially fraudulent card numbers based on suspicious behavior and/or other triggers. Information driven may include data from cyber intelligence, chatter and/or other external sources. Chatter may include communication between or among fraudsters in the criminal market who are buying, selling and/or transacting stolen data. An embodiment of the present invention may identify and access data indicative of fraudulent activity.
Another aspect of an embodiment of the present invention may involve inserting known fraudulent cards into the criminal market and monitoring their usage to gather additional information to better detect fraudulent activity.
Classification Module 230 may classify the detected activity into one or more categories or metrics. For example, the detected activity may be classified based on number of cards compromised, severity level, fraud loss type on impacted accounts, card-present, card-not-present and/or other factors. Card-present involves the situation where a card holder physically swipes the card product at the merchant location. Card-not-present involves the situation where information, such as a card number, expiration date, three-digit code and/or other identifier, is provided, such as an Internet, phone or other key-entered transaction. Fraud usage analysis may also be considered. For example, merchant type, merchant geography, transaction type, valid card verification value (CVV), valid card verification code (CVC) and/or other data may be identified.
According to an exemplary embodiment, the detected activity may be classified based on size, risk, severity level and/or other metric. Size may represent the number of accounts suspected to be in a compromise. For example, a small size merchant compromise may involve less than a predetermined number of accounts, such as 150 accounts. A medium size merchant compromise may involve between 150 to 500 accounts. A large size merchant compromise may involve over 500 accounts. Other ranges and/or thresholds may be applied to designate small, medium and large size merchants. Additional categories of merchants may be applied. Risk may be based on percentage of accounts converting to fraud after the compromise. For example, a high risk may be defined as greater than 25% account fraud rate. A medium risk may be defined as being between 15%-25% account fraud rate. A low risk may be defined as less than 15% account fraud rate. Other percentages may be applied to represent varying degrees of risk. Severity level may be based on a combination of size and risk. Other metrics for classifying the activity may be applied as well.
Drilldown Module 240 may involve analyzing additional variables, triggers and/or other factors. This may be particularly useful for medium risk and/or medium sized compromises but may be applied to other categories as well. According to another embodiment of the present invention, the drilldown process may monitor the usage of contact data for indicating a risk level on a compromise account. For example, an embodiment of the present invention may identify that for accounts that have been compromised, an increased number of calls have occurred. This activity supports the scenario where the fraudster is making calls on the account to obtain additional information about the account, card holder, etc. According to another example, an embodiment of the present invention may identify the time period (e.g., number of days, number of hours, etc.) between a contact and a first fraud transaction. For example, the contact may include a call to VRU (voice recognition unit) and/or other action. An embodiment of the present invention may observe whether the fraudster talked to an advisor, the type of contact, originating number, whether the zip code matched, whether the area code matched, and/or other indicators. An embodiment of the present invention recognizes that a counterfeit type of fraud that manufactured a fraudulent card product will have a different time period between the first fraud transaction and a contact. This data provides indicators useful for assessing risk levels and identifying appropriate mitigation strategies.
An embodiment of the present invention relates to a system and method for detecting probes. Probes may be used by fraudsters to test a compromised card product or other compromised data. In other words, a probe may determine whether compromised data is valid and usable by the fraudster. Card products and/or other data that has been probed are generally verified and worth more to fraudsters. For example, fraudsters can advertise that the account has been tested and can therefore sell the account information at a higher price. An embodiment of the present invention recognizes that a probe check may be a precursor to fraud usage.
Probes may be categorized as parasitic probes or pure probes. Parasitic probes may involve an abused merchants. For example, a fraudster may build an interface, such as a graphical user interface (GUI), on top of a website to perform a probe test. Such websites may be static, with little change. Examples include online newspaper websites, music download websites and any website that can support a high volume of small dollar transactions.
Pure probes may involve a fraudster created mechanism to process authorizations while posing as a merchant. For example, an authorization may be transmitted using the mechanism built into the acquirer to send an authorization request to approval a low dollar amount transaction. If the transaction is approved, the account is in good standing. Fraudsters will not want the probe transaction to be posted to the card holder's account because doing so will alert the customer of the fraud performed on the card. Probe transactions are usually of low dollar amount so as not to alert the card holder. In addition, the fraudster may cancel the transaction so as to avoid posting to the card holder's account. In this pure probe example, the merchant information may rapidly change. For example, the probe abuser may submit different merchant names, different merchant numbers, different locations, etc. The probe abuser may also use an existing legitimate merchant name or identifier. Other categorizes or types of probes may be recognized.
Probes may involve a transaction to check the accuracy of the compromised data. Some characteristics indicative of a probe may involve one or more of the following: low dollar amount, non-posted transaction, high fraud rates on an account level that occurs after the particular probe, multiple merchant names processed against a single merchant identifier, and/or other factors. Also, a rapidly changing merchant or method may be indicative of a probe where multiple names, multiple merchant identifiers, changing SIC codes, changing merchant category codes, city and state may be observed.
An embodiment of the present invention identifies merchants that fit the probe characteristics. For example, an online newspaper website may allow users to manage their monthly newspaper subscription. Probsters may test transactions on this type of website. An embodiment of the present invention may recognize that the website has high account fraud rates, no or very few posted transactions and most of the dollar amounts are very small. For example, account fraud rate may be 50%, which means that one in every two accounts that has an online newspaper website probe will go fraud within the next 30 days. By recognizing that the accounts have been probed, the risk level may be adjusted to be higher on those accounts.
Mitigation Module 250 may identify and apply an appropriate mitigation action to the detected data compromise. For example, the mitigation action may include one or more of the following: decline transaction, queue strategy analysis, card watch status, reissue with expiration date change only, proactive close and reissue, account compromise management process (ACMP) queue, block and reissue, follow-up with internal and external law enforcement, etc. Other mitigation action(s) may be applied. Also, Mitigation Module 250 may assess the compromise and apply an appropriate action. For a compromise involving a small size and high risk, a close and reissue action may be applied. For a compromise involving a medium size with high risk, close and reissue action may be applied. For a compromise involving a large size and low risk, a monitoring action may be applied.
Termination Module 260 may identify and/or apply a recommended action based on the analysis performed on the data compromise. An embodiment of the present invention provides for a more thorough investigation that will allow for better linking of accounts to ensure high risk accounts are properly identified and appropriate actions are taken (e.g., SARs are filed correctly). Moreover, earlier detection may lead to lower losses and damages.
At step 310, data compromises and/or other activity indicative of a fraud or potential fraud may be monitored for data compromise detection. External sources may provide information used for detecting a data compromise. External sources may include card associations, such as VISA, Mastercard, etc., industry sources, fraud investigations, global security and investigations alerts. Manual review of accounts may also be used to assist in identifying data compromises. An embodiment of the present invention may provide proactive compromise detection reports. Exemplary reports may include merchant compromise reports, region compromise reports and acquirer compromise reports. Other compromise detection reports may also be provided.
At step 320, the detected activity may be classified based on various factors. For example, the detected activity may be classified based on number of cards compromised, severity level, fraud loss type on impacted accounts, card-present, card-not-present and/or other factors. Fraud usage analysis may also be considered. For example, merchant type, merchant geography, transaction type, valid card verification value (CVV), valid card verification code and/or other data may be applied. Also, card associations publish compromise account management system (CAMS) alerts and details.
At step 330, a drilldown process may be applied to the detected activity for detailed analysis. For example, credit line, bank identification number (BIN), supersegment (e.g., product or partner affiliated entities, etc.), probe activity, strategic business unit (SBU) (e.g., mass affluence, high net worth, business cards, retail, etc.), card member geography, age of account, internal fraud investigation group (e.g., FOOTPRINTS) and/or other data may be applied. The drilldown process may also consider high credit line accounts (e.g., credit line accounts over $100,000 credit line), private bank accounts, etc. Also, an embodiment of the present invention may also consider voice recognition unit (VRU) calls with information mismatch (e.g., zip code, area code, calling number, etc.), high fraud rate Automatic Number Identification (ANI) calls, etc. As part of the drilldown process, probe checks may also be performed.
At step 340, an appropriate mitigation action may be applied. For example, based on the classification and/or drilldown process, an appropriate mitigation action may be applied to the detected data compromise. According to another example, the mitigation action may include one or more of the following: decline transaction, queue strategy analysis, card watch status, reissue with expiration date change only, proactive close and reissue, account compromise management process (ACMP) queue, block and reissue, etc. Other mitigation action(s) may be applied.
An embodiment of the present invention may apply a specific response to a detected probe activity. For example, an embodiment of the present invention may apply an appropriate action, such as contact the customer about the account, decline the probe, contact a card association and inform them about the probe activity, monitor the probe activity and/or other action. According to another example, an embodiment of the present invention may contact the customer to inquire whether a valid transaction has been made. If not, an embodiment of the present invention may then reissue the card to the customer. An embodiment of the present invention may target the probes and decline these probe transactions. By doing so, the fraudsters will become frustrated and move on to a different target. For example, the particular probe may be identified as having mismatched information where address information has not been verified. In response, the probe transaction may be declined. An embodiment of the present invention may also concurrently contact an appropriate card association and identify a possible breach. If the merchant and card association are affected, most, if not all, transactions may be declined. In addition, an embodiment of the present invention may monitor the probing activity and gain more information about the probing activity.
At step 350, an appropriate termination action may be applied. Termination actions may include monitor loss(es), post implementation review, action/strategy modification, SAR filing, charge back from card association, VISA, Mastercard, etc. Other termination action(s) may be applied. For example, manual review may involve analysis by a trained individual. Automated actions may include usage monitoring, law enforcement involvement, changes to account settings, tracking, automated SAR filing, authority reporting for illegal and/or other activity. An embodiment of the present invention facilitates case display for law enforcement to follow and act on.
A merchant compromise generally refers to a compromise occurring at a merchant. For example, an employee or other individual may access data stored at a merchant location or otherwise associated with a merchant. In this example, an employee at a merchant store may take information, such as card number, account number, etc., stored at the merchant location. The employee may then produce counterfeit cards or other fraudulent data and/or products. An embodiment of the present invention may recognize fraud activity on these cards. By running a merchant compromise report based on a prior time period (e.g., prior 90 days), an embodiment of the present invention may run site transactions on the cards and identify common merchants. One or more filters may be applied to the common merchants. For example, merchants with card transactions of less than 10 may be ignored. The report may include a list of merchants for a certain time period where those identified counterfeit cards were used. Fraud rates may be used to identify high risk merchants.
According to this exemplary embodiment, merchant compromises may include accounts tagged as counterfeit fraud. At step 410, a list of counterfeit card members may be identified. An embodiment of the present invention may identify a list of counterfeit card numbers with swiped or card-present fraud transactions setup during a time period, such as the last 45 days. Other time periods may be applied.
At step 412, swiped or card-present authorizations for a prior predetermined number of days may be identified. From the list of counterfeit card members, an embodiment of the present invention may identify swiped or card-present authorizations for a time period, such as prior 90 days, prior to the fraud transaction date. Other time periods may be applied.
At step 414, common merchants may be identified. An embodiment of the present invention may then identify common merchants where the swiped or card-present authorizations occurred. One or more filters may be applied to this data. According to an exemplary application, merchants with a lower number of fraud accounts may be ignored. For example, merchants with less than 10 fraud accounts may be ignored. Other thresholds may be applied.
At step 416, accounts with swiped or card-present authorizations at the merchants may be identified. For example, an embodiment of the present invention may identify accounts with swiped or card-present authorizations at the merchants in the 90 days prior to fraud setup to calculate an account fraud rate. Other time periods may be applied.
At step 418, an account fraud rate may be calculated for the identified merchants. For example, an account fraud rate may represent a number of counterfeit fraud accounts as compared to total accounts transacted at the merchant. Other representations or indications of fraud may be implemented and considered. An embodiment of the present invention may be directed to analyzing fraud rate after a fraud transaction to assess merchant risk levels. For example, for a group of customers who made transactions at a specific merchant, an embodiment of the present invention may calculate fraud rate over a period of time after the merchant visits. If the fraud rate is considered to be high, it may indicate a common point of compromise. It may also indicate that the merchant is being probed or tested by fraudsters.
At step 420, a list of merchants and “at risk” card numbers with an account fraud rate above a predetermined floor may be identified. An embodiment of the present invention may provide a listing of compromised merchant identifiers and “at risk” card numbers with an account fraud rate meeting a predetermined criteria. For example, the list of merchants may include card numbers with an account fraud rate above and/or equal to a predetermined minimum percentage, such as 2%. Other criteria may be applied. The listing of merchants may be validated for historical data and merchant compromises published by associations and/or data from other sources.
At step 610, a list of card numbers with transactions in one or more regions (e.g., non-US countries) for each month during a predetermined time period may be identified. An embodiment of the present invention may identify a list of card numbers that have transacted in various non-U.S. countries each month for the past one year. Other time periods may be applied. Some accounts may appear in different months and different countries for the same month (or other time period). Other regions, including territory, a country, multiple countries, and/or other area, may be identified.
At step 612, card numbers that turned fraud or confirmed fraud in the next predetermined number of days may be tagged. For example, card numbers that turned fraud or confirmed fraud in the next 180 days may be tagged. Other time periods may be applied.
At step 614, account level fraud rate by sales month and region (e.g., territory, country, multiple countries, etc.) may be identified. For example, an account fraud rate may represent a number of counterfeit fraud accounts as compared to total accounts for a time period. This information may be further compared to normalized account rates over a period of time.
At step 616, regions or countries may be classified into risk levels, such as high, medium and low, based on the number of fraud accounts per month. In addition, an embodiment of the present invention may identify an increase of fraud rate for a time period. Other time periods may be applied.
At step 618, regions or countries with an increase in account fraud rate on a month to month basis may be identified. An embodiment of the present invention recognizes that countries with an increase in account fraud rate month over month are in a compromise and exposure time period of compromise starts from the month first showing an increase in the rate.
At step 620, a list of regions or countries with account fraud rate by month for the predetermined time period may be identified. An embodiment of the present invention may provide a display with a list of regions or countries with account fraud rate by month for the past one year. Other time periods may be applied.
At step 810, a list of card numbers setup as fraud during a time period may be identified. An embodiment of the present invention may identify a list of card numbers setup as fraud in the last 90 days. Other time periods may be applied.
At step 812, swiped or card-present transaction history for prior number of days may be identified. An embodiment of the present invention may identify swiped or card-present transaction history for prior 180 days.
At step 814, common acquirers BINs (Bank Identifier Number) where swiped or card-present transactions were processed may be identified. One or more filters may be applied. For example, the BINs with less than 50 accounts processed may be ignored. Other filters or criteria may be applied.
At step 816, authorization history and fraud data for a time period, such as five months, on acquirer BINs may be identified. An embodiment of the present invention may identify authorization history and fraud data for five months on all acquirer BINs. A time period, such as 60 days prior to current, may be applied to allow for fraud tagging or fraud confirmation. The time period provides for confirmation of fraud, whether the customer is called to confirm fraudulent activity or the customer reports fraudulent activity.
At step 818, acquirer BIN may be classified by risk level, such as high, medium and low risk, by number of total fraud account during the time period, such as five months. Other categories of risk may be applied.
At step 820, a report displaying account fraud rate by month and acquirer BIN may be provided.
While the exemplary embodiments illustrated herein may show the various embodiments of the invention (or portions thereof) collocated, it is to be appreciated that the various components of the various embodiments may be located at distant portions of a distributed network, such as a local area network, a wide area network, a telecommunications network, an intranet and/or the Internet, or within a dedicated object handling system. Thus, it should be appreciated that the components of the various embodiments may be combined into one or more devices or collocated on a particular node of a distributed network, such as a telecommunications network, for example. As will be appreciated from the following description, and for reasons of computational efficiency, the components of the various embodiments may be arranged at any location within a distributed network without affecting the operation of the respective system.
Data and information maintained by Processor 110 may be stored and cataloged in Database 140 which may comprise or interface with a searchable database. Database 140 may comprise, include or interface to a relational database. Other databases, such as a query format database, a Standard Query Language (SQL) format database, a storage area network (SAN), or another similar data storage device, query format, platform or resource may be used. Database 140 may comprise a single database or a collection of databases, dedicated or otherwise. In one embodiment, Database 140 may store or cooperate with other databases to store the various data and information described herein. In some embodiments, Database 140 may comprise a file management system, program or application for storing and maintaining data and information used or generated by the various features and functions of the systems and methods described herein. In some embodiments, Database 140 may store, maintain and permit access to customer information, transaction information, account information, and general information used to process transactions as described herein. In some embodiments, Database 140 is connected directly to Processor 110, which, in some embodiments, it is accessible through a network, such as communication network, e.g., 122, 124, 126 illustrated in
Communications network, e.g., 122, 124, 126, may be comprised of, or may interface to any one or more of, the Internet, an intranet, a Personal Area Network (PAN), a Local Area Network (LAN), a Wide Area Network (WAN), a Metropolitan Area Network (MAN), a storage area network (SAN), a frame relay connection, an Advanced Intelligent Network (AIN) connection, a synchronous optical network (SONET) connection, a digital T1, T3, E1 or E3 line, a Digital Data Service (DDS) connection, a Digital Subscriber Line (DSL) connection, an Ethernet connection, an Integrated Services Digital Network (ISDN) line, a dial-up port such as a V.90, a V.34 or a V.34bis analog modem connection, a cable modem, an Asynchronous Transfer Mode (ATM) connection, a Fiber Distributed Data Interface (FDDI) connection, or a Copper Distributed Data Interface (CDDI) connection.
Communications network, e.g., 122, 124, 126, may also comprise, include or interface to any one or more of a Wireless Application Protocol (WAP) link, a General Packet Radio Service (GPRS) link, a Global System for Mobile Communication (GSM) link, a Code Division Multiple Access (CDMA) link or a Time Division Multiple Access (TDMA) link such as a cellular phone channel, a Global Positioning System (GPS) link, a cellular digital packet data (CDPD) link, a Research in Motion, Limited (RIM) duplex paging type device, a Bluetooth radio link, or an IEEE 802.11-based radio frequency link. Communications network 107 may further comprise, include or interface to any one or more of an RS-232 serial connection, an IEEE-1394 (Firewire) connection, a Fibre Channel connection, an infrared (IrDA) port, a Small Computer Systems Interface (SCSI) connection, a Universal Serial Bus (USB) connection or another wired or wireless, digital or analog interface or connection.
In some embodiments, communication network, e.g., 122, 124, 126, may comprise a satellite communications network, such as a direct broadcast communication system (DBS) having the requisite number of dishes, satellites and transmitter/receiver boxes, for example. Communications network, e.g., 122, 124, 126, may also comprise a telephone communications network, such as the Public Switched Telephone Network (PSTN). In another embodiment, communication network 120 may comprise a Personal Branch Exchange (PBX), which may further connect to the PSTN.
In some embodiments, Processor 110 may include any terminal (e.g., a typical home or personal computer system, telephone, personal digital assistant (PDA) or other like device) whereby a user may interact with a network, such as communications network, e.g., 122, 124, 126, for example, that is responsible for transmitting and delivering data and information used by the various systems and methods described herein. Processor 110 may include, for instance, a personal or laptop computer, a telephone, or PDA. Processor 110 may include a microprocessor, a microcontroller or other general or special purpose device operating under programmed control. Processor 110 may further include an electronic memory such as a random access memory (RAM) or electronically programmable read only memory (EPROM), a storage such as a hard drive, a CDROM or a rewritable CDROM or another magnetic, optical or other media, and other associated components connected over an electronic bus, as will be appreciated by persons skilled in the art. Processor 110 may be equipped with an integral or connectable cathode ray tube (CRT), a liquid crystal display (LCD), electroluminescent display, a light emitting diode (LED) or another display screen, panel or device for viewing and manipulating files, data and other resources, for instance using a graphical user interface (GUI) or a command line interface (CLI). Processor 110 may also include a network-enabled appliance, a browser-equipped or other network-enabled cellular telephone, or another TCP/IP client or other device.
As described above,
As noted above, the processing machine executes the instructions that are stored in the memory or memories to process data. This processing of data may be in response to commands by a user or users of the processing machine, in response to previous processing, in response to a request by another processing machine and/or any other input, for example. As described herein, a module performing functionality may comprise a processor and vice-versa.
As noted above, the processing machine used to implement the invention may be a general purpose computer. However, the processing machine described above may also utilize any of a wide variety of other technologies including a special purpose computer, a computer system including a microcomputer, mini-computer or mainframe for example, a programmed microprocessor, a micro-controller, a peripheral integrated circuit element, a CSIC (Customer Specific Integrated Circuit) or ASIC (Application Specific Integrated Circuit) or other integrated circuit, a logic circuit, a digital signal processor, a programmable logic device such as a FPGA, PLD, PLA or PAL, or any other device or arrangement of devices that is capable of implementing the steps of the process of the invention.
It is appreciated that in order to practice the method of the invention as described above, it is not necessary that the processors and/or the memories of the processing machine be physically located in the same geographical place. That is, each of the processors and the memories used in the invention may be located in geographically distinct locations and connected so as to communicate in any suitable manner. Additionally, it is appreciated that each of the processor and/or the memory may be composed of different physical pieces of equipment. Accordingly, it is not necessary that the processor be one single piece of equipment in one location and that the memory be another single piece of equipment in another location. That is, it is contemplated that the processor may be two pieces of equipment in two different physical locations. The two distinct pieces of equipment may be connected in any suitable manner. Additionally, the memory may include two or more portions of memory in two or more physical locations.
To explain further, processing as described above is performed by various components and various memories. However, it is appreciated that the processing performed by two distinct components as described above may, in accordance with a further embodiment of the invention, be performed by a single component. Further, the processing performed by one distinct component as described above may be performed by two distinct components. In a similar manner, the memory storage performed by two distinct memory portions as described above may, in accordance with a further embodiment of the invention, be performed by a single memory portion. Further, the memory storage performed by one distinct memory portion as described above may be performed by two memory portions.
Further, various technologies may be used to provide communication between the various processors and/or memories, as well as to allow the processors and/or the memories of the invention to communicate with any other entity; e.g., so as to obtain further instructions or to access and use remote memory stores, for example. Such technologies used to provide such communication might include a network, the Internet, Intranet, Extranet, LAN, an Ethernet, or any client server system that provides communication, for example. Such communications technologies may use any suitable protocol such as TCP/IP, UDP, or OSI, for example.
As described above, a set of instructions is used in the processing of the invention. The set of instructions may be in the form of a program or software. The software may be in the form of system software or application software, for example. The software might also be in the form of a collection of separate programs, a program module within a larger program, or a portion of a program module, for example The software used might also include modular programming in the form of object oriented programming. The software tells the processing machine what to do with the data being processed.
Further, it is appreciated that the instructions or set of instructions used in the implementation and operation of the invention may be in a suitable form such that the processing machine may read the instructions. For example, the instructions that form a program may be in the form of a suitable programming language, which is converted to machine language or object code to allow the processor or processors to read the instructions. That is, written lines of programming code or source code, in a particular programming language, are converted to machine language using a compiler, assembler or interpreter. The machine language is binary coded machine instructions that are specific to a particular type of processing machine, i.e., to a particular type of computer, for example. The computer understands the machine language.
Any suitable programming language may be used in accordance with the various embodiments of the invention. Illustratively, the programming language used may include assembly language, Ada, APL, Basic, C, C++, COBOL, dBase, Forth, Fortran, Java, Modula-2, Pascal, Prolog, REXX, Visual Basic, and/or JavaScript, for example. Further, it is not necessary that a single type of instructions or single programming language be utilized in conjunction with the operation of the system and method of the invention. Rather, any number of different programming languages may be utilized as is necessary or desirable.
Also, the instructions and/or data used in the practice of the invention may utilize any compression or encryption technique or algorithm, as may be desired. An encryption module might be used to encrypt data. Further, files or other data may be decrypted using a suitable decryption module, for example.
As described above, the invention may illustratively be embodied in the form of a processing machine, including a computer or computer system, for example, that includes at least one memory. It is to be appreciated that the set of instructions, i.e., the software for example, that enables the computer operating system to perform the operations described above may be contained on any of a wide variety of media or medium, as desired. Further, the data that is processed by the set of instructions might also be contained on any of a wide variety of media or medium. That is, the particular medium, i.e., the memory in the processing machine, utilized to hold the set of instructions and/or the data used in the invention may take on any of a variety of physical forms or transmissions, for example. Illustratively, the medium may be in the form of paper, paper transparencies, a compact disk, a DVD, an integrated circuit, a hard disk, a floppy disk, an optical disk, a magnetic tape, a RAM, a ROM, a PROM, a EPROM, a wire, a cable, a fiber, communications channel, a satellite transmissions or other remote transmission, as well as any other medium or source of data that may be read by the processors of the invention.
Further, the memory or memories used in the processing machine that implements the invention may be in any of a wide variety of forms to allow the memory to hold instructions, data, or other information, as is desired. Thus, the memory might be in the form of a database to hold data. The database might use any desired arrangement of files such as a flat file arrangement or a relational database arrangement, for example.
In the system and method of the invention, a variety of “user interfaces” may be utilized to allow a user to interface with the processing machine or machines that are used to implement the invention. As used herein, a user interface includes any hardware, software, or combination of hardware and software used by the processing machine that allows a user to interact with the processing machine. A user interface may be in the form of a dialogue screen for example. A user interface may also include any of a mouse, touch screen, keyboard, voice reader, voice recognizer, dialogue screen, menu box, list, checkbox, toggle switch, a pushbutton or any other device that allows a user to receive information regarding the operation of the processing machine as it processes a set of instructions and/or provide the processing machine with information. Accordingly, the user interface is any device that provides communication between a user and a processing machine. The information provided by the user to the processing machine through the user interface may be in the form of a command, a selection of data, or some other input, for example.
As discussed above, a user interface is utilized by the processing machine that performs a set of instructions such that the processing machine processes data for a user. The user interface is typically used by the processing machine for interacting with a user either to convey information or receive information from the user. However, it should be appreciated that in accordance with some embodiments of the system and method of the invention, it is not necessary that a human user actually interact with a user interface used by the processing machine of the invention. Rather, it is contemplated that the user interface of the invention might interact, i.e., convey and receive information, with another processing machine, rather than a human user. Accordingly, the other processing machine might be characterized as a user. Further, it is contemplated that a user interface utilized in the system and method of the invention may interact partially with another processing machine or processing machines, while also interacting partially with a human user.
It will be readily understood by those persons skilled in the art that the present invention is susceptible to broad utility and application. Many embodiments and adaptations of the present invention other than those herein described, as well as many variations, modifications and equivalent arrangements, will be apparent from or reasonably suggested by the present invention and foregoing description thereof, without departing from the substance or scope of the invention.
Accordingly, while the present invention has been described here in detail in relation to its exemplary embodiments, it is to be understood that this disclosure is only illustrative and exemplary of the present invention and is made to provide an enabling disclosure of the invention. Accordingly, the foregoing disclosure is not intended to be construed or to limit the present invention or otherwise to exclude any other such embodiments, adaptations, variations, modifications and equivalent arrangements.
The embodiments of the present inventions are not to be limited in scope by the specific embodiments described herein. For example, although many of the embodiments disclosed herein have been described with reference to identifying fraudulent activities, the principles herein are equally applicable to other applications. Indeed, various modifications of the embodiments of the present inventions, in addition to those described herein, will be apparent to those of ordinary skill in the art from the foregoing description and accompanying drawings. Thus, such modifications are intended to fall within the scope of the following appended claims.
Further, although the embodiments of the present inventions have been described herein in the context of a particular implementation in a particular environment for a particular purpose, those of ordinary skill in the art will recognize that its usefulness is not limited thereto and that the embodiments of the present inventions can be beneficially implemented in any number of environments for any number of purposes. Accordingly, the claims set forth below should be construed in view of the full breadth and spirit of the embodiments of the present inventions as disclosed herein.
Number | Name | Date | Kind |
---|---|---|---|
4321672 | Braun et al. | Mar 1982 | A |
4355372 | Goldberg | Oct 1982 | A |
4491725 | Pritchard | Jan 1985 | A |
4495018 | Vohrer | Jan 1985 | A |
4633397 | Macco | Dec 1986 | A |
4694397 | Vignola | Sep 1987 | A |
4722054 | Fukushima | Jan 1988 | A |
4745468 | Von Kohorn | May 1988 | A |
4752676 | Leonard et al. | Jun 1988 | A |
4752877 | Roberts et al. | Jun 1988 | A |
4774664 | Gottardy | Sep 1988 | A |
4797911 | Marks | Jan 1989 | A |
4812628 | Boston | Mar 1989 | A |
4914587 | Clouse | Apr 1990 | A |
4926255 | Von Kohorn | May 1990 | A |
4932046 | Ross | Jun 1990 | A |
4948174 | Thomson et al. | Aug 1990 | A |
4972504 | Daniel, Jr. | Nov 1990 | A |
4974878 | Josephson | Dec 1990 | A |
5025372 | Burton et al. | Jun 1991 | A |
5041972 | Frost | Aug 1991 | A |
5050207 | Hitchcock | Sep 1991 | A |
5121945 | Thomson et al. | Jun 1992 | A |
5122950 | Mee | Jun 1992 | A |
5157717 | Hitchcock | Oct 1992 | A |
5175682 | Higashiyama | Dec 1992 | A |
5179584 | Tsumura | Jan 1993 | A |
5220501 | Lawlor | Jun 1993 | A |
5225978 | Petersen et al. | Jul 1993 | A |
5237620 | Deaton | Aug 1993 | A |
5239642 | Gutierrez et al. | Aug 1993 | A |
5259023 | Katz | Nov 1993 | A |
5260778 | Kauffman | Nov 1993 | A |
5262941 | Saladin | Nov 1993 | A |
5287269 | Dorrough et al. | Feb 1994 | A |
5311594 | Penzias | May 1994 | A |
5326959 | Perazza | Jul 1994 | A |
5361201 | Jost et al. | Nov 1994 | A |
5383113 | Kight | Jan 1995 | A |
5402474 | Miller | Mar 1995 | A |
5424938 | Wagner | Jun 1995 | A |
5430644 | Deaton et al. | Jul 1995 | A |
5448471 | Deaton et al. | Sep 1995 | A |
5465206 | Hilt et al. | Nov 1995 | A |
5483444 | Malark | Jan 1996 | A |
5483445 | Pickering | Jan 1996 | A |
5490060 | Malec | Feb 1996 | A |
5513102 | Auriemma | Apr 1996 | A |
5523942 | Tyler | Jun 1996 | A |
5532920 | Hartrick | Jul 1996 | A |
5537314 | Kanter | Jul 1996 | A |
5550734 | Tarter | Aug 1996 | A |
5555299 | Maloney et al. | Sep 1996 | A |
5559855 | Dowens et al. | Sep 1996 | A |
5561707 | Katz | Oct 1996 | A |
5570465 | Tsakanikas | Oct 1996 | A |
5583759 | Geer | Dec 1996 | A |
5594791 | Szlam et al. | Jan 1997 | A |
5599528 | Igaki | Feb 1997 | A |
5615341 | Srikant | Mar 1997 | A |
5621812 | Deaton et al. | Apr 1997 | A |
5631828 | Hagan | May 1997 | A |
5638457 | Deaton et al. | Jun 1997 | A |
5642485 | Deaton et al. | Jun 1997 | A |
5644723 | Deaton et al. | Jul 1997 | A |
5649114 | Deaton et al. | Jul 1997 | A |
5652786 | Rogers | Jul 1997 | A |
5659165 | Jennings | Aug 1997 | A |
5659469 | Deaton et al. | Aug 1997 | A |
5684863 | Katz | Nov 1997 | A |
5687322 | Deaton et al. | Nov 1997 | A |
5689100 | Carrithers et al. | Nov 1997 | A |
5699527 | Davidson | Dec 1997 | A |
5699528 | Hogan | Dec 1997 | A |
5704044 | Tarter et al. | Dec 1997 | A |
5710889 | Clark et al. | Jan 1998 | A |
5715298 | Rogers | Feb 1998 | A |
5715450 | Ambrose | Feb 1998 | A |
5727249 | Pollin | Mar 1998 | A |
5734838 | Robinson | Mar 1998 | A |
5742775 | King | Apr 1998 | A |
5745706 | Wolfberg et al. | Apr 1998 | A |
5757904 | Anderson | May 1998 | A |
5758126 | Daniels et al. | May 1998 | A |
5761647 | Boushy | Jun 1998 | A |
5765142 | Allred et al. | Jun 1998 | A |
5787403 | Randle | Jul 1998 | A |
5793846 | Katz | Aug 1998 | A |
5794221 | Egendorf | Aug 1998 | A |
5802498 | Comesanas | Sep 1998 | A |
5802499 | Sampson et al. | Sep 1998 | A |
5815551 | Katz | Sep 1998 | A |
5819238 | Fernholz | Oct 1998 | A |
5826241 | Stein | Oct 1998 | A |
5832447 | Rieker | Nov 1998 | A |
5832457 | O'Brien | Nov 1998 | A |
5832460 | Bednar | Nov 1998 | A |
5835087 | Herz | Nov 1998 | A |
5835580 | Fraser | Nov 1998 | A |
5835603 | Coutts | Nov 1998 | A |
5842211 | Horadan | Nov 1998 | A |
5852811 | Atkins | Dec 1998 | A |
5862223 | Walker | Jan 1999 | A |
5870456 | Rogers | Feb 1999 | A |
5870721 | Norris | Feb 1999 | A |
5870724 | Lawlor | Feb 1999 | A |
5873072 | Kight | Feb 1999 | A |
5875437 | Atkins | Feb 1999 | A |
5884032 | Bateman | Mar 1999 | A |
5884288 | Chang | Mar 1999 | A |
5890140 | Clark et al. | Mar 1999 | A |
5897625 | Gustin | Apr 1999 | A |
5899982 | Randle | May 1999 | A |
5903881 | Schrader | May 1999 | A |
5920847 | Kolling et al. | Jul 1999 | A |
5923745 | Hurd | Jul 1999 | A |
5940811 | Norris | Aug 1999 | A |
5943656 | Crooks | Aug 1999 | A |
5949044 | Walker et al. | Sep 1999 | A |
5953406 | LaRue et al. | Sep 1999 | A |
5966695 | Melchione et al. | Oct 1999 | A |
5966698 | Pollin | Oct 1999 | A |
5970467 | Alavi | Oct 1999 | A |
5970480 | Kalina | Oct 1999 | A |
5974396 | Anderson | Oct 1999 | A |
5978780 | Watson | Nov 1999 | A |
5987434 | Libman | Nov 1999 | A |
5987435 | Weiss et al. | Nov 1999 | A |
5991736 | Ferguson et al. | Nov 1999 | A |
5991750 | Watson | Nov 1999 | A |
5995942 | Smith et al. | Nov 1999 | A |
5995948 | Whitford | Nov 1999 | A |
6006205 | Loeb et al. | Dec 1999 | A |
6006207 | Mumick et al. | Dec 1999 | A |
6009415 | Shurling et al. | Dec 1999 | A |
6016344 | Katz | Jan 2000 | A |
6016482 | Molinari et al. | Jan 2000 | A |
6018718 | Walker et al. | Jan 2000 | A |
6018722 | Ray et al. | Jan 2000 | A |
6026370 | Jermyn | Feb 2000 | A |
6029139 | Cunningham et al. | Feb 2000 | A |
6029153 | Bauchner et al. | Feb 2000 | A |
6032125 | Ando | Feb 2000 | A |
6032136 | Brake, Jr. et al. | Feb 2000 | A |
6038552 | Fleischl et al. | Mar 2000 | A |
6049782 | Gottesman et al. | Apr 2000 | A |
6055510 | Henrick | Apr 2000 | A |
6058378 | Clark et al. | May 2000 | A |
6067533 | McCauley et al. | May 2000 | A |
6070147 | Harms et al. | May 2000 | A |
6076072 | Libman | Jun 2000 | A |
6078892 | Anderson et al. | Jun 2000 | A |
6088685 | Kiron et al. | Jul 2000 | A |
6098052 | Kosiba et al. | Aug 2000 | A |
6100891 | Thorne | Aug 2000 | A |
6101486 | Roberts et al. | Aug 2000 | A |
6105007 | Norris | Aug 2000 | A |
6112190 | Fletcher et al. | Aug 2000 | A |
6128599 | Walker et al. | Oct 2000 | A |
6134563 | Clancey et al. | Oct 2000 | A |
6141666 | Tobin | Oct 2000 | A |
6148293 | King | Nov 2000 | A |
6151584 | Papierniak et al. | Nov 2000 | A |
6157924 | Austin | Dec 2000 | A |
6182059 | Angotti et al. | Jan 2001 | B1 |
6189787 | Dorf | Feb 2001 | B1 |
6195644 | Bowie | Feb 2001 | B1 |
6212178 | Beck et al. | Apr 2001 | B1 |
6222914 | McMullin | Apr 2001 | B1 |
6226623 | Schein et al. | May 2001 | B1 |
6230287 | Pinard et al. | May 2001 | B1 |
6233332 | Anderson et al. | May 2001 | B1 |
6233566 | Levine et al. | May 2001 | B1 |
6243688 | Kalina | Jun 2001 | B1 |
6267292 | Walker et al. | Jul 2001 | B1 |
6278981 | Dembo et al. | Aug 2001 | B1 |
6278996 | Richardson et al. | Aug 2001 | B1 |
6292786 | Deaton et al. | Sep 2001 | B1 |
6292789 | Schutzer | Sep 2001 | B1 |
6301567 | Leong et al. | Oct 2001 | B1 |
6304653 | O'Neil et al. | Oct 2001 | B1 |
6304858 | Mosler et al. | Oct 2001 | B1 |
6321212 | Lange | Nov 2001 | B1 |
6324524 | Lent et al. | Nov 2001 | B1 |
6327573 | Walker et al. | Dec 2001 | B1 |
6330543 | Kepecs | Dec 2001 | B1 |
6338047 | Wallman | Jan 2002 | B1 |
6349290 | Horowitz et al. | Feb 2002 | B1 |
6385594 | Lebda et al. | May 2002 | B1 |
6393409 | Young et al. | May 2002 | B2 |
6405179 | Rebane | Jun 2002 | B1 |
6405181 | Lent et al. | Jun 2002 | B2 |
6411947 | Rice et al. | Jun 2002 | B1 |
6415267 | Hagan | Jul 2002 | B1 |
6418419 | Nieboer et al. | Jul 2002 | B1 |
6424949 | Deaton et al. | Jul 2002 | B1 |
6430545 | Honarvar et al. | Aug 2002 | B1 |
6456983 | Keyes et al. | Sep 2002 | B1 |
6480850 | Veldhuisen | Nov 2002 | B1 |
6513019 | Lewis | Jan 2003 | B2 |
6553113 | Dhir et al. | Apr 2003 | B1 |
6567791 | Lent et al. | May 2003 | B2 |
6578015 | Haseltine et al. | Jun 2003 | B1 |
6606744 | Mikurak | Aug 2003 | B1 |
6609113 | O'Leary et al. | Aug 2003 | B1 |
6658393 | Basch et al. | Dec 2003 | B1 |
6704714 | O'Leary et al. | Mar 2004 | B1 |
6718388 | Yarborough et al. | Apr 2004 | B1 |
6804346 | Mewhinney | Oct 2004 | B1 |
6823319 | Lynch et al. | Nov 2004 | B1 |
6865547 | Brake, Jr. et al. | Mar 2005 | B1 |
6968319 | Remington et al. | Nov 2005 | B1 |
6970830 | Samra et al. | Nov 2005 | B1 |
6988082 | Williams et al. | Jan 2006 | B1 |
6999938 | Libman | Feb 2006 | B1 |
7068832 | Price et al. | Jun 2006 | B1 |
7139729 | Nault | Nov 2006 | B2 |
7143174 | Miller et al. | Nov 2006 | B2 |
7155614 | Ellmore | Dec 2006 | B2 |
7174302 | Patricelli et al. | Feb 2007 | B2 |
7206768 | deGroeve et al. | Apr 2007 | B1 |
7272857 | Everhart | Sep 2007 | B1 |
7310618 | Libman | Dec 2007 | B2 |
7353383 | Skingle | Apr 2008 | B2 |
7406426 | Pletz | Jul 2008 | B1 |
7444672 | Ellmore | Oct 2008 | B2 |
7461265 | Ellmore | Dec 2008 | B2 |
7472171 | Miller et al. | Dec 2008 | B2 |
7480631 | Merced et al. | Jan 2009 | B1 |
7580890 | Siegel et al. | Aug 2009 | B2 |
7587763 | Yodaiken | Sep 2009 | B2 |
8032936 | Yodaiken | Oct 2011 | B2 |
20010032158 | Starkman | Oct 2001 | A1 |
20010032159 | Starkman | Oct 2001 | A1 |
20010032176 | Starkman | Oct 2001 | A1 |
20010034663 | Teveler et al. | Oct 2001 | A1 |
20010034682 | Knight et al. | Oct 2001 | A1 |
20010042034 | Elliott | Nov 2001 | A1 |
20010044293 | Morgan | Nov 2001 | A1 |
20010044762 | Nault | Nov 2001 | A1 |
20010047342 | Cuervo | Nov 2001 | A1 |
20010047489 | Ito et al. | Nov 2001 | A1 |
20010054003 | Chien et al. | Dec 2001 | A1 |
20010056390 | Varadarajan et al. | Dec 2001 | A1 |
20020007313 | Mai et al. | Jan 2002 | A1 |
20020026394 | Savage et al. | Feb 2002 | A1 |
20020059141 | Davies et al. | May 2002 | A1 |
20020120570 | Loy | Aug 2002 | A1 |
20020194126 | Randell et al. | Dec 2002 | A1 |
20030101119 | Parsons et al. | May 2003 | A1 |
20030105981 | Miller et al. | Jun 2003 | A1 |
20030163403 | Chen et al. | Aug 2003 | A1 |
20030208441 | Poplawski et al. | Nov 2003 | A1 |
20030216997 | Cohen | Nov 2003 | A1 |
20030233321 | Scolini et al. | Dec 2003 | A1 |
20040024693 | Lawrence | Feb 2004 | A1 |
20040078328 | Talbert et al. | Apr 2004 | A1 |
20040123137 | Yodaiken | Jun 2004 | A1 |
20040200898 | Kepecs | Oct 2004 | A1 |
20040230483 | Kepecs | Nov 2004 | A1 |
20040236688 | Bozeman | Nov 2004 | A1 |
20040252823 | Becerra et al. | Dec 2004 | A1 |
20040255223 | Chawla | Dec 2004 | A1 |
20050091138 | Awatsu | Apr 2005 | A1 |
20050102212 | Roy | May 2005 | A1 |
20060036553 | Gupta et al. | Feb 2006 | A1 |
20080027841 | Eder | Jan 2008 | A1 |
20080052328 | Widhelm et al. | Feb 2008 | A1 |
20080121696 | Mock et al. | May 2008 | A1 |
20080304431 | Karaoguz | Dec 2008 | A1 |
20090261162 | Kargman et al. | Oct 2009 | A1 |
20100011444 | Yodaiken | Jan 2010 | A1 |
20100281223 | Wolfe et al. | Nov 2010 | A1 |
Number | Date | Country |
---|---|---|
7-152960 | Jun 1995 | JP |
2007-088822 | Apr 2007 | JP |
WO 0186524 | Nov 2001 | WO |
Entry |
---|
Delivering the Right Information to the Right Resource or Every Customer Interaction; Intelligent Callrouter, www.geotel.com/ solutions/icr/default/htm, 1 page, (1998). |
Friedman, Dictionary of Business Terms, Third Edition, (2000). |
Forbes, S I Eve, Fact and Comment, ProQuest, vol. 170, Issue 6, Sep. 30, 2002. |
Global Corruption Report 2004, Transparency International, Pluto Press, www.globalcorrupt, ISBN 07453 2231, Jun. 26, 2005. |
Rial, Astrid, How to Monitor Collectors, Credit Card Management, Jul. 2000, vol. 13, Iss. 3:p. 65, 4 pages. |
Keep the Change Savings Service, Bank of America, Retrieved from the internet on Jan. 27, 2006 at <https://www.bankofamerica.com/deposits/checksave/apps/ktc/ktc—terms.cfm>. |
Computer Telephony Solutions, The Customer Interaction Specialists, Computer Telephony Solutions, Internet, May 25, 1999. |