Identifying and filtering incoming telephone calls to enhance privacy

Information

  • Patent Grant
  • 10135976
  • Patent Number
    10,135,976
  • Date Filed
    Tuesday, July 12, 2016
    8 years ago
  • Date Issued
    Tuesday, November 20, 2018
    5 years ago
Abstract
A method for filtering a telephone call is provided. The method may comprise receiving from a caller the telephone call directed to a communication device associated with an intended call recipient. The received telephone call may be scored based on predetermined scoring criteria to create a score indicative of a desirability of the telephone call. Furthermore, the method may comprise comparing the score to a predetermined threshold score. Based on the comparison, the method may further comprise selectively classifying the telephone call as an unwanted telephone call. Furthermore, the method may comprise selectively rejecting the unwanted telephone call.
Description
TECHNICAL FIELD

This disclosure relates generally to privacy protection of telephone calls and, more particularly, to blocking of unwanted incoming telephone calls.


BACKGROUND

The approaches described in this section could be pursued but are not necessarily approaches that have previously been conceived or pursued. Therefore, unless otherwise indicated, it should not be assumed that any of the approaches described in this section qualify as prior art merely by virtue of their inclusion in this section.


With development of telephone communications, the number of parties offering services via telephone has increased. Telephone calls from these parties, such as telemarketers, pollsters, and charitable organizations, sometimes referred to as spam calls, robocalls, or junk calls, are typically unsolicited, placed in bulk and made indiscriminately. Telemarketers employ increasingly sophisticated technology to get their calls through to their target audience in spite of people's desire to be left alone. Furthermore, telemarketers can employ computers to dial a number until the call is answered, so they can automatically transfer the call to a human agent. This can be heard as a pause after the phone call is answered but before the called party can speak to the human agent.


Consumers can employ a variety of defenses against undesired callers. Some telephone services can allow anonymous call reject so that the phone does not ring if the caller Identification (ID) is blocked. Unfortunately, this strategy can prevent the receipt of calls from anyone who has set their outbound caller ID to be blocked in an attempt to maintain privacy. Legislation has been passed allowing consumers to opt out of receiving calls from telemarketers by adding their phone number to a national “do-not-call” list. However, the effectiveness of this list has been undermined by the fact that it does not apply to political organizations and charities, many telemarketers ignore the list, and the prohibition does not apply to any entity that has previously conducted business with the target. Some privacy protection techniques can employ a “whitelist” of known acceptable callers and a “blacklist” of known undesired callers, but blocking only blacklisted numbers often lets too many “junk” calls through due to the blacklist being incomplete, while allowing only calls from whitelisted numbers may be too restrictive. Moreover, managing blacklists and whitelists manually can be bothersome and often is not worth the effort.


In the meantime, computational power and network databases have advanced so much that a network operator can employ far more sophisticated and automated means for protecting privacy of the user more accurately and with minimal interaction with the user. Current methods of spam blocking are inadequate and a new approach to privacy protection is needed in the field.


SUMMARY

This summary is provided to introduce a selection of concepts in a simplified form that are further described in the Detailed Description below. This summary is not intended to identify key features or essential features of the claimed subject matter, nor is it intended to be used as an aid in determining the scope of the claimed subject matter.


The present disclosure is related to various techniques for filtering telephone calls. Specifically, a method for filtering a telephone call may comprise receiving, from a caller, a telephone call directed to a communication device associated with an intended call recipient. The received telephone call may be scored based on predetermined scoring criteria to create a score indicative of a desirability of the telephone call. Furthermore, the method may comprise comparing the score to a predetermined threshold score. The method may further comprise selectively classifying, based on the comparison, the telephone call as an unwanted telephone call. Furthermore, the method may comprise selectively rejecting the unwanted telephone call.


According to another approach of the present disclosure, there is provided a system for filtering a telephone call. The system may comprise a processor. The processor may be configured to receive, from a caller, the telephone call directed to a communication device associated with an intended call recipient. The processor may be further configured to score the telephone call based on predetermined scoring criteria to create a score indicative of a desirability of the telephone call. Furthermore, the processor may compare the created score to a predetermined threshold score and, based on the comparison, selectively classify the telephone call as an unwanted telephone call. Furthermore, the processor may selectively reject the unwanted telephone call. The system may further comprise a memory unit configured to store data associated with the predetermined scoring criteria, the predetermined threshold score, a whitelist, and a blacklist.


In further example embodiments of the present disclosure, the method steps are stored on a machine-readable medium comprising instructions, which when implemented by one or more processors perform the recited steps. In yet further example embodiments, hardware systems, or devices can be adapted to perform the recited steps. Other features, examples, and embodiments are described below.





BRIEF DESCRIPTION OF THE DRAWINGS

Embodiments are illustrated by way of example, and not by limitation, in the figures of the accompanying drawings, in which like references indicate similar elements and in which:



FIG. 1 shows an environment within which a method and a system for filtering a telephone call can be implemented, according to an example embodiment.



FIG. 2 is a process flow diagram showing a method for filtering a telephone call, according to an example embodiment.



FIG. 3 is a flow chart showing handling of an inbound call, according to an example embodiment.



FIG. 4 is a block diagram showing various modules of a system for filtering a telephone call, according to an example embodiment.



FIG. 5 is a flow chart showing logic for creating a whitelist, according to an example embodiment.



FIG. 6 is a flow chart showing logic for creating a blacklist, according to an example embodiment.



FIG. 7 is a flow chart showing logic for creating a graylist, according to an example embodiment.



FIG. 8 is a flow chart showing logic for handling of rejected calls, according to an example embodiment.



FIG. 9 illustrates a user interface for presenting a score of the caller, according to an example embodiment.



FIG. 10 shows a diagrammatic representation of a computing device for a machine in the example electronic form of a computer system, within which a set of instructions for causing the machine to perform any one or more of the methodologies discussed herein can be executed.





DETAILED DESCRIPTION

The following detailed description includes references to the accompanying drawings, which form a part of the detailed description. The drawings show illustrations in accordance with example embodiments. These example embodiments, which are also referred to herein as “examples,” are described in enough detail to enable those skilled in the art to practice the present subject matter. The embodiments can be combined, other embodiments can be utilized, or structural, logical, and electrical changes can be made without departing from the scope of what is claimed. The following detailed description is therefore not to be taken in a limiting sense, and the scope is defined by the appended claims and their equivalents. In this document, the terms “a” and “an” are used, as is common in patent documents, to include one or more than one. In this document, the term “or” is used to refer to a nonexclusive “or,” such that “A or B” includes “A but not B,” “B but not A,” and “A and B,” unless otherwise indicated.


The techniques of the embodiments disclosed herein may be implemented using a variety of technologies. For example, the methods described herein may be implemented in software executing on a computer system or in hardware utilizing either a combination of microprocessors or other specially designed application-specific integrated circuits (ASICs), programmable logic devices, or various combinations thereof. In particular, the methods described herein may be implemented by a series of computer-executable instructions residing on a storage medium such as a disk drive, or computer-readable medium. It should be noted that methods disclosed herein can be implemented by a computer (e.g., a desktop computer, a tablet computer, a laptop computer, and a server), game console, handheld gaming device, cellular phone, smart phone, smart television system, and so forth.


The present disclosure relates to a network-oriented, integrated and automated approach to avoid unwanted calls by exploiting the capabilities of a phone network and to provide sophisticated handling of incoming calls. In addition to user-defined blacklists (i.e., lists comprising numbers known to be associated with undesired callers) and whitelists (i.e., lists comprising numbers known to be associated with acceptable callers), the present disclosure implements processes for enhancing and automating creation of blacklists and whitelists and for scoring the likelihood of allowance of a call from an unfamiliar (but not anonymous) caller. The resulting scores may be used to provide the user with a control over calls from unfamiliar callers.


Various criteria may be used to score calls. For example, call desirability may be assessed based on network statistics gathered across the entire user base of the network, community whitelists and blacklists, and callbacks. Furthermore, machine detection may be employed to determine whether the caller is an automated system or a human. In an example embodiment, customized challenge and response may be used to provide the user with an opportunity to determine whether unfamiliar callers are known to the user and whether calls from the unfamiliar callers should be put through or blocked.


Thus, as a result of filtering a telephone call directed to privacy protection of the user, the call may be either rejected or put through to the user. During the scoring process, callers can be added to the community blacklist or whitelist or to a personal blacklist or whitelist of the person being called. If the call is put through, the system may provide the user with a user interface to allow the system to add the caller to the blacklist or the whitelist to streamline handling of calls from the same caller in the future. Alternatively, additions to the blacklist and whitelist may be automated based on programed rules prescribing flow of the call through the privacy protection system. Individual blacklists and whitelists may serve as resources to help determine the routing of future calls from the same callers to other users of the network. Additionally, preventive actions may be performed with regard to the undesired callers, such as automatically logging a complaint to an authority responsible for telephone call privacy identifying the caller as violating provisions related to the national “do-not-call” list. For the calls rejected by the system, the user may have several options for treating those calls to maximize the cost to the caller for having called, thereby increasing the chances that the caller will not call again and providing a measure of satisfaction to the user.


Referring now to the drawings, FIG. 1 illustrates an environment 100 within which a system and a method for filtering a telephone call can be implemented. The environment 100 may include a network 110, a system 400 for filtering a telephone call, a caller 120, and a user 130 being an intended call recipient and having one or more user communication devices 140. The network 110 may include the Internet or any other network capable of communicating data between devices. Suitable networks may include or interface with any one or more of, for instance, a local intranet, a PAN (Personal Area Network), a LAN (Local Area Network), a WAN (Wide Area Network), a MAN (Metropolitan Area Network), a virtual private network (VPN), a storage area network (SAN), a frame relay connection, an Advanced Intelligent Network (AIN) connection, a synchronous optical network (SONET) connection, a digital T1, T3, E1 or E3 line, Digital Data Service (DDS) connection, DSL (Digital Subscriber Line) connection, an Ethernet connection, an ISDN (Integrated Services Digital Network) line, a dial-up port such as a V.90 V.34 or V.34bis analog modem connection, a cable modem, an ATM (Asynchronous Transfer Mode) connection, or an FDDI (Fiber Distributed Data Interface) or CDDI (Copper Distributed Data Interface) connection. Furthermore, communications may also include links to any of a variety of wireless networks, including WAP (Wireless Application Protocol), GPRS (General Packet Radio Service), GSM (Global System for Mobile Communication), CDMA (Code Division Multiple Access) or TDMA (Time Division Multiple Access), cellular phone networks, GPS (Global Positioning System), CDPD (cellular digital packet data), RIM (Research in Motion, Limited) duplex paging network, Bluetooth radio, or an IEEE 802.11-based radio frequency network. The network 110 can further include or interface with any one or more of an RS-232 serial connection, an IEEE-1394 (Firewire) connection, a Fiber Channel connection, an IrDA (infrared) port, a SCSI (Small Computer Systems Interface) connection, a USB (Universal Serial Bus) connection or other wired or wireless, digital or analog interface or connection, mesh or Digi® networking. The network 110 may include a network of data processing nodes that are interconnected for the purpose of data communication. Furthermore, the network 110 may include a telephone network, such as a voice over Internet Protocol (VoIP) network.


The user communication device 140 may include a mobile phone, a smartphone, a tablet Personal Computer (PC), a laptop, a PC, and so forth. For the purposes of communication, the user communication device 140 may be compatible with one or more of the following network standards: GSM, CDMA, Long Term Evolution (LTE), IMS, Universal Mobile Telecommunication System (UMTS), 4G, 5G, 6G and upper, Radio-Frequency Identification (RFID), and so forth.


As shown, the caller 120 may perform a telephone call to the user communication device 140 via the network 110. The system 400 may be associated with the user communication device 140 to filter telephone calls from unwanted callers.



FIG. 2 is a flow chart illustrating a method 200 for filtering a telephone call, in accordance with an example embodiment. The method 200 may be performed by processing logic that may comprise hardware (e.g., decision making logic, dedicated logic, programmable logic, and microcode), software (such as software run on a general-purpose computer system or a dedicated machine), or a combination of both.


In the example shown, the method 200 may commence at operation 202 with receiving, by a processor, from a caller, the telephone call directed to a communication device associated with an intended call recipient. The method 200 may continue with operation 204, at which the processor may score the telephone call based on predetermined scoring criteria to create a score indicative of a desirability of the telephone call.


In an example embodiment, the scoring may be based on whether a caller or a caller communication device is associated with a whitelist, a blacklist, or a graylist. Furthermore, the scoring may be based on a number of calls made to different phone numbers by the caller, sequence of the calls made by the caller, and call sources. The predetermined scoring criteria may involve network statistics, machine detections, customized challenge and response, and so forth. In some embodiments, the predetermined scoring criteria may be provided by the intended call recipient.


At operation 206, the processor may compare the created score to a predetermined threshold score. In an example embodiment, the predetermined threshold score may be selected by the intended call recipient or set by the processor. The method 200 may continue to operation 208, at which, based on the comparison, the processor may selectively classify the telephone call as an unwanted telephone call.


At operation 210, the processor may selectively reject the unwanted telephone call. In an example embodiment, the telephone call may be automatically rejected if a caller ID associated with the telephone call is blocked or the call is anonymous. In an example embodiment, the method 200 may optionally comprise, based on the score, putting the telephone call through to the communication device.


Furthermore, based on the score, the caller or a caller communication device may be added to a whitelist or a blacklist. In an example embodiment, the intended call recipient may manually add the caller or caller communication device to the blacklist or the whitelist.


In an example embodiment, the method 200 may optionally comprise providing screening of the telephone call by the intended call recipient to allow the intended call recipient to manually reject or allow the telephone call. Furthermore, the intended call recipient may be allowed to grade the telephone call or effectiveness of the screening after the telephone call is completed.


In an example embodiment, the method 200 may comprise automatically requiring a call back number to complete the telephone call. In an example embodiment, the method 200 may optionally comprise providing a test to establish whether the caller is a human. The method 200 is further illustrated in detail in FIG. 3 showing a scheme 300 for handling of an inbound call. The inbound call may include a call from a caller to an intended call recipient before being put through to the intended call recipient and before being rejected.


In a telephone network, such as a VoIP network, at step 301 it may be determined that the caller ID associated with the inbound call is blocked or anonymous. In such a case, at step 303, it may be ascertained whether the intended call recipient has enabled anonymous call reject (ACR). In case the intended call recipient has enabled ACR, the call may be rejected at step 306. If, on the other hand, the intended call recipient has not enabled ACR, it may be ascertained, at step 309, whether further screening, namely a callback option, is enabled by the intended call recipient. If the callback option is not enabled for the inbound anonymous calls, the call may be completed at step 312. In some example embodiments, if it is determined, at step 309, that the callback option is not enabled for the inbound anonymous calls, method 300 may proceed to step 307 (described below) instead of completing the call directly.


If further screening is desired, the call flow may be passed to step 311 where a callback may be attempted. In this case, the call may be answered and the caller asked by an interactive voice response system (IVR) to provide a number at which the caller may be reached. Typically, undesired callers, such as telemarketers, may not have direct phone numbers at which they can be reached. Therefore, requiring the callback may help in overcoming spam attempts. In case of receiving the number at which the caller may be reached, the current call may be ended and the process may proceed with step 302 described below.


It is important to note that for telephone networks in general, and VoIP networks in particular, functions such as callbacks and IVRs can be implemented without tying up the phone line of the intended call recipient or inhibiting the use of the service by the intended call recipient. From the perspective of the intended call recipient, a system for filtering a telephone call may operate in the background and in general may be invisible to the intended call recipient.


For an inbound call where a caller ID may be available (or where a callback number has been obtained), the whitelist of previously identified acceptable callers of the intended call recipient may be queried at step 302. If the caller is on the whitelist, the call may be completed at step 305. If the caller is not on the white list, the blacklist associated with the intended call recipient may be queried at step 304. A call from the callers on the blacklist may be rejected at step 308. For the callers who are neither on the whitelist nor on the blacklist, the desirability of the caller can be established in step 307. In step 307, the call may be answered by the automated system with a greeting such as ‘Who is calling?’ The system purposely may have a short greeting asking a question so that it may measure the pause between the connecting of the call and the first words of the caller. If the pause is three seconds or longer, the call may be connected by an auto dialer and the telemarketer may never hear the greeting from the system. If the caller responds with ‘Hello’ it may be an indication that he did not hear the question because of the time lag introduced by the auto dialer. This may be the first measurement to be considered in grading the desirability of the caller. The desirability of the caller may be scored, and the resulting score may be compared with a predetermined threshold score at step 310. The predetermined threshold score may be defined by the system or selected by the intended call recipient. If the desirability score of the caller is higher than the predetermined threshold score, the caller may be classified as a desirable caller and the call may be put through to the intended call recipient at step 313. (See FIG. 7 and discussion below for further details of the desirability rating.)


At step 314, if the desirability score of the caller is below the predetermined threshold score, the intended call recipient may configure the system to provide further assurances that the call is worth taking by attempting to determine whether the caller is human and not an automated system. In an example embodiment, the determining whether the caller is human may be based on the IVR asking the caller to add two numbers. The score indicative of whether the caller is human may range from 1-100. This score may be used at step 315 to determine whether to reject the call. If the score associated with the caller fails to exceed the threshold score value of step 315, the call may be rejected and the caller may be added to the blacklist at step 316.


If the caller is determined to be human, the system may be configured to issue a customized challenge/response at step 317 to determine whether the caller should be allowed to complete the call. The customized challenge/response, employing the IVR and call recording mechanism, may be used to determine whether the caller is an acquaintance of the intended call recipient or to classify the caller as an undesired caller, such as a collection agent or a solicitor. The interactions of the challenge/response may be recorded for future use in documenting the violation by the caller of the rules related to the national “do-not-call” list. A customized challenge/response may comprise the IVR asking a question or series of questions of a personal nature that could only be answered by an acquaintance of the intended call recipient. The answer may require personal knowledge of the intended call recipient that an undesired caller, such as a telemarketer, may not possess. A small number of such questions could determine familiarity and, therefore, legitimacy of the caller at step 318. A call from an illegitimate caller may be rejected at step 321, while a call from a legitimate caller may be put through to the intended call recipient at step 319. With proper whitelist handling, only one pass through the question and answer process may be required for any legitimate caller. For subsequent calls from that caller, the whitelist may identify the caller as legitimate and the call may be connected without delay.


At step 320, once the call has been completed via any path through step 307, the intended call recipient may be given an opportunity, through an intended call recipient interface, to grade effectiveness of the screening. This feedback may allow the intended call recipient to put the caller onto a personal whitelist or blacklist in order to streamline processing of future calls from that caller.



FIG. 4 is a block diagram showing various modules of an exemplary system 400 for filtering telephone calls. Specifically, the system 400 may include a processor 402. The processor 402 may be configurable to receive a telephone call from a caller directed to a communication device associated with an intended call recipient. The processor 402 may score the received telephone call based on predetermined scoring criteria and create a score indicative of a desirability of the telephone call. The processor 402 may further compare the score to a predetermined threshold score. The predetermined threshold score may be selected by the intended call recipient or set by the processor 402. Based on the comparison, the processor 402 may be configured to selectively classify the telephone call as an unwanted telephone call and to selectively reject the unwanted telephone call.


In an example embodiment, the processor may be configured, based on the score, to selectively classify the telephone call as a wanted telephone call and put the telephone call through to the communication device of the intended call recipient. In an example embodiment, the processor may be configured, based on the score, to add the caller or a caller communication device to the whitelist or the blacklist.


In an example embodiment, the processor may be configured to automatically require a call back number to complete the telephone call. If the caller is unable to provide a call back number, the processor may be configured to reject the call.


In an example embodiment, the processor may be configured to provide a test to establish whether the caller is a human. If the caller passes the test, the processor may put the call through to the intended call recipient. If the caller fails the test, the processor may reject the call.


In a further example embodiment, the processor may be further configured to provide screening of the telephone call by the intended call recipient to allow the intended call recipient to manually reject or allow the telephone call. The intended call recipient may be further allowed to grade the telephone call or effectiveness of the screening after the telephone call is completed.


The system 400 may comprise a memory unit 404 configured to store data associated with the predetermined scoring criteria, the predetermined threshold score, a whitelist, a blacklist, and so forth.



FIG. 5 shows a flow chart 500 illustrating how the whitelist may be composed and evaluated for driving the decision in step 302 of FIG. 3. The whitelist may consist of numbers from which a user (i.e., an intended call recipient) wishes to receive calls. The user may be given several options as to what numbers may be included in the whitelist. Alternatively, the system 400 may be configured with a set of predetermined options for including a number in the whitelist. The list of options may include user generated numbers 502 (i.e., the numbers that the user enters as permitted numbers). Furthermore, the list of options may include numbers previously called by the user 503. The list of options may further include previously accepted numbers 504 (i.e., the numbers from callers whose calls were previously accepted by the user and were not subsequently removed from the whitelist or blacklisted). Furthermore, the list of options may include user rule numbers 505 (i.e., numbers from particular geographic locations, area codes or number wildcards, e.g., calls from a home town of the user, as determined by the area code). The list of options may comprise community list numbers 506 (i.e., numbers present in a public list).


The list of options may further include numbers calling special phone numbers 507 to which the user may give priority. For example, a user may have a virtual number that the user may provide to family or acquaintances known to be legitimate. The calls to this number may bypass the screening process entirely even for unfamiliar or blocked calling numbers.



FIG. 5 shows how various whitelists can be qualified against time/date criteria 508, such as time of day, day of week, or time of year; geographic location 510 of the caller as determined by the caller ID; or user status 509. These various lists and criteria may be combined in logic block 501 to produce a decision 511 as to whether the caller should definitely be allowed to complete the call.



FIG. 6 shows a flow chart 600 illustrating how the blacklist may be composed and evaluated for driving the decisions of step 304 of FIG. 3. The blacklist may consist of caller numbers associated with callers from whom the user may not want to receive calls. The user may be given several options for what numbers may go into the blacklist. The list of options may include user generated numbers 602 (i.e., numbers that the user may enter as blacklisted). The list of options may further include community list numbers 603 (i.e., numbers entered in a community blacklist that may be generated by the network operator or crowd-sourced from a community of users). The community blacklist may include undesired numbers from multiple sources, including numbers known to be used for calls to large numbers of subscribers in the network. Most residential telephone users may call fewer than 20 different numbers in a given month. Therefore, an inbound caller who has called, for example, 100 or even 1,000 different numbers on a single telephone network in a given month can reasonably be classified as a spam caller. Furthermore, the community blacklist may include numbers known from network calling records to be used for calling phone numbers in sequence.


In an example embodiment, the community blacklist may include numbers blacklisted in multiple blacklists of individual users, numbers found in global blacklists published by third party or public sources, and numbers that have previously called “honeypot” numbers. The “honeypot” numbers may be designed by the network operator to entice telemarketers and other undesired callers to call and engage in a manner which allows their intent to be revealed. For example, certain unassigned phone numbers may be connected to an IVR system which pretends to be a live caller asking a question. The IVR may then measure the pause before the caller responds, which may often be longer than usual when an automated telemarketing system connects the call. If the response detected does not begin with a predetermined phrase, then it may be concluded that the caller may not have heard the question due to the lag typically seen in automated dialers between call connection and an agent joining the call. This may mean that the caller may be an undesired caller, such as a telemarketer, and that the number should be blacklisted. Numbers that are known to have been out of service for a long time are also effective “honeypots” to attract undesired callers as it can be reasonably guessed that anyone calling a number that has been out of service for a long time may be either an undesired caller or has misdialed.


The list of options may further include previously blocked numbers 604 (i.e., numbers that have been previously blocked by the user). Furthermore, the list of options may include user rule numbers 605 (i.e., numbers from particular geographic locations, area codes or number wildcards for which the user has created a rule that causes them to be blocked).


Furthermore, the list of options may include numbers calling special phone numbers 606 that the user may provide only to persons from whom the user may never want to receive calls or whose calls would be sent directly to voicemail, and so forth. For example, such a number may be provided to a person with whom the user may not wish to continue a relationship.



FIG. 6 shows how the various blacklists can be further qualified by time of day, day of week, time of year 607, user status 608 provided by the user, or geographic location 609 of the caller as determined by the caller ID. A function can be generalized and integrated with a Do Not Disturb 610 function to give users greater control over their privacy. Do Not Disturb 610 can be further customized with a scheduling system that may identify personal preferences as to when calls may or may not be accepted. Once a call has passed the rejection tests, the call may be compared to the schedule before completion. The calls outside the assigned time windows may have their destination assigned in the same fashion as rejected calls or may be forwarded to a voicemail. The scheduled Do Not Disturb may have an Always OK list to allow certain numbers to call outside the scheduled times. The various lists and criteria may combine in logic block 601 to produce a decision 611 as to whether the caller may definitely be prevented from completing the call.



FIG. 7 shows a flow chart 700 illustrating how callers, who are neither whitelisted nor blacklisted, may be graded and assigned a score to determine if calls from the callers may be put through to the user. Unlike the definitive judgments made by the whitelist logic of FIG. 5 or blacklist logic of FIG. 6, the assessment of FIG. 7 may be considered as a “graylist” logic 701 in which judgment is not binary. Factors that affect the graylist score 707 or caller desirability score may be heuristic and dependent upon a wide variety of data sources, some of which were mentioned previously and which may also drive the whitelist or blacklist judgments (where the user has chosen to treat them as if they were binary factors). For example, some users may take an aggressive view toward any number that has called a honeypot and simply disallow calls from those numbers. Other users may choose not to include that factor in the blacklist determination and instead incorporate the factor into the graylist score 707. The factors which are particularly suited to influencing the graylist score 707 may include a call frequency factor 702 (i.e., the frequency with which the caller has called a network telephone number today or in the past week or month). If a network operator with several hundred thousand phone numbers detects that the same caller is calling many of its users over a short period of time, then this may be a strong indication that the caller is a junk caller and, therefore, an undesired caller. For a typical network with high affinity (the tendency for users who are acquaintances to share the same telephone service), a typical residential user may call fewer than 25 numbers per million subscribers in a given month. In the current embodiment, a graylist score 707 may be set to equal the number of different on-network phone numbers called by the user in the last 30 days, minus 25.


The factors influencing the graylist score 707 may further include a call locality factor 703. If a caller calls numbers in sequence, that may be an indication of a junk caller. In an example embodiment, the factor may be an integer indicating the length of the longest run of sequentially dialed numbers in the last 30 days. In an example embodiment, this graylist component score may be set to 20 times this number. Accordingly, a caller who dialed five phone numbers in sequence may get a graylist score 707 of 100 for just this component and the caller may be almost certainly a junk caller and, therefore, an undesired caller.


A blacklist prevalence factor 704 may mean that the caller has been blacklisted by many other users, which may be an indication of a junk caller. In an example embodiment, this graylist component score may be set to be 10 times an integer indicating the number of users in the network who have this number in their personal blacklist.


A whitelist prevalence factor 705 may mean that a caller is found on many user whitelists or has been called frequently by users on the network, which may be an indication that the call is valuable and should not be blocked. In an example embodiment, this graylist component score may be set to be −10 times an integer indicating the number of users in the network who have this number in their personal whitelist.


A call source type factor 706 may include calls from mobile numbers that are unlikely to be from junk callers, as most telemarketers operate from fixed locations using landlines instead of mobile phones. In an example embodiment, this graylist component score may be set to be −50 if the call is coming from a mobile phone number.


According to an example embodiment, each of the factors above may be weighted and added or subtracted as they apply to a particular caller. The resulting score may be normalized in a range from 0 to 100 and then passed to the rest of the screening process.



FIG. 8 is a flow chart 800 illustrating the treatment for rejected calls. Since the system described herein effectively filters calls without any disturbance of the user, it is possible to engage in a counterattack against telemarketers for calls that are rejected in order to discourage their activities and make them less profitable when calling numbers on the protected network. The system may provide the user with several options for handling rejected callers.


The handling options may include a dial tone option 802 so that the rejected calls are terminated with a dial tone. Furthermore, in case of a voicemail option 803, rejected calls may be passed to voicemail. The voicemail may be flagged so that the user does not waste time listening to the message. Alternatively, the user may want to listen to such messages to make sure that the settings are not inadvertently blocking calls from legitimate users.


An out-of-service tone option 804 may be used, according to which rejected calls may be given the out-of-service tone, which may cause some automatic dialing systems to remove the number from their phone number list.


This call treatment may prevent the undesired caller from wasting resources on a call that will never get through to an intended user. Furthermore, according to an example embodiment of option 805, there is no need to ring the phone for screened calls.


Furthermore, an IVR Turing test option 806 may be used to direct rejected calls to an IVR that simulates human conversation to prolong the time wasted by the undesired caller. For example, the IVR may answer the phone by asking the caller to wait for several minutes followed by a long pause. This may be followed by a Turing test wherein the challenge of the undesired caller may be to determine whether the IVR is a human or a machine while the purpose of the IVR challenge may be to prolong the conversation for as long as possible. Various personalities may be selected by the user for their Turing tests, for example a confused person whose conversation wanders, a drug addict who tries to take a message and forgets what he is doing and starts over, a dog owner who excuses himself to try to quiet the hound with barking in the background, and so forth.


A variation of the Turing Test call treatment may be an IVR Evidence 807 in which a further objective of the IVR may be to isolate the name of the caller and to determine whether the caller is covered by the constraints of the national “do not call” list. If the caller is covered, then the system may record the call as evidence for a complaint to an authority for the violation.


Selection 811 of which rejected call treatment to use may be based on what factors of logic block 801 caused the call to be rejected in the first place. For example, if the call is rejected because the caller is found on a user-generated blacklist 809 (implying that the user is quite confident that the call is from an undesired caller), then a more aggressive call treatment such as the Turing test may be employed. If, however, the call is rejected because of a marginal score 810 from the graylist scoring engine, then the call may be sent to voicemail in case a mistake was made in the scoring. Selection 811 may be also based on whether the caller is found on a user-generated whitelist 808. Selecting a call treatment may be as simple as directing the call to a server (such as a voicemail server or an announcement server) that connects the call and implements the desired treatment. Such servers are commonly employed using FreeSwitch or Asterisk soft switches.



FIG. 9 illustrates a user interface 900, according to an example embodiment. The user interface 900 may present the information gathered by the system that may be relevant to a user deciding to answer a call. This information may be presented on the screen of a PC or, in compressed form, on the screen of a Digital Enhanced Cordless Telephone (DECT) handset or smartphone. The objective of the user interface 900 may be to provide the user with as much information as possible that may be relevant to the decision of the user to accept the call. The location 901 of the caller may be approximated based on the originating phone number or IP address (in the case of a VoIP call). Statistics 902 for previous calls, such as a number of calls over a given time period, or what their duration was, may be presented. The number 903 of calls to the service provider network may be provided along with their average call duration. Finally, the privacy score 904 from the graylist calculator might be provided. Notes 905 associated with the caller may be presented. Furthermore, on-screen options to whitelist or blacklist the caller may be presented.



FIG. 10 shows a diagrammatic representation of a machine in the example electronic form of a computer system 1000, within which a set of instructions for causing the machine to perform any one or more of the methodologies discussed herein may be executed. In various example embodiments, the machine operates as a standalone device or may be connected (e.g., networked) to other machines. In a networked deployment, the machine may operate in the capacity of a server or a client machine in a server-client network environment, or as a peer machine in a peer-to-peer (or distributed) network environment. The machine may be a PC, a tablet PC, a set-top box (STB), a cellular telephone, a portable music player (e.g., a portable hard drive audio device such as an Moving Picture Experts Group Audio Layer 3 (MP3) player), a web appliance, a network router, switch or bridge, or any machine capable of executing a set of instructions (sequential or otherwise) that specify actions to be taken by that machine. Further, while only a single machine is illustrated, the term “machine” shall also be taken to include any collection of machines that individually or jointly execute a set (or multiple sets) of instructions to perform any one or more of the methodologies discussed herein.


The example computer system 1000 includes a processor or multiple processors 1002 (e.g., a central processing unit (CPU), a graphics processing unit (GPU), or both), a main memory 1004 and a static memory 1006, which communicate with each other via a bus 1008. The computer system 1000 may further include a video display unit 1010 (e.g., a liquid crystal display (LCD) or a cathode ray tube (CRT)). The computer system 1000 may also include an alphanumeric input device 1012 (e.g., a keyboard), a cursor control device 1014 (e.g., a mouse), a disk drive unit 1013, a signal generation device 1028 (e.g., a speaker), and a network interface device 1020.


The disk drive unit 1013 includes a non-transitory computer-readable medium 1022, on which is stored one or more sets of instructions and data structures (e.g., instructions 1024) embodying or utilized by any one or more of the methodologies or functions described herein. The instructions 1024 may also reside, completely or at least partially, within the main memory 1004 and/or within the processors 1002 during execution thereof by the computer system 1000. The main memory 1004 and the processors 1002 may also constitute machine-readable media.


The instructions 1024 may further be transmitted or received over a network 1026 via the network interface device 1020 utilizing any one of a number of well-known transfer protocols (e.g., Hyper Text Transfer Protocol (HTTP)).


In some embodiments, the computer system 1000 may be implemented as a cloud-based computing environment, such as a virtual machine operating within a computing cloud. In other embodiments, the computer system 1000 may itself include a cloud-based computing environment, where the functionalities of the computer system 1000 are executed in a distributed fashion. Thus, the computer system 1000, when configured as a computing cloud, may include pluralities of computing devices in various forms, as will be described in greater detail below.


In general, a cloud-based computing environment is a resource that typically combines the computational power of a large grouping of processors (such as within web servers) and/or that combines the storage capacity of a large grouping of computer memories or storage devices. Systems that provide cloud-based resources may be utilized exclusively by their owners or such systems may be accessible to outside users who deploy applications within the computing infrastructure to obtain the benefit of large computational or storage resources.


The cloud may be formed, for example, by a network of web servers that comprise a plurality of computing devices, with each server (or at least a plurality thereof) providing processor and/or storage resources. These servers may manage workloads provided by multiple users (e.g., cloud resource customers or other users). Typically, each user places workload demands upon the cloud that vary in real-time, sometimes dramatically. The nature and extent of these variations typically depends on the type of business associated with the user.


It is noteworthy that any hardware platform suitable for performing the processing described herein is suitable for use with the technology. The terms “computer-readable storage medium” and “computer-readable storage media” as used herein refer to any medium or media that participate in providing instructions to a CPU for execution. Such media can take many forms, including, but not limited to, non-volatile media, volatile media and transmission media. Non-volatile media include, for example, optical or magnetic disks, such as a fixed disk. Volatile media include dynamic memory, such as system RAM. Transmission media include coaxial cables, copper wire and fiber optics, among others, including the wires that comprise one embodiment of a bus. Transmission media can also take the form of acoustic or light waves, such as those generated during radio frequency (RF) and infrared (IR) data communications. Common forms of computer-readable media include, for example, a floppy disk, a flexible disk, a hard disk, magnetic tape, any other magnetic medium, a CD-ROM disk, digital video disk (DVD), any other optical medium, any other physical medium with patterns of marks or holes, a RAM, a PROM, an EPROM, an EEPROM, a FLASHEPROM, any other memory chip or data exchange adapter, a carrier wave, or any other medium from which a computer can read.


Various forms of computer-readable media may be involved in carrying one or more sequences of one or more instructions to a CPU for execution. A bus carries the data to system RAM, from which a CPU retrieves and executes the instructions. The instructions received by system RAM can optionally be stored on a fixed disk either before or after execution by a CPU.


Computer program code for carrying out operations for aspects of the present technology may be written in any combination of one or more programming languages, including an object oriented programming language such as Java, Smalltalk, C++ or the like and conventional procedural programming languages, such as the “C” programming language or similar programming languages. The program code may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer or entirely on the remote computer or server. In the latter scenario, the remote computer may be connected to the user's computer through any type of network, including a local area network (LAN) or a wide area network (WAN), or the connection may be made to an external computer (for example, through the Internet using an Internet Service Provider).


The corresponding structures, materials, acts, and equivalents of all means or step plus function elements in the claims below are intended to include any structure, material, or act for performing the function in combination with other claimed elements as specifically claimed. The description of the present technology has been presented for purposes of illustration and description, but is not intended to be exhaustive or limited to the invention in the form disclosed. Many modifications and variations will be apparent to those of ordinary skill in the art without departing from the scope and spirit of the invention. Exemplary embodiments were chosen and described in order to best explain the principles of the present technology and its practical application, and to enable others of ordinary skill in the art to understand the invention for various embodiments with various modifications as are suited to the particular use contemplated.


Aspects of the present technology are described above with reference to flowchart illustrations and/or block diagrams of methods, apparatus (systems) and computer program products according to embodiments of the invention. It will be understood that each block of the flowchart illustrations and/or block diagrams, and combinations of blocks in the flowchart illustrations and/or block diagrams, can be implemented by computer program instructions. These computer program instructions may be provided to a processor of a general purpose computer, special purpose computer, or other programmable data processing apparatus to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable data processing apparatus, create means for implementing the functions/acts specified in the flowchart and/or block diagram block or blocks.


These computer program instructions may also be stored in a computer readable medium that can direct a computer, other programmable data processing apparatus, or other devices to function in a particular manner, such that the instructions stored in the computer readable medium produce an article of manufacture including instructions which implement the function/act specified in the flowchart and/or block diagram block or blocks.


Thus, methods and systems for filtering telephone calls have been disclosed. Although embodiments have been described with reference to specific example embodiments, it will be evident that various modifications and changes can be made to these example embodiments without departing from the broader spirit and scope of the present application. Accordingly, the specification and drawings are to be regarded in an illustrative rather than a restrictive sense.

Claims
  • 1. A method for generating a community blacklist comprising: determining, using a processor, a first telephone call originates from an originating telephone number, the first telephone call being directed to a first telephone number, the first telephone number being assigned to a first user;determining, using the processor, a second telephone call originates from the originating telephone number, the second telephone call being directed to a second telephone number, the second telephone number being assigned to a second user and being non-sequential from the first telephone number;determining, using the processor, a third telephone call originates from the originating telephone number, the third telephone call being directed to a third telephone number, the third telephone number being assigned to a third user and being non-sequential from the first and second telephone numbers;adding, using the processor, the originating telephone number to the community blacklist, when the originating telephone number is in at least one of a first user blacklist associated with the first user, a second user blacklist associated with the second user, and a third user blacklist associated with the third user, and the first telephone call, the second telephone call, and the third telephone call originate from the originating telephone number; andrejecting a fourth telephone call from the originating telephone number using the community blacklist.
  • 2. The method of claim 1 wherein determining the first telephone call, the second telephone call, and the third telephone call originate from the originating telephone number is based at least on one of a caller ID and a callback number.
  • 3. The method of claim 2 wherein the first telephone number, the second telephone number, and the third telephone number are each associated with a telephone service.
  • 4. The method of claim 3 further comprising: receiving, using the processor, the first, second, and third telephone calls through the Internet.
  • 5. The method of claim 4 further comprising: determining, using the processor, a plurality of telephone calls originates from the originating telephone number; andadding, using the processor, the originating telephone number to the community blacklist further based on the plurality of telephone calls being greater than or equal to a predetermined proportion of all telephone numbers associated with a service provider.
  • 6. The method of claim 5 further comprising: adding, using the processor, the originating telephone number to the community blacklist further based on at least one of time of day, day of week, and time of year.
  • 7. The method of claim 5 wherein the predetermined proportion is 0.0001.
  • 8. A system for generating a community blacklist comprising: a processor; anda memory coupled to the processor, the memory storing instructions executable by the processor to perform a method comprising: determining a first telephone call originates from an originating telephone number, the first telephone call being directed to a first telephone number, the first telephone number being assigned to a first user;determining a second telephone call originates from the originating telephone number, the second telephone call being directed to a second telephone number, the second telephone number being assigned to a second user and being non-sequential from the first telephone number;determining a third telephone call originates from the originating telephone number, the third telephone call being directed to a third telephone number, the third telephone number being assigned to a third user and being non-sequential from the first and second telephone numbers;adding the originating telephone number to the community blacklist, when the originating telephone number is in at least one of a first user blacklist associated with the first user, a second user blacklist associated with the second user, and a third user blacklist associated with the third user, and the first telephone call, the second telephone call, and the third telephone call originate from the originating telephone number; andrejecting a fourth telephone call from the originating telephone number using the community blacklist.
  • 9. The system of claim 8 wherein determining the first telephone call, the second telephone call, and the third telephone call originate from the originating telephone number is based at least on one of a caller ID and a callback number.
  • 10. The system of claim 9 wherein the first telephone number, the second telephone number, and the third telephone number are each associated with a telephone service.
  • 11. The system of claim 10 wherein the method further comprises: receiving, using the processor, the first, second, and third telephone calls through the Internet.
  • 12. The system of claim 11 wherein the method further comprises: determining, using the processor, a plurality of telephone calls originates from the originating telephone number; andadding, using the processor, the originating telephone number to the community blacklist further based on the plurality of telephone calls being greater than or equal to a predetermined proportion of all telephone numbers associated with a service provider.
  • 13. The system of claim 12 wherein the method further comprises: adding, using the processor, the originating telephone number to the community blacklist further based on at least one of time of day, day of week, and time of year.
  • 14. The system of claim 12 wherein the predetermined proportion is 0.0001.
  • 15. A system for generating a community blacklist comprising: means for determining a first telephone call originates from an originating telephone number, the first telephone call being directed to a first telephone number, the first telephone number being assigned to a first user;means for determining a second telephone call originates from the originating telephone number, the second telephone call being directed to a second telephone number, the second telephone number being assigned to a second user and being non-sequential from the first telephone number;means for determining a third telephone call originates from the originating telephone number, the third telephone call being directed to a third telephone number, the third telephone number being assigned to a third user and being non-sequential from the first and second telephone numbers;means for adding the originating telephone number to the community blacklist, when the originating telephone number is in at least one of a first user blacklist associated with the first user, a second user blacklist associated with the second user, and a third user blacklist associated with the third user, and the first telephone call, the second telephone call, and the third telephone call originate from the originating telephone number; andmeans for rejecting a fourth telephone call from the originating telephone number using the community blacklist.
CROSS-REFERENCE TO RELATED APPLICATIONS

This application is a continuation of U.S. patent application Ser. No. 14/318,630 filed on Jun. 28, 2014, which is a continuation of U.S. patent application Ser. No. 14/034,457 filed on Sep. 23, 2013, now U.S. Pat. No. 9,386,148 issued on Jul. 5, 2016. The disclosure of the above-reference patent application is incorporated herein by reference in its entirety for all purposes.

US Referenced Citations (250)
Number Name Date Kind
5425085 Weinberger et al. Jun 1995 A
5463595 Rodhall et al. Oct 1995 A
5519769 Weinberger et al. May 1996 A
5796736 Suzuki Aug 1998 A
5999611 Tatchell et al. Dec 1999 A
6023724 Bhatia et al. Feb 2000 A
6377938 Block et al. Apr 2002 B1
6487197 Elliott Nov 2002 B1
6615264 Stoltz et al. Sep 2003 B1
6661340 Saylor et al. Dec 2003 B1
6697358 Bernstein Feb 2004 B2
6714545 Hugenberg et al. Mar 2004 B1
6775267 Kung et al. Aug 2004 B1
6778517 Lou et al. Aug 2004 B1
6778528 Blair et al. Aug 2004 B1
6781983 Armistead Aug 2004 B1
6914900 Komatsu et al. Jul 2005 B1
6934258 Smith et al. Aug 2005 B1
7113090 Saylor et al. Sep 2006 B1
7124506 Yamanashi et al. Oct 2006 B2
7127043 Morris Oct 2006 B2
7127506 Schmidt et al. Oct 2006 B1
7154891 Callon Dec 2006 B1
7280495 Zweig et al. Oct 2007 B1
7295660 Higginbotham et al. Nov 2007 B1
7342925 Cherchali et al. Mar 2008 B2
7376124 Lee et al. May 2008 B2
7394803 Petit-Huguenin et al. Jul 2008 B1
8140392 Altberg et al. Mar 2012 B2
8331547 Smith et al. Dec 2012 B2
8350694 Trundle et al. Jan 2013 B1
8515021 Farrand et al. Aug 2013 B2
8634520 Morrison et al. Jan 2014 B1
8837698 Altberg et al. Sep 2014 B2
9225626 Capper et al. Dec 2015 B2
9386148 Farrand et al. Jul 2016 B2
9386414 Mayor et al. Jul 2016 B1
9560198 Farrand et al. Jan 2017 B2
9633547 Farrand et al. Apr 2017 B2
9667782 Farrand et al. May 2017 B2
9787611 Gillon et al. Oct 2017 B2
9905103 Hsieh Feb 2018 B2
9929981 Gillon et al. Mar 2018 B2
10009286 Gillon et al. Jun 2018 B2
20010053194 Johnson Dec 2001 A1
20020016718 Rothschild et al. Feb 2002 A1
20020035556 Shah et al. Mar 2002 A1
20020037750 Hussain et al. Mar 2002 A1
20020038167 Chirnomas Mar 2002 A1
20020085692 Katz Jul 2002 A1
20020130784 Suzuki et al. Sep 2002 A1
20020133614 Weerahandi et al. Sep 2002 A1
20020140549 Tseng Oct 2002 A1
20020165966 Widegren et al. Nov 2002 A1
20030027602 Han et al. Feb 2003 A1
20030058844 Sojka et al. Mar 2003 A1
20030099334 Contractor May 2003 A1
20030119492 Timmins et al. Jun 2003 A1
20030141093 Tirosh et al. Jul 2003 A1
20030164877 Murai Sep 2003 A1
20030184436 Seales et al. Oct 2003 A1
20030189928 Xiong Oct 2003 A1
20040001512 Challener et al. Jan 2004 A1
20040010472 Hilby et al. Jan 2004 A1
20040010569 Thomas et al. Jan 2004 A1
20040017803 Lim et al. Jan 2004 A1
20040059821 Tang et al. Mar 2004 A1
20040086093 Schranz May 2004 A1
20040090968 Kimber et al. May 2004 A1
20040105444 Korotin et al. Jun 2004 A1
20040160956 Hardy et al. Aug 2004 A1
20040235509 Burritt et al. Nov 2004 A1
20050027887 Zimler et al. Feb 2005 A1
20050036590 Pearson et al. Feb 2005 A1
20050074114 Fotta et al. Apr 2005 A1
20050078681 Sanuki et al. Apr 2005 A1
20050089018 Schessel Apr 2005 A1
20050097222 Jiang et al. May 2005 A1
20050105708 Kouchri et al. May 2005 A1
20050141485 Miyajima et al. Jun 2005 A1
20050169247 Chen Aug 2005 A1
20050222820 Chung Oct 2005 A1
20050238034 Gillespie et al. Oct 2005 A1
20050246174 DeGolia Nov 2005 A1
20050259637 Chu et al. Nov 2005 A1
20060007915 Frame Jan 2006 A1
20060009240 Katz Jan 2006 A1
20060013195 Son et al. Jan 2006 A1
20060071775 Otto et al. Apr 2006 A1
20060092011 Simon et al. May 2006 A1
20060114894 Cherchali et al. Jun 2006 A1
20060140352 Morris Jun 2006 A1
20060156251 Suhail et al. Jul 2006 A1
20060167746 Zucker Jul 2006 A1
20060187898 Chou et al. Aug 2006 A1
20060187900 Akbar Aug 2006 A1
20060243797 Apte et al. Nov 2006 A1
20060251048 Yoshino et al. Nov 2006 A1
20060258341 Miller et al. Nov 2006 A1
20060259767 Mansz et al. Nov 2006 A1
20060268828 Yarlagadda Nov 2006 A1
20060268848 Larsson et al. Nov 2006 A1
20070030161 Yang Feb 2007 A1
20070036314 Kloberdans et al. Feb 2007 A1
20070037560 Yun et al. Feb 2007 A1
20070037605 Logan Feb 2007 A1
20070041517 Clarke et al. Feb 2007 A1
20070049342 Mayer et al. Mar 2007 A1
20070054645 Pan Mar 2007 A1
20070061735 Hoffberg et al. Mar 2007 A1
20070067219 Altberg et al. Mar 2007 A1
20070071212 Quittek et al. Mar 2007 A1
20070118750 Owen et al. May 2007 A1
20070121593 Vance et al. May 2007 A1
20070121596 Kurapati et al. May 2007 A1
20070132844 Katz Jun 2007 A1
20070133757 Girouard et al. Jun 2007 A1
20070153776 Joseph et al. Jul 2007 A1
20070165811 Reumann et al. Jul 2007 A1
20070183407 Bennett et al. Aug 2007 A1
20070203999 Townsley et al. Aug 2007 A1
20070223455 Chang et al. Sep 2007 A1
20070238472 Wanless Oct 2007 A1
20070255702 Orme Nov 2007 A1
20070283430 Lai et al. Dec 2007 A1
20070298772 Owens et al. Dec 2007 A1
20080049748 Bugenhagen et al. Feb 2008 A1
20080075248 Kim Mar 2008 A1
20080075257 Nguyen et al. Mar 2008 A1
20080084975 Schwartz Apr 2008 A1
20080089325 Sung Apr 2008 A1
20080097819 Whitman Apr 2008 A1
20080111765 Kim May 2008 A1
20080118039 Elliot et al. May 2008 A1
20080125095 Mornhineway et al. May 2008 A1
20080144625 Wu et al. Jun 2008 A1
20080144884 Habibi Jun 2008 A1
20080159515 Rines Jul 2008 A1
20080168145 Wilson Jul 2008 A1
20080196099 Shastri Aug 2008 A1
20080200142 Abdel-Kader et al. Aug 2008 A1
20080205386 Pumadi et al. Aug 2008 A1
20080225749 Peng et al. Sep 2008 A1
20080247401 Bhal et al. Oct 2008 A1
20080298348 Frame et al. Dec 2008 A1
20080313297 Heron et al. Dec 2008 A1
20080316946 Capper et al. Dec 2008 A1
20090106318 Mantripragada et al. Apr 2009 A1
20090135008 Kirchmeier et al. May 2009 A1
20090168755 Peng et al. Jul 2009 A1
20090213999 Farrand et al. Aug 2009 A1
20090224931 Dietz et al. Sep 2009 A1
20090240586 Ramer et al. Sep 2009 A1
20090253428 Bhatia et al. Oct 2009 A1
20090295572 Grim, III et al. Dec 2009 A1
20090303042 Song et al. Dec 2009 A1
20090319271 Gross Dec 2009 A1
20100034121 Bozionek Feb 2010 A1
20100046530 Hautakorpi et al. Feb 2010 A1
20100046731 Gisby et al. Feb 2010 A1
20100098034 Tang et al. Apr 2010 A1
20100098058 Delangis Apr 2010 A1
20100098235 Cadiz et al. Apr 2010 A1
20100114896 Clark et al. May 2010 A1
20100136982 Zabawskyj et al. Jun 2010 A1
20100158223 Fang et al. Jun 2010 A1
20100191829 Cagenius Jul 2010 A1
20100229452 Suk Sep 2010 A1
20100277307 Horton et al. Nov 2010 A1
20100302025 Script Dec 2010 A1
20110054689 Nielsen et al. Mar 2011 A1
20110111728 Ferguson et al. May 2011 A1
20110140868 Hovang Jun 2011 A1
20110170680 Chislett et al. Jul 2011 A1
20110183652 Eng et al. Jul 2011 A1
20110265145 Prasad et al. Oct 2011 A1
20120010955 Ramer et al. Jan 2012 A1
20120027191 Baril et al. Feb 2012 A1
20120035993 Nangia Feb 2012 A1
20120036576 Iyer Feb 2012 A1
20120047442 Nicolaou et al. Feb 2012 A1
20120092158 Kumbhar et al. Apr 2012 A1
20120099716 Rae et al. Apr 2012 A1
20120284778 Chiou et al. Nov 2012 A1
20120320905 Ilagan Dec 2012 A1
20120329420 Zotti et al. Dec 2012 A1
20130018509 Korus Jan 2013 A1
20130035774 Warren et al. Feb 2013 A1
20130053005 Ramer et al. Feb 2013 A1
20130070928 Ellis et al. Mar 2013 A1
20130154822 Kumar et al. Jun 2013 A1
20130214925 Weiss Aug 2013 A1
20130267791 Halperin et al. Oct 2013 A1
20130272219 Singh et al. Oct 2013 A1
20130288639 Varsavsky Waisman-Diamond Oct 2013 A1
20130293368 Ottah et al. Nov 2013 A1
20130336174 Rubin et al. Dec 2013 A1
20140011470 D'Amato et al. Jan 2014 A1
20140022915 Caron et al. Jan 2014 A1
20140084165 Fadell et al. Mar 2014 A1
20140085093 Mittleman et al. Mar 2014 A1
20140101082 Matsuoka et al. Apr 2014 A1
20140120863 Ferguson et al. May 2014 A1
20140169274 Kweon et al. Jun 2014 A1
20140172953 Blanksteen Jun 2014 A1
20140222436 Binder et al. Aug 2014 A1
20140253326 Cho et al. Sep 2014 A1
20140266699 Poder et al. Sep 2014 A1
20140273912 Peh et al. Sep 2014 A1
20140273979 Van Os et al. Sep 2014 A1
20140306802 Hibbs, Jr. Oct 2014 A1
20140358666 Baghaie et al. Dec 2014 A1
20150065078 Mejia et al. Mar 2015 A1
20150071450 Boyden et al. Mar 2015 A1
20150086001 Farrand et al. Mar 2015 A1
20150087280 Farrand et al. Mar 2015 A1
20150100167 Sloo et al. Apr 2015 A1
20150117624 Rosenshine Apr 2015 A1
20150138333 DeVaul et al. May 2015 A1
20150145693 Toriumi et al. May 2015 A1
20150177114 Kapoor et al. Jun 2015 A1
20150229770 Shuman et al. Aug 2015 A1
20150244873 Boyden et al. Aug 2015 A1
20150262435 Delong et al. Sep 2015 A1
20150281450 Shapiro et al. Oct 2015 A1
20150302725 Sager et al. Oct 2015 A1
20150334227 Whitten et al. Nov 2015 A1
20150339912 Farrand et al. Nov 2015 A1
20150379562 Spievak et al. Dec 2015 A1
20160012702 Hart et al. Jan 2016 A1
20160036751 Ban Feb 2016 A1
20160078750 King et al. Mar 2016 A1
20160117684 Khor et al. Apr 2016 A1
20160142758 Karp et al. May 2016 A1
20160173693 Spievak et al. Jun 2016 A1
20160248847 Saxena et al. Aug 2016 A1
20160277573 Farrand et al. Sep 2016 A1
20160300260 Cigich et al. Oct 2016 A1
20160330108 Gillon et al. Nov 2016 A1
20160330319 Farrand et al. Nov 2016 A1
20160330770 Lee et al. Nov 2016 A1
20160373372 Gillon et al. Dec 2016 A1
20170021802 Mims Jan 2017 A1
20170034044 Gillon et al. Feb 2017 A1
20170034045 Gillon et al. Feb 2017 A1
20170034062 Gillon et al. Feb 2017 A1
20170034081 Gillon et al. Feb 2017 A1
20170084164 Farrand et al. Mar 2017 A1
20170104875 Im et al. Apr 2017 A1
20170270569 Altberg et al. Sep 2017 A1
Foreign Referenced Citations (9)
Number Date Country
3050287 Aug 2016 EP
3146516 Mar 2017 EP
3167340 May 2017 EP
3295620 Mar 2018 EP
WO2015041738 Mar 2015 WO
WO2015179120 Nov 2015 WO
WO2016007244 Jan 2016 WO
WO2016182796 Nov 2016 WO
WO2018044657 Mar 2018 WO
Non-Patent Literature Citations (16)
Entry
International Search Report and Written Opinion dated Nov. 7, 2014 for App. No. PCT/US2014/044945, filed Jun. 30, 2014.
International Search Report and Written Opinion dated Jul. 27, 2015 for App. No. PCT/US2015/029109, filed May 4, 2015.
International Search Report and Written Opinion dated Nov. 2, 2015 for App. No. PCT/US2015/034054, filed Jun. 3, 2015.
Life Alert. “Life Alert's Four Layers of Protection, First Layer of Protection: Protection at Home.” https://web.archive.org/web/20121127094247/http://www.lifealert.net/products/homeprotection.html. [retrieved Oct. 13, 2015].
International Search Report and Written Opinion dated Jun. 30, 2016 for App. No. PCT/US2016/030597, filed May 3, 2016.
European Patent Application No. 14845956.3, “Extended European Search Report,” dated Feb. 16, 2017, 8 pages.
Canadian Patent Application No. 2949211, “Office Action,” dated Aug. 16, 2017, 4 pages.
“Office Action,” Canadian Patent Application No. 2954351, dated Oct. 27, 2017, 3 pages.
“International Search Report” and “Written Opinion of the International Searching Authority,” Patent Cooperation Treaty Application No. PCT/US2017/048284, dated Nov. 8, 2017, 8 pages.
“Extended European Search Report,” European Patent Application No. 15796148.3, dated Jan. 8, 2018, 8 pages.
“Office Action,” European Patent Application No. 14845956.3, dated Apr. 9, 2018, 4 pages.
“Extended European Search Report,” European Patent Application No. 15818258.4, dated Feb. 26, 2018, 8 pages.
Osterlund, Karl et al., “Communications Network Failure Detection and Remediation,” U.S. Appl. No. 16/011,479, filed Jun. 18, 2018, Specification, Claims, Abstract, and Drawings, 92 pages.
“Notice of Allowance”, Canadian Patent Application No. 2949211, dated Jul. 31, 2018, 1 page.
“Office Action,” Canadian Patent Application No. 2954351, dated Aug. 22, 2018, 4 pages.
“Notice of Allowance,” European Patent Application No. 14845956.3, dated Jul. 11, 2018, 7 pages.
Related Publications (1)
Number Date Country
20160323446 A1 Nov 2016 US
Continuations (2)
Number Date Country
Parent 14318630 Jun 2014 US
Child 15208004 US
Parent 14034457 Sep 2013 US
Child 14318630 US