Method for identifying and prioritizing fault location in a cable plant

Information

  • Patent Grant
  • 10477199
  • Patent Number
    10,477,199
  • Date Filed
    Friday, March 15, 2013
    11 years ago
  • Date Issued
    Tuesday, November 12, 2019
    4 years ago
Abstract
A method of prioritizing estimated fault locations within a network includes monitoring multiple different performance parameters for unacceptable threshold levels via communications with a set of terminal network elements and separately analyzing the different performance parameters to identify potential network fault locations on the network. Accordingly, a plurality of priority rankings of potential network fault locations can be generated, one for each performance parameter monitored, and then combined to generate an overall priority ranking of potential fault locations including at least a highest priority inspection point estimated as being a most likely source of a fault on the network.
Description
BACKGROUND

Program providers such as multiple system operators, television networks and stations, cable TV operators, satellite TV operators, studios, wireless service providers, and Internet broadcasters/service providers, among others, require broadband communication systems to deliver programming and like content to consumers/subscribers over networks via digital or analog signals. Such networks and physical plants tend to be extensive and complex and therefore are difficult to manage and monitor for faults, impairments, maintenance issues and the like.


Monitoring network maintenance activities particularly presents problems to operators of extensive cable networks. For purposes of example, a cable network may include a headend which is connected to several nodes that may provide access to IP or ISPN networks. The cable network may also include a variety of cables such as coaxial cables, optical fiber cables, or a Hybrid Fiber/Coaxial (HFC) cable system which interconnect terminal network elements of subscribers to the headend in a tree and branch structure. The terminal network elements (media terminal adaptors (MTAs), cable modem, set top box, etc.) reside on the nodes which may be combined and serviced by common components at the headend.


Cable modems may support data connection to the Internet and other computer networks via the cable network. Thus, cable networks provide bi-directional communication systems in which data can be sent downstream from the headend to a subscriber and upstream from a subscriber to the headend. The headend typically interfaces with cable modems via a cable modem termination system (CMTS) which has several receivers. Each receiver of the CMTS may connect to numerous nodes which, in turn, may connect to numerous network elements, such as modems, media terminal adaptors (MTAs), set top boxes, terminal devices, customer premises equipment (CPE) or like devices of subscribers. A single receiver of the CMTS, for instance, may connect to several hundred or more network elements.


The conventional process for tracking which terminal devices are attached to which optical node and like information is a manual process. For instance, when a new customer's services are first enabled, a network operator may identify the specific node or location of the user and enter this information manually into a customer management database. This information can be valuable for resolving physical layer communications issues, performing periodic plant maintenance, and planning future service expansions. However, when the data is inaccurate or incomplete, it can lead to misdiagnosis of issues, excessive costs associated with maintenance, and prolonged new deployments. In addition, as communication traffic increases or new services are deployed, the need to understand loading of parts of the network becomes important, particularly if existing subscribers must be reallocated to different nodes or parts of the network.


Based on conventional practice, locating and identifying network and physical plant issues essentially relies upon the receipt of customer calls and manual technician analysis in response thereto.





BRIEF DESCRIPTION OF THE DRAWINGS

Various features of the embodiments described in the following detailed description can be more fully appreciated when considered with reference to the accompanying figures, wherein the same numbers refer to the same elements.



FIG. 1 is a snapshot screen view of a so-called dashboard of a graphical user interface according to an embodiment.



FIG. 2 is a view of a panel of the dashboard showing a cluster of objects displayed on top of a satellite image of a geographic area into which a network extends according to an embodiment.



FIG. 3 is a view of an interactive user interface display which may provide a starting point of the dashboard once a user logs into the system according to an embodiment.



FIG. 4 is a view similar to FIG. 3 with the map further zoomed-in to a particular region of the network service area according to an embodiment.



FIG. 5 is a view of an interactive user interface display which shows an alarm tree for use in investigating information of alarms shown on the display according to an embodiment.



FIG. 6 is a view similar to FIG. 5 with the alarm tree further expanded in accordance with an embodiment.



FIG. 7 is a view of a graphical user interface with a local geographic map showing a node location, terminal network elements, network path, and alarms in accordance with an embodiment.



FIG. 8 is a view of a graphical user interface similar to FIG. 7 with a cluster of terminal network elements highlighted based on geo-proximity in accordance with an embodiment.



FIG. 9 is a view of a graphical user interface similar to FIG. 8 that is displayed on a satellite image of the geographic area according to an embodiment.



FIG. 10 is a view of a graphical user interface similar to FIG. 9 and including a listing of alarms for the cable modems displayed on the map according to an embodiment.



FIG. 11 is a view of a graphical user interface similar to FIG. 10 and including a listing of a particular performance parameter (in this instance, downstream microreflections in dBs for absolute and delta values) for the cable modems displayed on the map and channels used thereby according to an embodiment.



FIG. 12 is a view of a wireless communication tablet having a display screen that may be used by a field technician in accordance with an embodiment.



FIG. 13 is a snapshot view of a display screen of the tablet providing a list of faulted modems in accordance with an embodiment.



FIG. 14 is a snapshot view of a display screen of the tablet providing the geographic locations of the faulted modems on a street map in accordance with an embodiment.



FIG. 15 is a view of a section of a network extending downstream from a fiber-optic node and in which cable modems affected by downstream signal-to-noise ratio at threshold levels are shown in accordance with an embodiment.



FIG. 16 is a view of the same section of the network as FIG. 15 in which cable modems affected by downstream power at threshold levels are shown in accordance with an embodiment.



FIG. 17 is a view of the same section of the network as FIG. 15 in which cable modems affected by upstream echo at threshold levels are shown in accordance with an embodiment.



FIG. 18 is a view of the same section of the network as FIG. 15 in which cable modems affected by downstream micro-reflection at threshold levels are shown in accordance with an embodiment.



FIG. 19 is a fault overlay view in which the estimated fault locations shown in FIGS. 15-18 are combined in a single topology image in accordance with an embodiment.



FIG. 20 is a topology image displaying a final prioritized inspection list for potential fault locations for the network shown in FIGS. 15-19 in accordance with an embodiment.



FIG. 21 is a flowchart of a method of estimating and prioritizing a location of a defect within a network in accordance with an embodiment.





DETAILED DESCRIPTION

For simplicity and illustrative purposes, the principles of embodiments are described by referring mainly to examples thereof. In the following description, numerous specific details are set forth in order to provide a thorough understanding of the embodiments. It will be apparent however, to one of ordinary skill in the art, that the embodiments may be practiced without limitation to these specific details. In some instances, well known methods and structures have not been described in detail so as not to unnecessarily obscure the embodiments.


Embodiments disclosed herein are directed to automated management and monitoring systems, tools, and methods that enable issues occurring in a network, such as a cable network, to be proactively and automatically detected and located. The embodiments leverage a combination of key data and network topology such as information concerning the geographical location of an issue, the nature of the issue, and/or the severity of an issue to permit a network operator to quickly detect, isolate, locate and address problems. In addition, collection and analysis of historical, long term and periodic health information of a network provided by the embodiments can aid in determining trends that may indicate slow and steady degradation of a network element or component. Such degradation has conventionally remained undetected when relying only on manual spot checks by field technicians and only becomes detectable upon component failure.


According to embodiments, the above referenced tasks are accomplished automatically by a management and monitoring tool that is able to scale across extremely large networks thereby enabling network operators to become more proactive with network maintenance activities and to achieve higher levels of network availability and reliability. Operational costs can be reduced by decreasing the need for troubleshooting at a time after the occurrence of the problem or issue. In addition, the periodic collection and analysis of network conditions provides a view into critical network indicators and aids in resolving issues prior to customer impact.


Network monitoring can be performed such that information concerning geographic location of monitored network elements, such as cable modems or the like, and associated network component topology, such as HFC components and the like, are automatically populated into a network management database or the like for purposes of providing a visual display, such as a geographically accurate street map or satellite image of a region of a service area, that clearly indicates a fault or other issue and the geographical location thereof. Examples are provided by FIGS. 15-20. Thus, the path that the network takes geographically is displayed on the map along with the physical location of network elements and components within the network. Such a map provides a useful network management tool to network operators and field technicians for resolving issues in an efficient and prompt manner.


As one contemplated example, the map can be provided as part of a graphical interface which displays faults of varying severity levels ranging from critical to completely non-service affecting. Accordingly, in at least some embodiments, the severity of a fault on the network can be determined and displayed with the estimated geographic location of the fault on the map.


In addition, the network monitoring and management system or tool can be provided and fully integrated into software that is loaded and resides on a server or remote server connected to or communicating with the network. Of course, the software may reside on other devices and equipment such as equipment located at the headend of the network, cloud devices, and portable or mobile devices. Utilization of the software eliminates the need for manual analysis of data and permits large amounts of data to be automatically analyzed electronically by microprocessors or the like on a large scale.


The network management tool or software may estimate and make assumptions regarding probable tap and passive locations, and couple this information with known optical node location data, and with walking directions data from a geographical data (geodata) services provider. Walking directions data may be in accordance with an appropriate format, language, or standard; examples include, but are not limited to, data in Keyhole Markup Language (KML), e.g., Open Geospatial Consortium (OGC) KML, or the OpenGIS KML Encoding Standard. From this cumulative information, the network management tool or software can estimate and automatically populate a map or the like of a given service area with monitored cable modem locations and associated network component topology. See FIGS. 15-20 for examples.


The geographic location of a fault and surrounding network path can be estimated, isolated, and displayed despite minimum information and manually entered data concerning the actual network path or network element location being available. The graphical interface can identify and display specific network elements as problematic. As an example, a network or HFC component such as cables, taps, passives, or the like that is identified as a suspect component potentially contributing to linear distortion, excessive loss impairments, or the like may be identified and displayed as a location of a fault. Whether a fault impacts a single subscriber or a group of subscribers may also be estimated and shown in the display.


Still further, the network management tool may be used to identify clusters or groups of network elements or cable modems that may share network or HFC infrastructure, such as common components including optics, nodes, amps, cables, taps, passives, and the like. In this regard, Management Information Base (MIB) information for service groups readily available via data pulls from a CMTS or like equipment at the headend of the network can be used in conjunction with the above referenced geographical location information. Network element groups or clusters can be readily displayed via the graphical interface and without the need for the software to reference other sources, perform testing, or wait for common impairment signature alarms to be raised.


Still further, the severity of a fault may be estimated with respect to upstream impairments through association of physical layer metrics including pre and post forward error correction (FEC) along with the number of impacted network elements or subscribers. Higher priority alarms can be assigned to groups of network elements or subscribers that exceed threshold values. In contrast, lower priority alarms can be assigned to faults such as detected for single network elements or subscribers.


According to an embodiment, the graphical interface referenced above may be presented in the form of a so-called “dashboard” to a user such as personnel of a network operations center. Critical alarms may be shown across the entire network in a geographical display of the network or parts thereof. In addition, access may be provided to statistics via use of the dashboard to allow the user to monitor the overall health of their network.


By way of example, various snap-shot views of a graphical user interface are provided in FIGS. 1-14. It should be understood that these displays are disclosed for purposes of example only and may be altered as desired.


A first example of a dashboard 10 which may be displayed to a user via a monitor or like electronic display screen is shown in FIG. 1. In this example, a first panel 12 of the dashboard 10 provides information of “Active Alarms” including a list of alarms or potential faults 14, a second panel 16 provides a so-called “Physical View” of the network, and a third panel 18 provides a geographically-accurate street map 20 showing the geographical location of the alarms listed in panel 12 along with the nearest node 22 or other network component. The map 20 may include roads and streets and names thereof. In addition, as best illustrated in FIG. 2, alarms can be overlaid on images 24, for instance satellite images, of the geographical service area in which the alarms are located.


When an issue, fault or alarm is identified, it can be associated and displayed with other issues, faults and alarms based on geographical proximity. For instance, see the alarms 14 within circle 26 in FIG. 1. This group or cluster of alarms provides a visual indicator of the network elements affected and can indicated a center point of a potential problem causing the cluster of alarms. For instance, see the center point 28 in FIG. 2. A user which selects the center point may be provided with a listing of problem network elements or modems. In addition, the cluster of alarms may have a single corresponding “alarm” object to thereby reduce the number of alarms displayed to the user.


After an issue is first identified by the network monitoring and management system, tool or software, the operator or user may be provided with several options to further investigate the apparent problem or problems. For instance, network issues may be isolated by “serving group” or “geographic proximity” (i.e., clustering) and may be prioritized by severity based on the number of customers/subscribers affected and the extent to which faults are service-affecting. The network faults can be linked by the management software to a map interface which enables the fault to be connected to a physical location in the network.



FIGS. 3-11 provide further examples of views of a dashboard which may be displayed to a network operator. Any type or number of available charts, maps, or alert views can be viewed and organized in the dashboard. By way of example, the dashboard 30 shown in FIG. 3 may be configured as a starting point when a user first logs onto the network monitoring and management software or system. Here, a “zoomed-out” view of the network is initially provided to permit an overall view of the network, which may span a large geographic area. Data is collected and analyzed by the network monitoring and management tool to identify a type of fault or faults and the estimated geographic location of the fault(s) solely based on analysis of the data.



FIG. 3 provides an entire network view 32 based on a geographic format and provides an indication of so-called “hot-spots” 34 of alarms. A listing 36 of alarms can be provided in a panel 38 which can also indicate the severity and location of the hot-spots 34. Charts such as a FEC deltas/CMTS channel exceeding threshold chart 40, a Flap deltas/CMTS channel exceeding threshold chart 42, and a CMTS channel utilization threshold crossing chart 44 can be displayed in a panel 46 and correspond to the alarms shown in the listing 36. Of course, these charts provide just a few examples of possible charts. A further example of such a dashboard is shown in FIG. 4 which provides a display of a section of the map 48 in greater detail.


In FIG. 5, a dashboard is shown in which panel 50 provides information on network topology. Here, the topology is provided in a form of a so-called alarm tree which enables a user to gain further information with respect to more narrowly defined sections of the network. For example, the topology could list CMTSs (such as CMTS-1, CMTS-2, CMTS-3, CMTS-4, and CMTS-5). Further, the fiber nodes (i.e., FN-A and FN-B) can be shown for any of the CMTSs and a number of network elements associated with an alarm can be listed. As shown in FIG. 6, the panel 50 can also be expanded to show the number of network elements associated with alarms per severity of alarm (i.e., critical, major, and minor).


A more local view of a street map 52 is shown in FIG. 7. Here a single fiber node 54 of the network is shown as is the network path 56 extending from the node 54 to terminal network elements 58, such as cable modems, serviced via the node 54. The shade (or color, etc.) of the terminal networks elements 58 can be used to visually indicate an alarm on the map 52. For instance, terminal network element 58a is shown in a dark shade (or a particularly color, such as red) which may indicate an alarm of critical severity whereas terminal network elements displayed in lighter shades (other colors, such as yellow) may indicate an alarm of a minor severity. This same map 52 can be further investigated as shown in FIG. 8 in which a geo-proximity cluster 60 is shown highlighted. The path 56 of the cable plant may be estimated and shown such as in FIGS. 7 and 8. If desired, the user of the management tool is able to adjust the path 56 or enter in any known network topology information into the management software or tool should the estimated path and view be inaccurate.


Another view similar to FIG. 7 is shown in the map 62 of FIG. 9. Here the street map 52 has been modified to show actual satellite imagery of the surrounding geographic area. The node 54, path 56, and terminal network elements 58 are overlaid on the satellite imagery as are the alarms and other network topology. For purposes of further investigating a potential network fault, the “cable modems” illustrated in FIG. 9 can be shown in a drop down window 64 such as shown in FIG. 10. Here the MAC address, power status, noise status, upstream reflection status, downstream reflection status, FEC status for each cable modem or terminal network element 58. Some of these cable modems and listed statuses have no alarms whereas others have alarms of “minor” severity while others have alarms of “critical” severity. FIG. 11 shows the ability of the tool to further investigate network issues. Here, measurements corresponding to downstream microreflections in dBs are listed (as absolute and delta values) and shown in a window 66 so that a user may view these or any other values that are or are not the subject of an alarm.


Accordingly, after a network operator center user views the above referenced dashboards and investigates alarms therewith, for instance as shown above, and has identified a particular issue that needs to be resolved, the network monitoring and management tool, software or system can be used to assist the user in sending an appropriate field technician to the correct geographical location. The user can also use the management tool or software to assess the urgency with respect to the need to resolve the issue.


The network monitoring and management system, tool or software can also be used by a service technician in the field. For example, the network monitoring and management software may be run on a remote server that is accessible by the technician such as via a secure wireless web interface. For instance, a mobile device, such as a portable, lap-top, notebook, or tablet computer, a smart phone, or the like may be used to obtain various views, information and maps as discussed above. Accordingly, provided information can be used for rapid, real-time debugging of field issues and provide geographic information, provide real-time monitoring of upstream and downstream performance metrics and error states, and permit a technician to see the interdependency of multiple issues. The above can reduce the need for the technician to access the inside of residences, reduce the number of calls the technician needs to make to the head-end, and enable the technician to update network topology information while in the field. For purposes of this disclosure, “real-time” includes a level of responsiveness that is sufficiently fast to provide meaningful data that reflects current or recent network conditions as well as a level of responsiveness that tolerates a degree of lateness or built-in delay.


A tablet 70 is shown in FIGS. 12-14 that may be used by a field technician to connect to the network monitoring and management software. In FIG. 12, the technician is provided with a display 72 that includes an icon 74 for a list of the CMTSs, an icon 76 for network wide alerts, an icon 78 for scanning or uploading information into the system, and a settings icon 80. FIG. 13 shows a display 82 providing a tabular view of network devices 84 having faults, and FIG. 14 shows a display 86 showing the same network devices 84 in a geographical map-style platform with the closest fiber node 88 or like network component. All of the above provides helpful and useful information to the field technician.


Various methods can be used by the network monitoring and management system, software, and tool described above that enables fault determination, fault location, mapping of the network geographically, displaying of faults with and without network topology information, displaying a cluster of network elements impacted by the same fault, and the severity of the fault. For example, a combination of monitored parameters and network topology information can be used to identify the likely physical locations of cable network defects. This approach is able to be implemented in software utilizing numerical analysis. In addition, a combination of sub-algorithms can be used to locate a common network failure point even when several different and potentially, seemingly unrelated, issues are observed.


Often, a single defect within a plant can cause multiple types of impairments to be recognized which may otherwise appear to be independent and arise from separate issues. Each of these impairments may trigger multiple, independent fault detection mechanisms within the network monitoring tool. However, not all of the fault detection algorithms may identify the same network element as a primary fault location (i.e. a location estimated to be the most likely point or source of the fault). However, the results of all of these independent fault detection/identification algorithms can be viewed together in an effort to significantly improve the accuracy of identifying a root cause of an issue in the presence of multiple fault signatures.


According to an embodiment, the above referenced network monitoring tool can be configured to automatically evaluate groups of separate issues affecting a set of cable modems sharing common network components to quickly and accurately identify a root cause of a particular issue. Thus, a grouping of otherwise seemingly unrelated alarms is analyzed, and a determination is made with respect to whether or not the issues might actually be related. If a relation is found, the relation is shown on a map or provided in an alternate form (such as within a listing) so that the root cause can be quickly located and addressed. Each alarm within the grouping is assessed and analyzed independently, and then the results are evaluated as a set to accurately locate the issue within the plant. In this manner, a plurality of alarm topologies is considered, and then a single accurate inspection list is generated for the root cause issue.


By way of example, the following algorithm may be used to prioritize fault location based upon the occurrence of multiple alarms. First, all active threshold alarms that may be associated with a particular fiber node being evaluated are automatically retrieved. Each alarm associated with the fiber node is evaluated independently with respect to estimated fault location on the network. Two or more of the alarms are considered to be part of the same issue if they share any fault topology points in common. Thereafter, an inspection list is generated and prioritized based on a priority ranking for each particular type of alarm (i.e., highest priority alarm, second highest, third highest, etc. . . ). Here, the estimation as to fault location may be different depending upon the type of alarm and algorithm for such an alarm that is used.


The inspection points for each alarm are scored based upon their priority ranking. Simply for purposes of example, the highest priority alarm associated with each alarm may receive a score of ten (10) points and a second highest priority alarm for each alarm may receive a score of nine (9) points. Thereafter, all of the scores for a given inspection point (across all of the alarm types) are then added up and this sum total is assigned to the inspection point. The inspection point with the highest point totals is then given the highest priority as the root cause issue of all the alarms, the inspection point with the second highest point total is given the second highest priority as the root cause issued of all alarms, and so forth. Here, a point system is disclosed by way of example and the disclosed point system could be replaced by any type of point value system and/or ranking system involving letter grades or the like.


For purposes of providing an example with respect to the above described algorithm, four different alarm topologies 100, 102, 104 and 106 associated with a single plant defect and the same part of a network is shown in FIGS. 15-18. In each of these topologies 100, 102, 104 and 106, the physical location of a network fault may be estimated by receiving different types of information via data pulls including information concerning network components and geographic locations of the network components and terminal network elements and geographic locations of the terminal network elements. The existence of a network fault within the network can be automatically and electronically detected by monitoring various performance parameters, for instance, obtained via upstream communications from terminal network elements on the network.


A physical location of the network fault on the network may be estimated based on the particular performance parameter detected, the information of the physical topology of the network obtained, and the terminal network element or elements from which the performance parameter was received that indicated the network fault. Thereafter, a list of network components that may require inspection and may provide a source of the network fault can be automatically generated based on analysis of the performance parameter. By way of example, the listed network components may include drop cables, taps, trunk cables, splitters, amplifiers, nodes, and like components and the types of performance parameters may include downstream or upstream signal-to-noise ratio (SNR), absolute and delta downstream power (DS Power) level, absolute and delta upstream power (US Power) level, upstream echo (US Echo) level, downstream micro-reflection (DS Micro) level, upstream filter coefficient ratio, carrier-to-noise ratio (CNR), and modulation error ratio (MER).


For purposes of example, the network shown in FIGS. 15-20 may be a hybrid fiber-coaxial (HFC) network which interconnects terminal network elements, such as cable modems, to a headend (not shown) of the network having a cable modem termination system (CMTS) (not shown) via a tree and branch network structure. The upstream communications are herein defined as communications transmitted in a direction from the terminal network elements toward the headend.


A geographically-accurate map may be automatically and electronically populated with the geographic locations of network components to which a network fault is attributed, a geographic location of each the terminal network elements impacted by the network fault, and a diagnostic alarm identifying the network fault. The map may be displayable, for instance, with the use of geospatial software.


Different algorithms are used in each of the topologies 100, 102, 104 and 106 of FIGS. 15-18 to estimate the physical location of a fault. For this purpose, data is automatically gathered in real time and/or with an acceptable amount of delay by the CMTS, a server, or other equipment from cable modems in use by subscribers in the network to locate issues within the cable plant. As shown in FIGS. 15-18, a plurality of cable modems 110 are shown connected in tree and branch architecture via a node 112 which connects to the headend (not shown) of the network. The tree and branch architecture defines the path the network follows to each cable modem 110 and common network components on the network that may be shared by different subsets of cable modems.


In FIG. 15, a plurality of the modems 114 is identified in a part of the network as having unacceptable downstream signal-to-noise-ratio (DS SNR) levels. Based on an analysis of the DS SNR data, it is determined that taps and splitter identified by the circle 116 represents the most likely source of the DS SNR issue. Thus, for example, this location may be assigned a value of ten (10) points. Likewise, the taps and splitter identified by the circle 118 may be determined to be the second most likely source of the DS SNR issue. Thus, for example, this location may be assigned a value of nine (9) points. The third most likely source of the issue shown in FIG. 15 is the cable 120 extending between the two above referenced splitters.


In FIG. 16, a number of the modems 124 in the same part of the network is identified as having unacceptable downstream power (DS Power) levels. Based on an analysis of the DS Power data, it is determined that taps and splitter identified by the circle 126 represents the most likely source of the DS Power issue. Thus, for example, this location may be assigned a value of ten (10) points. Likewise, the taps and splitter identified by the circle 128 may be determined to be the second most likely source of the DS Power issue. Thus, for example, this location may be assigned a value of nine (9) points. The third most likely source of the issue shown in FIG. 16 is the splitter and taps referenced by circle 130.


In FIG. 17, a number of the modems 134 in the same part of the network is identified as having unacceptable upstream echo (US Echo) levels. Based on an analysis of the US Echo data, it is determined that taps and splitter identified by the circle 136 represents the most likely source of the US Echo issue. Thus, for example, this location may be assigned a value of ten (10) points Likewise, the taps and splitter identified by the circle 138 may be determined to be the second most likely source of the US Echo issue. Thus, for example, this location may be assigned a value of nine (9) points. The third most likely source of the issue shown in FIG. 17 is the cable 140 extending between the above referenced splitters.


Finally, in FIG. 18, a number of the modems 144 in the same part of the network is identified as having unacceptable downstream microreflection (DS Micro) levels. Based on an analysis of the DS Micro data, it is determined that taps and splitter identified by the circle 146 represents the most likely source of the DS Micro issue. Thus, for example, this location may be assigned a value of ten (10) points Likewise, the taps and splitter identified by the circle 148 may be determined to be the second most likely source of the DS Micro issue. Thus, for example, this location may be assigned a value of nine (9) points. The third most likely source of the issue shown in FIG. 18 is the cable 150 extending between the above referenced splitters.


As shown in the example of above, the four topologies 100, 102, 104 and 106 identify alarms based on different cable modem performance parameters (i.e., for DS SNR, DS Power, US Echo, and DS Micro) that may not be separate issues and that be associated with a single plant defect or root cause. As shown in FIGS. 15-18, the analysis of each of the different parameter (regardless of algorithm utilized) produces different priority lists with respect to the location that represents the first, second and third most likely locations of the fault. As also shown in FIGS. 15-18, points can be assigned to each of the top two highest priority inspection points in each topology (in this example, ten points for the highest priority inspection point and nine points for the second highest inspection point).


In FIG. 19, each of the alarm topology scores discussed above with respect to topologies 100, 102, 104 and 106 and FIGS. 15-18 are totaled at each corresponding inspection point to create a topology map 152 in which an overall priority ranking based on a combination of the analysis performed independently for each of the DS SNR, DS Power, US Echo and DS Micro cable modem performance parameters. This essentially combines these four separate fault location determinations to provide an overall presentation of the problem. As shown in FIG. 19, inspection point 154 received thirty points, inspection point 156 received nineteen points, and each of inspection points 158, 160 and 162 received nine points.



FIG. 20 provides a final topology map 164 identifying a final inspection list. According to the above described algorithm, inspection point 154 is identified as a location of the most likely source of the issue. In addition, inspection point 156 is shown as being the second most likely location of the issue, and inspection point 158 is shown as being the third most likely source of the fault or defect. Here, although inspection points 158, 160 and 162 each received the same amount of points, inspection point 158 is given priority over inspection points 160 and 162 since inspection point 158 is located further upstream and closest to the node 112.



FIG. 21 provides a flowchart for the above referenced algorithm using fault signature overlay to prioritize network locations/components to be inspected in view of issues detected based on different performance parameters for cable modems served by a common node. In step 170, a first performance parameter is monitored for alarms with respect to unacceptable threshold levels via communications with a set of terminal network elements, such as cable modems, on the network. The performance parameter reported by the alarm is analyzed in step 172 to identify potential network fault locations on the network. In step 174, a priority ranking of potential network fault locations including a most likely estimated fault location is generated, and scores of pre-determined values are assigned in step 176 to at least the most likely estimated fault location and a second most likely estimated fault location within the priority ranking for the performance parameter.


Thereafter, in step 178, the above referenced steps are repeated a number of times for different types of performance parameters. As examples of performance parameters, any of the following may be monitored and analyzed: downstream or upstream signal-to-noise ratio (SNR); absolute and delta downstream power (DS Power) level; absolute and delta upstream power (US Power) level; upstream echo (US Echo) level; downstream micro-reflection (DS Micro) level; upstream filter coefficient ratio; carrier-to-noise ratio (CNR); and modulation error ratio (MER).


After all desired performance parameters have been monitored, analyzed, and used to generate and apply scores to priority rankings, the plurality of priority rankings can be combined. For example, the scores attributed to each corresponding potential network fault location given by the plurality of priority rankings can be totaled. See step 180. From this information, an overall priority ranking can be generated which includes at least a highest priority inspection point estimated as being a most likely source of a fault on the network. See step 182. Thereafter, if desired, a list of network components that require inspection and that includes the highest priority inspection point estimated as being the most likely source of the fault on the network can be generated. See step 184. As a further option, a geographically-accurate map can be populated with a geographic location of a network component determined to be the most likely source of the fault, a geographic location of each the terminal network elements impacted by the network fault, and a diagnostic alarm identifying the network fault (see step 186) and displayed.


A signal processing electronic device, such as a server, remote server, CMTS or the like can run a software application to provide the above process steps and analysis. In addition, a non-transitory computer readable storage medium having computer program instructions stored thereon that, when executed by a processor, cause the processor to perform the above discussed operations can also be provided.


The above referenced signal processing electronic devices for carrying out the above methods can physically be provided on a circuit board or within another electronic device and can include various processors, microprocessors, controllers, chips, disk drives, and the like. It will be apparent to one of ordinary skill in the art the modules, processors, controllers, units, and the like may be implemented as electronic components, software, hardware or a combination of hardware and software.


While the principles of the invention have been described above in connection with specific networks, devices, apparatus, systems, and methods, it is to be clearly understood that this description is made only by way of example and not as limitation on the scope of the invention as defined in the appended claims.

Claims
  • 1. A method of prioritizing estimated fault locations within a network, comprising the steps of: in at least one processor communicatively coupled to the network, monitoring a plurality of different performance parameters for unacceptable threshold levels via communications with a set of terminal network elements on the network;in the at least one processor, separately analyzing each one of the plurality of different performance parameters identifying unacceptable threshold levels to identify potential network fault locations on the network;in the at least one processor, generating a plurality of priority rankings of potential network fault locations, each of the plurality of priority rankings associated respectively with a most likely estimated fault location, and each of the plurality of priority rankings being determined from a different one of the plurality of different performance parameters identifying unacceptable threshold levels; andin the at least one processor, combining information of each of the plurality of priority rankings to generate an overall priority ranking of potential network fault locations, wherein the overall priority ranking is associated with at least a highest priority inspection point, and wherein generating the overall priority ranking further includes estimating that one of the potential network fault locations is a most likely source of a fault on the network, and identifying said most likely source of a fault on the network as being the highest priority inspection point.
  • 2. A method according to claim 1, further comprising the step of, for each of the plurality of priority rankings, assigning scores of pre-determined values to at least the most likely estimated fault location and a second most likely estimated fault location.
  • 3. A method according to claim 2, wherein said combining step includes a step of totaling the scores attributed to each corresponding potential network fault location across the plurality of priority rankings to generate the overall priority ranking.
  • 4. A method according to claim 3, wherein, for different potential network fault locations having an equal score as determined during said totaling step, a potential network fault location that is most upstream on a network path relative to a node component is assigned a higher ranking than other potential network fault locations having the equal score.
  • 5. A method according to claim 1, further comprising a step of automatically generating a list of network components that require inspection and that includes at least the highest priority inspection point.
  • 6. A method according to claim 1, wherein the set of terminal network elements includes terminal network elements sharing at least one of the potential network fault locations in common.
  • 7. A method according to claim 4, wherein the set of terminal network elements are connected to a headend of the network via a common node of the network via paths arranged in a tree and branch network structure.
  • 8. A method according to claim 1, further comprising the step of receiving information electronically of a physical topology of the network.
  • 9. A method according to claim 8, wherein said receiving step including data pulls of information concerning network components and geographic locations of the network components and terminal network elements and geographic locations of the terminal network elements.
  • 10. A method according to claim 9, further comprising the step of automatically and electronically populating a geographically-accurate map with a geographic location of a network component determined to be the most likely source of the fault, a geographic location of each the terminal network elements impacted by the network fault, and a diagnostic alarm identifying the network fault.
  • 11. A method according to claim 10, further comprising the step of displaying the map with geospatial software.
  • 12. A method according to claim 9, wherein the network components are selected from a group consisting of drop cables, taps, trunk cables, amplifiers, splitters, and node components.
  • 13. The method according to claim 1, wherein the network is a cable network interconnecting the terminal network elements which include cable modems to a headend of the network having a cable modem termination system (CMTS), and wherein the communications include upstream communications in a direction from the terminal network elements to the headend.
  • 14. The method according to claim 1, wherein a performance parameter of the plurality of different performance parameters monitored comprises one of: a downstream signal-to-noise ratio, an upstream signal-to-noise ratio, an absolute and delta downstream power level, an absolute and delta upstream power level, an upstream echo level, a downstream micro-reflection level, an upstream filter coefficient ratio, a carrier-to-noise ratio, and a modulation error ratio.
  • 15. A signal processing electronic device for prioritizing estimated fault locations within a network, comprising at least one processing unit configured to: monitor a plurality of different performance parameters for unacceptable threshold levels via communications with a set of terminal network elements on the network;separately analyze each one of the plurality of different performance parameters identifying unacceptable threshold levels to identify potential network fault locations on the network;generate a plurality of priority rankings of potential network fault locations, each of the plurality of priority rankings associated respectively with a most likely estimated fault location, and each of the plurality of priority rankings being determined from a different one of the plurality of different performance parameters identifying unacceptable threshold levels; andcombine information of each of the plurality of priority rankings to generate an overall priority ranking of potential network fault locations, wherein the overall priority ranking is associated with at least a highest priority inspection point, and wherein to generate the overall priority ranking further includes to estimate that one of the potential network fault locations is a most likely source of a fault on the network, and to identify said most likely source of a fault on the network as being the highest priority inspection point.
  • 16. A signal processing electronic device according to claim 15, the at least one processing unit being further configured to: assign scores of pre-determined values to at least the most likely estimated fault location and a second most likely estimated fault location for each of the plurality of priority rankings and being configured to total the scores attributed to each corresponding potential network fault location across the plurality of priority rankings to generate the overall priority ranking.
  • 17. A signal processing electronic device according to claim 15, the at least one processing unit being further configured to: automatically generate a list of network components that require inspection and that includes the at least highest priority inspection point estimated as being the most likely source of the fault on the network.
  • 18. A signal processing electronic device according to claim 15, the at least one processing unit being further configured to: populate a geographically-accurate map with a geographic location of a network component determined to be the most likely source of the fault, a geographic location of each the terminal network elements impacted by the network fault, and a diagnostic alarm identifying the network fault.
  • 19. A signal processing electronic device according to claim 15, the at least one processing unit being further configured to: monitor performance parameters including at least one of a downstream signal-to-noise ratio, an upstream signal-to-noise ratio, an absolute and delta downstream power level, an absolute and delta upstream power level, an upstream echo level, a downstream micro-reflection level, an upstream filter coefficient ratio, a carrier-to-noise ratio, and a modulation error ratio.
  • 20. At least one non-transitory computer readable storage medium having computer program instructions stored thereon that, when executed by at least one processor, cause the at least one processor to perform the following operations: monitor a plurality of different performance parameters for unacceptable threshold levels via communications with a set of terminal network elements on a network;separately analyze each one of the plurality of different performance parameters identifying unacceptable threshold levels to identify potential network fault locations on the network;generate a plurality of priority rankings of potential network fault locations, each of the plurality of priority rankings associated respectively with a most likely estimated fault location, and each of the plurality of priority rankings being determined from a different one of the plurality of different performance parameters identifying unacceptable threshold levels; andcombine information of each of the plurality of priority rankings to generate an overall priority ranking of potential network fault locations, wherein the overall priority ranking is associated with at least a highest priority inspection point, and wherein to generate the overall priority ranking further includes to estimate that one of the potential network fault locations is a most likely source of a fault on the network, and to identify said most likely source of a fault on the network as being the highest priority inspection point.
US Referenced Citations (346)
Number Name Date Kind
3838221 Schmidt et al. Sep 1974 A
4245342 Entenman et al. Jan 1981 A
4385392 Angell et al. May 1983 A
4811360 Potter Mar 1989 A
4999787 McNally et al. Mar 1991 A
5197064 Chao et al. Mar 1993 A
5228060 Uchiyama Jul 1993 A
5251324 McMullan et al. Oct 1993 A
5271060 Moran et al. Dec 1993 A
5278977 Spencer et al. Jan 1994 A
5347539 Sridhar et al. Sep 1994 A
5390339 Bruckert et al. Feb 1995 A
5463661 Moran et al. Oct 1995 A
5532865 Utsumi et al. Jul 1996 A
5557603 Barlett et al. Sep 1996 A
5606725 Hart Feb 1997 A
5631846 Szurkowski May 1997 A
5692010 Nielsen Nov 1997 A
5694437 Yang et al. Dec 1997 A
5732104 Brown et al. Mar 1998 A
5757526 Shiragaki et al. May 1998 A
5771274 Harris Jun 1998 A
5790523 Ritchie et al. Aug 1998 A
5862451 Grau et al. Jan 1999 A
5867539 Koslov Feb 1999 A
5870429 Moran et al. Feb 1999 A
5886749 Williams et al. Mar 1999 A
5939887 Schmidt et al. Aug 1999 A
5943604 Chen et al. Aug 1999 A
6032019 Chen et al. Feb 2000 A
6061393 Tsui et al. May 2000 A
6108351 Hardy et al. Aug 2000 A
6154503 Strolle et al. Nov 2000 A
6229792 Anderson et al. May 2001 B1
6230326 Unger et al. May 2001 B1
6233274 Tsui et al. May 2001 B1
6240553 Son et al. May 2001 B1
6272150 Hrastar et al. Aug 2001 B1
6278730 Tsui et al. Aug 2001 B1
6308286 Richmond et al. Oct 2001 B1
6310909 Jones Oct 2001 B1
6321384 Eldering Nov 2001 B1
6330221 Gomez Dec 2001 B1
6334219 Hill et al. Dec 2001 B1
6377552 Moran, III et al. Apr 2002 B1
6385773 Schwartzman et al. May 2002 B1
6389068 Smith et al. May 2002 B1
6434583 Dapper et al. Aug 2002 B1
6445734 Chen et al. Sep 2002 B1
6456597 Bare Sep 2002 B1
6459703 Grimwood et al. Oct 2002 B1
6477197 Unger Nov 2002 B1
6477526 Hayashi et al. Nov 2002 B2
6480469 Moore et al. Nov 2002 B1
6483033 Simoes et al. Nov 2002 B1
6498663 Farhan et al. Dec 2002 B1
6512616 Nishihara Jan 2003 B1
6526260 Hick et al. Feb 2003 B1
6546557 Ovadia Apr 2003 B1
6556239 Al-Araji et al. Apr 2003 B1
6556562 Bhagavath et al. Apr 2003 B1
6556660 Li et al. Apr 2003 B1
6559756 Al-Araji et al. May 2003 B2
6563868 Zhang et al. May 2003 B1
6570394 Williams May 2003 B1
6570913 Chen May 2003 B1
6574797 Naegeli et al. Jun 2003 B1
6588016 Chen et al. Jul 2003 B1
6606351 Dapper et al. Aug 2003 B1
6611795 Cooper Aug 2003 B2
6646677 Noro et al. Nov 2003 B2
6662135 Burns et al. Dec 2003 B1
6662368 Cloonan et al. Dec 2003 B1
6671334 Kuntz et al. Dec 2003 B1
6687632 Rittman Feb 2004 B1
6690655 Miner et al. Feb 2004 B1
6700875 Schroeder et al. Mar 2004 B1
6700927 Esliger et al. Mar 2004 B1
6711134 Wichelman et al. Mar 2004 B1
6741947 Wichelman et al. May 2004 B1
6748551 Furudate et al. Jun 2004 B2
6757253 Cooper et al. Jun 2004 B1
6772388 Cooper et al. Aug 2004 B2
6772437 Cooper et al. Aug 2004 B1
6775840 Naegel et al. Aug 2004 B1
6785473 Sasaki et al. Aug 2004 B1
6816463 Cooper et al. Nov 2004 B2
6834057 Rabenko et al. Dec 2004 B1
6839829 Daruwalla et al. Jan 2005 B1
6853932 Wichelman et al. Feb 2005 B1
6877166 Roeck et al. Apr 2005 B1
6895043 Naegeli et al. May 2005 B1
6895594 Simoes et al. May 2005 B1
6906526 Hart, Jr. et al. Jun 2005 B2
6928475 Schenkel et al. Aug 2005 B2
6944881 Vogel Sep 2005 B1
6961314 Quigley et al. Nov 2005 B1
6961370 Chappell Nov 2005 B2
6967994 Boer et al. Nov 2005 B2
6973141 Isaksen et al. Dec 2005 B1
6985437 Vogel Jan 2006 B1
6987754 Shahar et al. Jan 2006 B2
6999408 Gomez Feb 2006 B1
7002899 Azenkot et al. Feb 2006 B2
7010002 Chow et al. Mar 2006 B2
7017176 Lee et al. Mar 2006 B1
7032159 Lusky et al. Apr 2006 B2
7039939 Millet et al. May 2006 B1
7050419 Azenkot et al. May 2006 B2
7054554 McNamara et al. May 2006 B1
7058007 Daruwalla et al. Jun 2006 B1
7072365 Ansley Jul 2006 B1
7079457 Wakabayashi et al. Jul 2006 B2
7099412 Coffey Aug 2006 B2
7099580 Bulbul Aug 2006 B1
7139283 Quigley et al. Nov 2006 B2
7142609 Terreault et al. Nov 2006 B2
7145887 Akgun et al. Dec 2006 B1
7152025 Lusky et al. Dec 2006 B2
7158542 Zeng et al. Jan 2007 B1
7164694 Nodoushani et al. Jan 2007 B1
7177324 Choudhury et al. Feb 2007 B1
7197067 Lusky et al. Mar 2007 B2
7222255 Claessens et al. May 2007 B1
7227863 Leung et al. Jun 2007 B1
7242862 Saunders et al. Jul 2007 B2
7246368 Millet et al. Jul 2007 B1
7263123 Yousef Aug 2007 B2
7274735 Lusky et al. Sep 2007 B2
7286756 Marshall et al. Oct 2007 B1
7295518 Monk et al. Nov 2007 B1
7315573 Lusky et al. Jan 2008 B2
7315967 Azenko et al. Jan 2008 B2
7400677 Jones Jul 2008 B2
7421276 Steer et al. Sep 2008 B2
7451472 Williams Nov 2008 B2
7492703 Lusky et al. Feb 2009 B2
7554902 Kim et al. Jun 2009 B2
7573884 Klimker et al. Aug 2009 B2
7573935 Min et al. Aug 2009 B2
7584298 Klinker et al. Sep 2009 B2
7616654 Moran, III et al. Nov 2009 B2
7650112 Utsumi et al. Jan 2010 B2
7672310 Cooper et al. Mar 2010 B2
7684315 Beser Mar 2010 B1
7684341 Howald Mar 2010 B2
7693042 Wei Apr 2010 B1
7693090 Kimpe Apr 2010 B1
7701938 Bernstein et al. Apr 2010 B1
7716712 Booth et al. May 2010 B2
7739359 Millet et al. Jun 2010 B1
7742697 Cooper et al. Jun 2010 B2
7742771 Thibeault Jun 2010 B2
7760624 Goodson et al. Jul 2010 B1
7764231 Karr et al. Jul 2010 B1
7778314 Wajcer et al. Aug 2010 B2
7787557 Kim et al. Aug 2010 B2
7792183 Massey et al. Sep 2010 B2
7826569 Popper et al. Nov 2010 B2
7856049 Currivan et al. Dec 2010 B2
7876697 Thompson et al. Jan 2011 B2
7953144 Allen et al. May 2011 B2
7958534 Beser Jun 2011 B1
7970010 Denney et al. Jun 2011 B2
7983162 Ford et al. Jul 2011 B1
8000254 Thompson et al. Aug 2011 B2
8037541 Montague et al. Oct 2011 B2
8040915 Cummings Oct 2011 B2
8059546 Pai et al. Nov 2011 B2
8081674 Thompson et al. Dec 2011 B2
8116360 Thibeault Feb 2012 B2
8265559 Cooper et al. Sep 2012 B2
8279764 Cooper et al. Oct 2012 B2
8284828 Cooper et al. Oct 2012 B2
8345557 Thibeault et al. Jan 2013 B2
8526485 Thompson et al. Sep 2013 B2
8537972 Thompson et al. Sep 2013 B2
8594118 Cooper et al. Nov 2013 B2
20010055319 Quigley et al. Dec 2001 A1
20020038461 White Mar 2002 A1
20020044531 Cooper et al. Apr 2002 A1
20020091970 Furudate et al. Jul 2002 A1
20020116493 Schenkel et al. Aug 2002 A1
20020119783 Bourlas et al. Aug 2002 A1
20020154620 Azenkot et al. Oct 2002 A1
20020168131 Walter et al. Nov 2002 A1
20020181395 Foster et al. Dec 2002 A1
20030028898 Howald Feb 2003 A1
20030043732 Walton et al. Mar 2003 A1
20030067883 Azenkot et al. Apr 2003 A1
20030067944 Sala et al. Apr 2003 A1
20030101463 Greene et al. May 2003 A1
20030108052 Inoue et al. Jun 2003 A1
20030120819 Abramson Jun 2003 A1
20030138250 Glynn Jul 2003 A1
20030149991 Reidhead et al. Aug 2003 A1
20030158940 Leigh Aug 2003 A1
20030179768 Lusky et al. Sep 2003 A1
20030179770 Reznic et al. Sep 2003 A1
20030179821 Lusky et al. Sep 2003 A1
20030181185 Lusky et al. Sep 2003 A1
20030182664 Lusky et al. Sep 2003 A1
20030185176 Lusky et al. Oct 2003 A1
20030188254 Lusky et al. Oct 2003 A1
20030200317 Zeitak et al. Oct 2003 A1
20030212999 Cai Nov 2003 A1
20040015765 Cooper et al. Jan 2004 A1
20040042385 Kim et al. Mar 2004 A1
20040047284 Eidson Mar 2004 A1
20040052248 Frank et al. Mar 2004 A1
20040052356 McKinzie et al. Mar 2004 A1
20040062548 Obeda et al. Apr 2004 A1
20040073937 Williams Apr 2004 A1
20040096216 Ito May 2004 A1
20040109661 Bierman et al. Jun 2004 A1
20040139473 Greene Jul 2004 A1
20040153855 Titmuss Aug 2004 A1
20040163129 Chapman et al. Aug 2004 A1
20040181811 Rakib Sep 2004 A1
20040208513 Peddanarappagari et al. Oct 2004 A1
20040233234 Chaudhry et al. Nov 2004 A1
20040233926 Cummings Nov 2004 A1
20040248520 Miyoshi Dec 2004 A1
20040261119 Williams Dec 2004 A1
20050010958 Rakib et al. Jan 2005 A1
20050025145 Rakib et al. Feb 2005 A1
20050034159 Ophir et al. Feb 2005 A1
20050039103 Azenko et al. Feb 2005 A1
20050058082 Moran et al. Mar 2005 A1
20050064890 Johan et al. Mar 2005 A1
20050097617 Currivan et al. May 2005 A1
20050099951 Mohan et al. May 2005 A1
20050108763 Baran et al. May 2005 A1
20050122996 Azenkot et al. Jun 2005 A1
20050163088 Yamano et al. Jul 2005 A1
20050175080 Bouillett Aug 2005 A1
20050183130 Sadja et al. Aug 2005 A1
20050198688 Fong Sep 2005 A1
20050226161 Jaworski Oct 2005 A1
20050281200 Terreault Dec 2005 A1
20060013147 Terpstra Jan 2006 A1
20060088056 Quigley et al. Apr 2006 A1
20060121946 Walton et al. Jun 2006 A1
20060250967 Miller et al. Nov 2006 A1
20060262722 Chapman et al. Nov 2006 A1
20060291503 Chapman Dec 2006 A1
20070002752 Thibeault et al. Jan 2007 A1
20070030805 Pantelias et al. Feb 2007 A1
20070058542 Thibeault Mar 2007 A1
20070076592 Thibeault Apr 2007 A1
20070076789 Thibeault Apr 2007 A1
20070076790 Thibeault et al. Apr 2007 A1
20070086328 Kao et al. Apr 2007 A1
20070094691 Gazdzinski Apr 2007 A1
20070097907 Cummings May 2007 A1
20070121712 Okamoto May 2007 A1
20070133672 Lee et al. Jun 2007 A1
20070143654 Joyce et al. Jun 2007 A1
20070147489 Sun Jun 2007 A1
20070177526 Siripunkaw et al. Aug 2007 A1
20070184835 Bitran et al. Aug 2007 A1
20070189770 Sucharczuk et al. Aug 2007 A1
20070201547 Willcocks et al. Aug 2007 A1
20070206600 Klimker et al. Sep 2007 A1
20070206625 Maeda Sep 2007 A1
20070211618 Cooper et al. Sep 2007 A1
20070223512 Cooper et al. Sep 2007 A1
20070223513 Pantelias et al. Sep 2007 A1
20070223920 Moore et al. Sep 2007 A1
20070245177 Cooper et al. Oct 2007 A1
20080056713 Cooper et al. Mar 2008 A1
20080062888 Lusky et al. Mar 2008 A1
20080065960 Cheng et al. Mar 2008 A1
20080069006 Walter et al. Mar 2008 A1
20080075157 Allen et al. Mar 2008 A1
20080101210 Thompson et al. May 2008 A1
20080125984 Skendzic et al. May 2008 A1
20080140823 Thompson et al. Jun 2008 A1
20080193137 Thompson et al. Aug 2008 A1
20080200129 Cooper et al. Aug 2008 A1
20080242339 Anderson Oct 2008 A1
20080250508 Montague et al. Oct 2008 A1
20080274700 Li Nov 2008 A1
20080291840 Cooper et al. Nov 2008 A1
20090031384 Brooks et al. Jan 2009 A1
20090103557 Hong et al. Apr 2009 A1
20090103669 Kolze et al. Apr 2009 A1
20090109877 Murray et al. Apr 2009 A1
20090158096 Ali et al. Jun 2009 A1
20090249421 Liu et al. Oct 2009 A1
20090252234 Samdani et al. Oct 2009 A1
20100083356 Steckley et al. Apr 2010 A1
20100095360 Pavlovski et al. Apr 2010 A1
20100128739 Jung et al. May 2010 A1
20100154016 Li et al. Jun 2010 A1
20100154017 An et al. Jun 2010 A1
20100157824 Thompson et al. Jun 2010 A1
20100158093 Thompson et al. Jun 2010 A1
20100185391 Lee et al. Jul 2010 A1
20100223650 Millet et al. Sep 2010 A1
20100251320 Shafer et al. Sep 2010 A1
20100322390 Bialk et al. Dec 2010 A1
20110026577 Primo et al. Feb 2011 A1
20110030019 Ulm et al. Feb 2011 A1
20110069745 Thompson et al. Mar 2011 A1
20110072127 Gerber et al. Mar 2011 A1
20110099570 Sadja et al. Apr 2011 A1
20110110415 Cooper et al. May 2011 A1
20110116387 Beeco et al. May 2011 A1
20110150058 Oh Jun 2011 A1
20110153683 Hoskinson Jun 2011 A1
20110194418 Wolcott et al. Aug 2011 A1
20110194597 Wolcott et al. Aug 2011 A1
20110197071 Wolcott et al. Aug 2011 A1
20110243214 Wolcott et al. Oct 2011 A1
20110252337 Pignataro Oct 2011 A1
20120027069 Clausen et al. Feb 2012 A1
20120054312 Salinger Mar 2012 A1
20120084416 Thibeault et al. Apr 2012 A1
20120093240 McFarland et al. Apr 2012 A1
20120115505 Miyake et al. May 2012 A1
20120147751 Ulm Jun 2012 A1
20120190380 Duprey et al. Jul 2012 A1
20130003565 Gotwals et al. Jan 2013 A1
20130041990 Thibeault et al. Feb 2013 A1
20130051442 Cooper et al. Feb 2013 A1
20130070772 Watson et al. Mar 2013 A1
20130128723 Thibeault et al. May 2013 A1
20130148707 Thibeault et al. Jun 2013 A1
20130286852 Bowler et al. Oct 2013 A1
20130290783 Bowler et al. Oct 2013 A1
20130290791 Basile et al. Oct 2013 A1
20130291034 Basile et al. Oct 2013 A1
20130294489 Thibeault et al. Nov 2013 A1
20140029654 Thompson et al. Jan 2014 A1
20140029655 Thompson et al. Jan 2014 A1
20140133533 Thibeault et al. May 2014 A1
20140185428 Thibeault et al. Jul 2014 A1
20140198630 Nof Jul 2014 A1
20140204954 Nee Jul 2014 A1
20140267788 Bowler et al. Sep 2014 A1
20140269416 Bowler et al. Sep 2014 A1
20140269869 Bowler et al. Sep 2014 A1
20140270095 Bowler et al. Sep 2014 A1
20140278273 Bowler et al. Sep 2014 A1
20140355454 Serban Dec 2014 A1
Foreign Referenced Citations (21)
Number Date Country
69631420 Dec 2004 DE
1235402 Aug 2002 EP
1341335 Sep 2003 EP
55132161 Oct 1980 JP
04208707 Jul 1992 JP
06120896 Apr 1994 JP
06177840 Jun 1994 JP
09008738 Jan 1997 JP
09162816 Jun 1997 JP
10247893 Sep 1998 JP
11230857 Aug 1999 JP
2001044956 Feb 2001 JP
2003530761 Oct 2003 JP
2004172783 Jun 2004 JP
2004343678 Dec 2004 JP
0192901 Dec 2001 WO
2002033980 Apr 2002 WO
0233974 Apr 2002 WO
2004062124 Jul 2004 WO
2008103262 Aug 2008 WO
2009146426 Dec 2009 WO
Non-Patent Literature Citations (38)
Entry
Maintenance Using Pre-Equalization, DOCSIS Best Practices and Guidelines, Cable Television Laboratories, Inc., CM-GL-PNMP-V02-110623, Jun. 23, 2011.
“Proactive Network Maintenance Using Pre-Equalization,” DOCSIS Best Practices and Guidelines, Cable Television Laboratories, Inc., CM-GL-PNMP-V02-110623, Jun. 23, 2011, pp. 133.
R. Hranac, “BER and MER fundamentals”, CISCO, 2007, Retrieved from the Internet: URL:http://www.gcscte.org/presentations/2008/Ron.Hranac_Presentation-BER%20+%20MER%20Fun.pdf, retrieved on Dec. 11, 2014, all pages.
Cable Television Laboratories, Inc., “A Simple Algorithm for Fault Localization Using Naming Convention and Micro-reflection Signature,” Invention Disclosure 60193, 2 pages, Jun. 2008.
Cable Television Laboratories, Inc., “Data-Over-Cable Service Interface Specifications DOCSIS® 3.0—MAC and Upper Layer Protocols Interface Specification,” CM-SP-MULPIv3.0-117-111117, Nov. 2011.
Cable Television Laboratories, Inc., “DOCSIS Best Practices and Guidelines: Proactive Network Maintenace Using Pre-Equalization,” CM-GL-PNMP-V01-100415, Apr. 2010.
Cable Television Laboratories, Inc., “Pre-Equalization Based Pro-active Network Maintenance Process Model for CMs Transmitting on Multiple Upstream Channels,” Invention Disclosure 60203, 2 pages, May 2009.
Cable Television Laboratories, Inc., “Pre-Equalization based pro-active network maintenance process model”, Invention Disclosure 60177, 2 pages, Jun. 2008.
Cable Television Laboratories, Inc., “DOCSIS® Best Practices and Guidelines: Proactive Network Maintenance Using Pre-equalization,” CM-GL-PNMP-V02-110623, Jun. 2011.
Cable Television Laboratories, Inc., “Data-Over-Cable Service Interface Specifications: DOCSIS 2.0 Radio Frequency Interface Specification,” CM-SP-RFIv2.0-I06-040804, Apr. 2004.
L.A. Campos, et al., “Pre-equalization based Pro-active Network Maintenance Methodology”, Cable Television Laboratories, Inc., presentation, 32 pages, 2008.
R.L. Howald, et al., “Customized Broadband—Analysis Techniques for Blended Multiplexes,” NCTA Technical Papers, 2002.
R. Howald, “Access Networks Solutions: Introduction to S-CDMA,” Presentation to Society of Cable Telecommunications Engineers (SCTE) South Florida Chapter, 2009.
R. Howald, “Upstream Snapshots & Indicators (2009),” Regional Samples, Presentation to Society of Cable Telecommunications Engineers (SCTE) South Florida Chapter, Jan. 2010.
R.L. Howald et al., “Characterizing and Aligning the HFC Return Path for Successful DOCSIS 3.0 Rollouts”, SCTE Cable—Tec Expo, Oct. 2009.
R. Howald, et al., “DOCSIS 3.0 Upstream: Readiness & Qualification,” SCTE Cable—Tec Expo, Oct. 2009.
R. Howald, et al., “The Grown-Up Potential of a Teenage PHY”, NCTA Convention and Exposition, May 2012.
R. Howald, “DOCSIS 3.0 Upstream: Technology, RF Variables & Case Studies,” Access Networks Solutions, 2009, presentation to Society of Cable Telecommunications Engineers (SCTE) South Florida Chapter, 23 pages, Jan. 2010.
R. Hranac, “Linear Distortions, Part 1,” Communication Technology, Jul. 2005.
X. Liu, et al., “Variable Bit Rate Video Services in DOCSIS 3.0 Networks,” NCTA Technical Papers, 2008.
Motorola, Inc., “White Paper: Expanding Bandwidth Using Advanced Spectrum Management,” Sep. 25, 2003.
H. Newton, Newton's Telecom Dictionary, Flatiron Publishing, 9th ed., pp. 216 and 1023 (definitions of “carrier to noise ratio” and “signal to noise ratio”), Sep. 1995.
M. Patrick, et al., “Delivering Economical IP Video over DOCSIS by Bypassing the M-CMTS with DIBA,” SCTE 2007 EmergingTechnologies, NCTA Technical Papers, 2007.
A. Popper, et al., “An Advanced Receiver with Interference Cancellation for Broadband Cable Networks,” 2002 International Zurich Seminar on Broadband Communications—Access, Transmission, Networking, pp. 23-1-23-6, IEEE, 2002.
A. Popper, et al, “Ingress Noise Cancellation for the Upstream Channel in Broadband Cable Access Systems,” 2002 IEEE International Conference on Communications, vol. 3, pp. 1808-1812. IEEE, 2002.
S.U.H. Qureshi, “Adaptive Equalization,” Proceedings of the IEEE, vol. 73, No. 9, pp. 1349-1387, Sep. 1985.
S. Ramakrishnan, “Scaling the DOCSIS Network for IPTV,” SCTE Conference on Emerging Technologies, NCTA Cable Show, Apr. 2009.
Y.R. Shelke, “Knowledge Based Topology Discovery and Geo-localization”, Thesis, Master of Science, Ohio State University, 2010.
R. Thompson, et al., “256-QAM for Upstream HFC,” NCTA 2010 Spring Technical Forum Proceedings, pp. 142-152 May 2010.
R. Thompson, et al., “256-QAM for Upstream HFC Part Two”, SCTE Cable-Tec Expo 2011, Technical Paper, Nov. 2011.
R. Thompson, et al., “Multiple Access Made Easy,” SCTE Cable-Tec Expo 2011, Technical Paper, Nov. 2011.
R. Thompson, et al., “Optimizing Upstream Throughput Using Equalization Coefficient Analysis”, National Cable & Telecommunications Association (NCTA) Technical Papers, Apr. 2009.
R. Thompson, et al., “Practical Considerations for Migrating the Network Toward All-Digital”, Society of Cable Telecommunications Engineers (SCTE) Cable-Tec Expo, Oct. 2009.
R. Thompson, et al., “64-QAM, 6.4MHz Upstream Deployment Challenges,” SCTE Canadian Summit, Toronto, Canada, Technical Paper, Mar. 2011.
B. Volpe, et al., “Cable-Tec Expo 2011: Advanced Troubleshooting in a DOCSIS© 3.0 Plant,” Nov. 2011.
L. Wolcott, “Modem Signal Usage and Fault Isolation,” U.S. Appl. No. 61/301,835, filed Feb. 5, 2010.
F. Zhao, et al., “Techniques for minimizing error propagation in decision feedback detectors for recording channels,” IEEE Transactions on Magnetics, vol. 37, No. 1, pp. 592-602, Jan. 2001.
Y. Morishita, et al., “An LMS adaptive equalizer using threshold in impulse noise environments”, IEEE, ICT 2003 10th International Conference on Telecommunications, vol. 1, pp. 578-582, Feb. 2003.
Related Publications (1)
Number Date Country
20140267788 A1 Sep 2014 US