Software applications executing on a computer system may fail for a wide variety of reasons, such as code bugs, user errors, bad input data, unavailable resources, or the like. Such application failures may result in loss of data and application downtime, and may incur costs and time related to recovery of the application and data. Applications running in a common environment or from a common installation may be expected to encounter the same failures given the same inputs, conditions, and/or circumstances. Such may be the case with applications running in a virtualized application environment.
Application virtualization allows software applications executed by a computer to be decoupled from the hardware, operating system (“OS”), and local configuration of the computer. Application virtualization may remove the requirement for an application to be installed, configured, and maintained locally on the computer. Instead, a virtual application environment may execute on the computer and stream the application components across a network from a virtualized application package maintained centrally on a virtual application server.
It is with respect to these considerations and others that the disclosure made herein is presented.
Technologies are described herein for differentiating normal operation of an application program from error conditions to predict, diagnose, and recover from application failures. When an application program is running in a virtual application environment, the virtualization layer or virtual application environment may have knowledge of and control over requests the application makes for resources, such as reads from data files, writes to registry keys, and the like. Utilizing the technologies described herein, the virtualization layer may log accesses to resources and, over time, establish common patterns of resource usage. Once such resource access patterns are established, the virtualization layer may continue to monitor resource usage by the application program and provide a warning or alert when the patterns change. This proactive warning may provide the user of the application program or an administrator a chance to take diagnostic or corrective actions quickly, thus reducing or even preventing downtime and data loss.
It will be appreciated that application programs typically have a means of displaying errors to the user or administrator, like a pop-up dialog or an event logged in an application or system event log. The quality and utility of these error messages, however, may vary dramatically from application to application. Providing high-quality error messages that allow the source of the error to be determined requires a significant investment, and not all software vendors make this investment. Leveraging the log of accesses to resources and common patterns established described herein to correlate which resources were being accessed or how the pattern changed just before the error condition may allow users and/or administrators to more quickly diagnose application failures and implement recovery actions, reducing application downtime.
In addition, because the virtualization layer is aware of all resource usage by the application program, it can log resource additions, modifications, or deletions along with the data used in these resource modifications that occur over time. If the application program fails, a second instance of the application program can be started immediately and the log of resource modifications and data can be replayed, thus restoring the application state to a point just before failure of the first instance of the application. Such rapid failover between application instances may limit further downtime.
According to embodiments, access to resources by the application program executing in a virtual application environment is monitored, and resource access events are logged in a resource access log. Resource access patterns are established from the logged resource access events utilizing computer pattern recognition techniques. If subsequent access to resources by the application program deviates from the established patterns, then a user and/or administrator of the application program is notified of a potential error condition based on the detected deviation.
Further, sequences of resource access events that deviate from the established resources access patterns may be correlated with an error condition based on a temporal proximity to the time of occurrence of the error condition to provide diagnostic information to the user and/or administrator regarding the error. Finally, resource access events regarding the addition, modification, or deletion of data logged in the resource access log may be replayed in the event of an application failure and subsequent restart of the application to re-establish the application state of the application program.
It should be appreciated that the above-described subject matter may be implemented as a computer-controlled apparatus, a computer process, a computing system, or as an article of manufacture such as a computer-readable medium. These and various other features will be apparent from a reading of the following Detailed Description and a review of the associated drawings.
This Summary is provided to introduce a selection of concepts in a simplified form that are further described below in the Detailed Description. This Summary is not intended to identify key features or essential features of the claimed subject matter, nor is it intended that this Summary be used to limit the scope of the claimed subject matter. Furthermore, the claimed subject matter is not limited to implementations that solve any or all disadvantages noted in any part of this disclosure.
The following detailed description is directed to technologies for differentiating normal operation of an application program from error conditions based on patterns of resource access in order to predict, diagnose, and/or recover from application failures. While the subject matter described herein is presented in the general context of program modules that execute in conjunction with the execution of an operating system and application programs on a computer system, those skilled in the art will recognize that other implementations may be performed in combination with other types of program modules. Generally, program modules include routines, programs, components, data structures, and other types of structures that perform particular tasks or implement particular abstract data types. Moreover, those skilled in the art will appreciate that the subject matter described herein may be practiced with other computer system configurations, including hand-held devices, multiprocessor systems, microprocessor-based or programmable consumer electronics, minicomputers, mainframe computers, and the like.
In the following detailed description, references are made to the accompanying drawings that form a part hereof and that show, by way of illustration, specific embodiments or examples. In the accompanying drawings, like numerals represent like elements through the several figures.
A software application program 104 executes on the computer 102. According to embodiments, the application program 104 may execute inside a virtual application environment 106. The virtual application environment 106 may allow the computer 102 to launch and execute application programs that have not been previously installed on the computer. The virtual application environment 106 may instead stream the components of the application program 104 in real-time or near real-time over a network 110 from a virtual application server 112. The virtual application environment 106 and virtual application server 112 may be based upon the MICROSOFT® APP-V technology from MICROSOFT Corporation of Redmond, Wash., the CITRIX XENAPP™ technology from CITRIX SYSTEMS Inc. of Fort Lauderdale, Fla., or any other application streaming and virtualization platform or technologies. The network 110 may be a LAN, a wide-area network (“WAN”), the Internet, or any other networking topology that connects the computer 102 to the virtual application server 112.
The software components of the application program 104 may be stored in a virtualized application package 114 located on a storage device accessible by the virtual application server 112. According to embodiments, the virtualized application package 114 consists of a number of blocks of data that contain application program structure information as well as the individual component files and other elements of the application. The virtualized application package 114 may further contain metadata regarding the location and configuration of local and remote resources utilized by the application program 104 during execution. The virtualized application package 114 may be created by an administrator of the application program 104 by performing a typical installation of the application on a management server and recording the changes made to the local file system, registry, and other, for example. The blocks in the virtualized application package 114 may then be streamed to the virtual application environment 106 to allow the application program 104 to be executed on the computer 102
The virtual application environment 106 may create a separate virtual runtime environment, referred to as an “application sandbox,” to execute each application program 104 streamed from the virtual application server 112. The application sandbox allows the components of the application program 104 to execute in isolation from the remainder of the system. The virtual application environment 106 may further provide a virtualization layer 116 that abstracts access to local resources 118A and remote resources 118B (referred to herein generally as resources 118) utilized by the application program 104 during execution. The resources 118 may include system memory, local processor time or processing threads, files stored in a file system, data stored in a registry database, application services, presentation services, database services, and the like available locally on the computer 102 or remotely across the network 110.
The application program 104 may access the local and remote resources 118 through resource application programming interfaces (“APIs”) 120 implemented by an operating system 122 or other standard software libraries installed on the computer 102. According to embodiments, the virtualization layer 116 abstracts the resource APIs 120 in order monitor and control access requests to the local and remote resources 118 by the application program 104 executing in the virtual application environment 106. In addition, the virtualization layer 116 may log access to the resources 118 by the application program 104 in a resource access log 124. The resource access log 124 may comprise a log file in the local file system, a number of database tables on a remote database server, a combination of the two, or any other data storage system accessible by the computer 102.
The resource access log 124 may contain a log of resource access events 126. The resource access events 126 may include details of calls to the resource APIs 120 made by the application program 104 executing in the virtual application environment 106. Each of the resource access events 126 may include a timestamp indicating when the resource access occurred, an identifier of the individual resource API 120 called, and a number of parameter values indicating the resource type, location, or other aspects of the local or remote resources 118 being accessed. The resource access events 126 may be stored as an entries in a log file, as rows in a database table, as objects in a dictionary, or in any other data structure or format known in the art.
In one embodiment, the resource access log 124 further contains resource access patterns 128. The resource access patterns 128 may comprise patterns of resource access by the application program 104 that regularly occur. For example, the application program 104 may read a particular registry key at time T1 and then write to a particular file located on a remote file system at time T1+240 msec. Further this pattern of reading the registry key and writing to the file may occur more than once, either in response to a particular event or condition, periodically, or at a particular time of day, for example. Resource access patterns 128 may be established between specific API calls, between API calls for specific resources, or between API calls for specific quantities of resources, such as an amount of memory allocated or a number of threads started, for example.
The resource access patterns 128 may be established from the resource access events 126 collected over some period of time using pattern recognition techniques. For example, a subset of event types may be determined to be important, and Bayesian learning techniques may be utilized to establish the resource access patterns 128 across those types of resource access events 126 within the collection period. The generated resource access patterns 128 may be stored as Markov chains or probability trees indicating a relative probability of occurrence between the resource access events, for example.
According to one embodiment, the resource access patterns 128 may be generated in near-real-time on the computer from the log of resource access events 126 in a background process running in parallel to the execution of the application program 104. In another embodiment, resource access events 126 logged by the virtualization layer 116 on a number of computers 102 executing the application program 104 in the virtual application environment 106 may be aggregated at a central location. The aggregated event data may be genericized, by removing computer-dependent resource paths, for example, and resource access patterns 128 may be established from the aggregated and genericized event data. The generated resource access patterns 128 may then be sent to each of the computers 102 executing the application program 104 in order for the patterns to be utilized in predicting application failures, as will be described in more detail below in regard to
In a further embodiment, the resource access log 124 also contains resource write data 130. The resource write data 130 may contain a log of calls by the application program 104 to resource APIs 120 that add, modify, or delete data, such as a registry value write or an I/O buffer write. The resource write data 130 may contain a “deep copy” from pointers or structure-type parameters in order to include the data being written. In addition, any filename, key name, address, or other location parameters may be genericized using the current context of the executing application program 104. It will be appreciated that the resource write data 130 and the resource access events 126 may be integrated into a single log file or other structure in the resource access log 124. The resource write data 130 may be utilized to restore the application state of the application program 104 during recovery from an application failure, as will be described in more detail below in regard to
Referring now to
The routine 200 begins at operation 202, where the virtualization layer 116 monitors access to local and remote resources 118 by the application program 104 executing in the virtual application environment 106, and logs these accesses to the resource access log 124. As described above in regard to
From operation 202, the routine 200 proceeds to operation 204, where resource access patterns 128 are established. It will be appreciated that, over some period of time, a significant number of resource access events 126 may be logged in the resource access log 124 by the virtualization layer 116. As described above in regard to
The resource access patterns 128 may be generated in near-real-time on the computer 102 by the virtualization layer 116. Alternatively, logged resource access events 126 may be aggregated from a number of computers 102 at a central location, genericized, and utilized to establish resource access patterns 128 across multiple instances of the application program 104 executing in the virtual application environment 106 across the computers. The generic resource access patterns 128 established from the aggregated resource access events 126 may then be utilized to predict error conditions on any computer 102 executing the application program 104 in the manner described herein.
The routine 200 proceeds from operation 204 to operation 206, where the virtualization layer 116 detects a deviation from the established resource access patterns 128 by the application program 104 executing in the virtual application environment 106. For example, the virtualization layer 116 may detect a sequence of resource API calls that has a probability of occurrence below a specified threshold based on Bayesian analysis of the resource access patterns 128. Similarly, the virtualization layer 116 may detect a sequence of resource API calls that has a high probability of occurrence with a known error condition, as established in the resource access patterns 128. In one embodiment, if the probability of the sequence of resource API calls does not fall below the specific threshold, then the virtualization layer 116 logs the corresponding resource access events 126 so that the resource access patterns 128 may be updated with new probabilities in the background process described above. In this way, the resource access patterns 128 may be continually updated during execution of the application program 104 in the virtual application environment 106.
If a deviation from the established resource access patterns 128 by the application program 104 is detected, then the routine 200 proceeds from operation 206 to operation 208, where the virtualization layer 116 raises an alert regarding the deviation in the pattern. The alert may be directed to a user or administrator of the application program 104. The alert may be sent via e-mail, text message, or system message queue; raised as a system level event; logged in an application or system event log, or otherwise transmitted to the administrator via a messaging system accessible by the computer 102. This proactive alert may give the administrator a chance to take diagnostic or corrective actions quickly, reducing or perhaps preventing downtime and data loss in the wake of a potential, pending error condition. From operation 208, the routine 200 ends.
The routine 300 begins at operation 302, where the virtualization layer 116 monitors access to local and remote resources 118 by the application program 104 executing in the virtual application environment 106 and logs these accesses to the resource access log 124, in the manner described above in regard to operation 202. The routine 300 then proceeds to operation 304, where an error condition in the application program 104 is detected. For example, the error condition may be detected in the application program 104 by a user or administrator of the application through traditional means, such as a pop-up error dialog, an event logged in an application or system event log, or the like.
From operation 304, the routine 300 proceeds to operation 306, where resource access events 126 in the resource access log 124 are correlated with the detected error condition. The administrator may provide a time of occurrence of the error condition, or the time of occurrence of the error condition may be identified from specific calls to resource APIs 120 logged in the resource access events 126. The virtualization layer 116 or other module may then identify a subset of resource access events 126 in the resource access log 124 within a temporal proximity to the time of occurrence of the error condition. For example, all resource access events 126 that occurred within a 10 second window leading up to the error condition may be correlated with the error condition.
In one embodiment, only sequences of resource API calls logged in the resource access events 126 within the temporal proximity to the error condition that deviate from established resource access patterns 128 are correlated with the error condition. For example, the virtualization layer 116 or other module may identify sequences of resource API calls logged in the resource access events 126 within 10 seconds of the time occurrence of the error condition that have a probability of occurrence below a specified threshold based on Bayesian analysis of the resource access patterns 128 in the resource access log 124. It will be appreciated that the specified threshold of probability for correlating deviations in resource access patterns 128 with an error condition may be higher than the threshold of probability for predicting an error condition based on deviations in resource access patterns described above in regard to operation 206.
The routine 300 then proceeds from operation 306 to operation 308, where the resource access events 126 correlated with the error condition are shown to the user or administrator of the application program 104. The correlated resource access events 126 may be shown through a user interface dialog or in a report transmitted via e-mail, text message, system message queue, or the like. Providing resource access events 126 from the resource access log 124 that are correlated in time with a particular error condition may allow the user or administrator to quickly diagnose the cause of the error condition and implement the appropriate recovery actions in order to reduce downtime and data loss. From operation 308, the routine 300 ends.
The routine 400 begins at operation 402, where the virtualization layer 116 monitors access to local and remote resources 118 by the application program 104 executing in the virtual application environment 106 and logs these accesses to the resource access log 124, in the manner described above in regard to operation 202. In addition, the virtualization layer 116 logs resource write data 130 for calls by the application program 104 to resource APIs 120 that add, modify, or delete data, as described above in regard to
From operation 402, the routine 400 proceeds to operation 404, where the application program 104 fails due to an error condition. For instance, the application program 104 may fail because of a software bug, a user error, bad input data, unavailable resources, a hardware failure in the computer 102, or the like. Upon failure of the application program 104, the routine 400 moves to operation 406, where the application program 104 is restarted. The application program 104 may be restarted automatically by the virtualization layer 116 or another module executing on the computer 102, or it may be restarted manually by a system administrator on the same computer or another computer system with similar configuration.
The routine 400 proceeds from operation 406 to operation 408, where the virtualization layer 116 executing in the virtual application environment 106 on the computer 102 where the application program 104 was restarted replays certain resource access events 126 logged in the resource access log 124 in order to restore the application state to a point before the failure occurred. For example, the virtualization layer 116 may replay all resource access events 126 corresponding to calls to resource APIs 120 that write to volatile or cached storage locations, such as system memory addresses, I/O buffers, cached files, or the like.
In a further embodiment, the virtualization layer 116 may replay all resource access events 126 corresponding to the writing of data that occurred since a last snapshot of the application state, or “checkpoint,” was taken and stored by the virtualization layer and/or application program 104 before the application failure. The virtualization layer 116 may utilize the resource write data 130 in the resource access log 124 to replay the selected resource access events 126, in order to ensure the proper data is written in restoring the application state. Logging resource access events 126 that add, modify, or delete data along with the corresponding resource write data 130 in order to replay the writes to restore application state may allow for faster recovery from application failures, thus reducing application downtime. From operation 408, the routine 400 ends.
While the present disclosure is described in the context of a virtualized application environment 106, it will be appreciated that the methods presented herein for differentiating normal operation of an application program from error conditions to predict, diagnose, and recover from application failures may be implemented in any other application environment where the access of application programs 104 to local and remote resources 118 may be monitored. For example a module similar to the virtualization layer 116 may be implemented that utilizes methods known in the art to hook the resources APIs 120 implemented by the operating system 122 in order to monitor requests for local and remote resources 118 by a locally executing applications program 104 outside of a virtual application environment 106. The module may log the accesses to the resources and establish common patterns of resource usage by the application program 104. Once such resource access patterns are established, the module may continue to monitor resource usage by the application program 104 and provide a warning or alert when the patterns change, in the manner described herein.
The computer architecture shown in
The computer architecture further includes a system memory 508, including a random access memory (“RAM”) 514 and a read-only memory 516 (“ROM”), and a system bus 504 that couples the memory to the CPUs 502. A basic input/output system containing the basic routines that help to transfer information between elements within the computer 500, such as during startup, is stored in the ROM 516. The computer 500 also includes a mass storage device 510 for storing an operating system 122, application programs, and other program modules, which are described in greater detail herein.
The mass storage device 510 is connected to the CPUs 502 through a mass storage controller (not shown) connected to the bus 504. The mass storage device 510 provides non-volatile storage for the computer 500. The computer 500 may store information on the mass storage device 510 by transforming the physical state of the device to reflect the information being stored. The specific transformation of physical state may depend on various factors, in different implementations of this description. Examples of such factors may include, but are not limited to, the technology used to implement the mass storage device, whether the mass storage device is characterized as primary or secondary storage, and the like.
For example, the computer 500 may store information to the mass storage device 510 by issuing instructions to the mass storage controller to alter the magnetic characteristics of a particular location within a magnetic disk drive, the reflective or refractive characteristics of a particular location in an optical storage device, or the electrical characteristics of a particular capacitor, transistor, or other discrete component in a solid-state storage device. Other transformations of physical media are possible without departing from the scope and spirit of the present description. The computer 500 may further read information from the mass storage device 510 by detecting the physical states or characteristics of one or more particular locations within the mass storage device.
As mentioned briefly above, a number of program modules and data files may be stored in the mass storage device 510 and RAM 514 of the computer 500, including an operating system 518 suitable for controlling the operation of a computer. The mass storage device 510 and RAM 514 may also store one or more program modules. In particular, the mass storage device 510 and the RAM 514 may store the virtualization layer 116, which was described in detail above in regard to
In addition to the mass storage device 510 described above, the computer 500 may have access to other computer-readable media to store and retrieve information, such as program modules, data structures, or other data. It should be appreciated by those skilled in the art that computer-readable media may be any available media that can be accessed by the computer 500, including computer-readable storage media and communications media. Communications media includes transitory signals. Computer-readable storage media includes volatile and non-volatile, removable and non-removable media implemented in any method or technology for the storage of information, such as computer-readable instructions, data structures, program modules, or other data. For example, computer-readable storage media includes, but is not limited to, RAM, ROM, EPROM, EEPROM, flash memory or other solid state memory technology, CD-ROM, digital versatile disks (DVD), HD-DVD, BLU-RAY, or other optical storage, magnetic cassettes, magnetic tape, magnetic disk storage or other magnetic storage devices, or any other medium that can be used to store the desired information and that can be accessed by the computer 500.
The computer-readable storage medium may be encoded with computer-executable instructions that, when loaded into the computer 500, may transform the computer system from a general-purpose computing system into a special-purpose computer capable of implementing the embodiments described herein. The computer-executable instructions may be encoded on the computer-readable storage medium by altering the electrical, optical, magnetic, or other physical characteristics of particular locations within the media. These computer-executable instructions transform the computer 500 by specifying how the CPUs 502 transition between states, as described above. According to one embodiment, the computer 500 may have access to computer-readable storage media storing computer-executable instructions that, when executed by the computer, perform the routines 200, 300, and/or 400 for differentiating normal operation of an application program from error conditions to predict, diagnose, and recover from application failures, described above in regard to
According to various embodiments, the computer 500 may operate in a networked environment using logical connections to remote computing devices and computer systems through the network 110, such as a LAN, a WAN, the Internet, or a network of any topology known in the art. The computer 500 may connect to the network 110 through a network interface unit 506 connected to the bus 504. It should be appreciated that the network interface unit 506 may also be utilized to connect to other types of networks and remote computer systems.
The computer 500 may also include an input/output controller 512 for receiving and processing input from a number of input devices, including a keyboard, a mouse, a touchpad, a touch screen, an electronic stylus, or other type of input device. Similarly, the input/output controller 512 may provide output to a display device, such as a computer monitor, a flat-panel display, a digital projector, a printer, a plotter, or other type of output device. It will be appreciated that the computer 500 may not include all of the components shown in
Based on the foregoing, it should be appreciated that technologies for differentiating normal operation of an application program from error conditions to predict, diagnose, and recover from application failures are provided herein. Although the subject matter presented herein has been described in language specific to computer structural features, methodological acts, and computer-readable storage media, it is to be understood that the invention defined in the appended claims is not necessarily limited to the specific features, acts, or media described herein. Rather, the specific features, acts, and mediums are disclosed as example forms of implementing the claims.
The subject matter described above is provided by way of illustration only and should not be construed as limiting. Various modifications and changes may be made to the subject matter described herein without following the example embodiments and applications illustrated and described, and without departing from the true spirit and scope of the present invention, which is set forth in the following claims.
This application is a continuation of co-pending U.S. patent application Ser. No. 14/938,542, filed Nov. 11, 2015, entitled “PREDICTING, DIAGNOSING, AND RECOVERING FROM APPLICATION FAILURES BASED ON RESOURCE ACCESS PATTERNS,” which is a continuation of U.S. patent application Ser. No. 12/978,663, filed Dec. 27, 2010, entitled “PREDICTING, DIAGNOSING, AND RECOVERING FROM APPLICATION FAILURES BASED ON RESOURCE ACCESS PATTERNS,” and which issued as U.S. Pat. No. 9,189,308 on Nov. 17, 2015. The entire contents of each of these applications are incorporated by reference herein in their entireties.
Number | Name | Date | Kind |
---|---|---|---|
6065146 | Bosshart | May 2000 | A |
6149887 | Lagas | Nov 2000 | A |
6360344 | Khoche | Mar 2002 | B1 |
6442726 | Knefel | Aug 2002 | B1 |
6745350 | Cline | Jun 2004 | B1 |
6763370 | Schmeidler | Jul 2004 | B1 |
6938096 | Greschler | Aug 2005 | B1 |
7017188 | Schmeidler | Mar 2006 | B1 |
7028305 | Schaefer | Apr 2006 | B2 |
7103625 | Hipp | Sep 2006 | B1 |
7117393 | Baba | Oct 2006 | B2 |
7165190 | Srivastava | Jan 2007 | B1 |
7200632 | Greschler | Apr 2007 | B1 |
7225264 | Croman | May 2007 | B2 |
7243267 | Klemm | Jul 2007 | B2 |
7426661 | Schaefer | Sep 2008 | B2 |
7451451 | Schaefer | Nov 2008 | B2 |
7467333 | Keeton | Dec 2008 | B2 |
7590743 | Willis | Sep 2009 | B2 |
7634681 | Takamoto | Dec 2009 | B2 |
7647530 | Britt | Jan 2010 | B2 |
7690039 | Schmeidler | Mar 2010 | B2 |
7707641 | Schmeidler | Apr 2010 | B2 |
7730169 | Greschler | Jun 2010 | B1 |
7770064 | Harper | Aug 2010 | B2 |
7774647 | Vertes | Aug 2010 | B2 |
7797372 | Greschler | Sep 2010 | B2 |
7921341 | Abe | Apr 2011 | B2 |
8024607 | Ladd | Sep 2011 | B2 |
8069374 | Panigrahy | Nov 2011 | B2 |
8103944 | Treigherman | Jan 2012 | B2 |
8190982 | Wang | May 2012 | B2 |
20050102567 | McGuire | May 2005 | A1 |
20060242288 | Masurkar | Oct 2006 | A1 |
20070220376 | Furukawa | Sep 2007 | A1 |
20080250265 | Chang | Oct 2008 | A1 |
20090119549 | Vertes | May 2009 | A1 |
20090165135 | Lomont | Jun 2009 | A1 |
20090199178 | Keller | Aug 2009 | A1 |
20090222815 | Dake | Sep 2009 | A1 |
20100058108 | Nammatsu | Mar 2010 | A1 |
20100211829 | Ziskind | Aug 2010 | A1 |
20100223499 | Panigrahy | Sep 2010 | A1 |
20110006808 | de Gruijl | Jan 2011 | A1 |
20120166869 | Young | Jun 2012 | A1 |
Number | Date | Country | |
---|---|---|---|
20190073258 A1 | Mar 2019 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 14938542 | Nov 2015 | US |
Child | 16179360 | US | |
Parent | 12978663 | Dec 2010 | US |
Child | 14938542 | US |