Enterprise threat detection (ETD) typically collects and stores a large amount/large sets of log data associated with various heterogeneous systems (often referred to as “big data”). The stored data can be analyzed computationally using forensic-type data analysis tools to identify suspicious behavior in revealed patterns, trends, interactions, and associations, especially relating to ETD behavior. Appropriate responses can then be taken if malicious behavior is suspected or identified. The forensic environment permits the creation of ETD patterns based on created filters to generate alerts when log data meeting the filter criteria is detected.
The present disclosure describes creating patterns in enterprise threat detection (ETD).
In an implementation, Search results are received from an initiated free text search of log data from one or more logs, where the free text is performed using search terms entered into a free text search graphical user interface. A set of at least one search result is selected from the search results containing an event desired to be identified in a completed ETD pattern (or pattern). A forensic lab application is rendered to complete an ETD pattern. An event filter is added for an event type based on normalized log data to a path. A relative ETD pattern time range is set and an ETD pattern is based on the added event filter.
The previously described implementation is implementable using a computer-implemented method; a non-transitory, computer-readable medium storing computer-readable instructions to perform the computer-implemented method; and a computer-implemented system comprising a computer memory interoperably coupled with a hardware processor configured to perform the computer-implemented method/the instructions stored on the non-transitory, computer-readable medium.
The subject matter described in this specification can be implemented in particular implementations, so as to realize one or more of the following advantages. First, the described methodology permits search over raw logs in an interactive manner. Second, once search results correspond to desired requirements, an ETD pattern based on normalized logs can be completed. Third, a user does not need knowledge about ETD events as searches for known search terms in ETD logs can be used to complete ETD patterns. Other advantages will be apparent to those of ordinary skill in the art.
The details of one or more implementations of the subject matter of this specification are set forth in the accompanying drawings and the description. Other features, aspects, and advantages of the subject matter will become apparent from the description, the drawings, and the claims.
Like reference numbers and designations in the various drawings indicate like elements.
The following detailed description describes completing patterns in enterprise threat detection (ETD), and is presented to enable any person skilled in the art to make and use the disclosed subject matter in the context of one or more particular implementations. Various modifications, alterations, and permutations of the disclosed implementations can be made and will be readily apparent to those or ordinary skill in the art, and the general principles defined may be applied to other implementations and applications, without departing from scope of the disclosure. In some instances, details unnecessary to obtain an understanding of the described subject matter may be omitted so as to not obscure one or more described implementations with unnecessary detail and inasmuch as such details are within the skill of one of ordinary skill in the art. The present disclosure is not intended to be limited to the described or illustrated implementations, but to be accorded the widest scope consistent with the described principles and features.
ETD typically collects and stores a large amount/large sets of log data associated with various heterogeneous systems (often referred to as “big data”). The collected log data is usually analyzed using forensic-type data analysis tools in a forensic environment to identify suspicious behavior and to allow an appropriate response.
ETD normalizes logs and keeps a reference (for example, with a universally unique identifier (UUID)) from each raw log to a corresponding normalized log. In the forensic lab environment, event types (semantic attributes) based on normalized log fields are offered to search over logs, to create a series (“path”) of one or more filters on these sematic attributes. ETD patterns for attack detection can be based on paths. When log data meeting the filter criteria is detected, an alert can be generated for further investigation.
Additionally, in the forensic lab environment, an attribute “Event, Original Message” is offered. This event type contains raw log data corresponding to a normalized log. In the forensic lab environment, it is possible to create efficient ETD patterns based on normalized log fields and less efficient ETD patterns based on the attribute “Event, Original Message”, for which execution takes more time and consumes more resources.
ETD also offers a search UI free search functionality (illustrated as “Searchlog”) permitting a free search over persisted log data (for example, for multiple different systems and time frames). In some implementations, the search can be performed independently of whether the log data is recognized (that is, interpreted) or normalized. This means that unknown log formats and raw data can be searched by the provided search functionality.
To assist with understanding, the disclosure provides an example of the free search and associated ETD pattern creation functionality using a filter for log data using a protocol with the “Fail write file” action. As will be appreciated by those of ordinary skill in the art, this is only one possible example. This example is not meant to limit or to constrain the disclosure in any way. Other examples and functionality consistent with this disclosure are considered to be within the scope of this disclosure.
Received refined search results 302 also include three log entries 306a, 306b, and 306c. Note that each log entry includes the terms “Failed,” “write,” and “file” (at 308) as specified for refinement. Similar to the original search, the results are typically based on an “AND” type search (that is, “Failed” AND “write” AND “file”), other possible search types for refinement are also possible (for example, an “OR”-type search). Additionally, some implementations can permit advanced search term specification with more complicated logical relationships or regular expressions.
The ability to use a free term search is a natural method of searching. The ease of use permits fast and instinctive changing and refinement of search terms to search for desired results.
Raw data is data produced by systems or applications. In some implementations, the raw data (for example, log entries) is persisted in ETD in one database field as one binary large object (BLOB) or one STRING per log. Such raw log data makes it possible to search by a user's own search terms in the BLOB/STRING field of raw data in a free text search manner.
In contrast, the content of each raw data log is categorized/distributed into many different database fields when generating normalized log data. Each normalized log has additional database fields added with context information (for example, for a particular dynamic IP address, a device MAC address and host name can be saved). Performance optimized ETD patterns can only run over normalized data. General ETD patterns running over raw data are significantly slower as data is not categorized/distributed and must be searched.
In the forensic lab 502, individual filters based on raw or normalized data can be defined. The filters are associated with a path 503 (here “Path 1”).
The bubblegram 504 presents many “bubbles” in a user interface. The distribution of events associated with path 503 is illustrated to be by distribution value 505 (here “Correlation ID).” Here, “Correlation ID” is one of possible selectable dimensions of the applicable log data. Bubblegram 504 displays the entire data distribution and count of events for path 503. Generally, all event types that contain values are shown in the bubblegram 504. Some technical events, such as timestamp or timestamp of insertion (where the distribution of values is very high), can, in some implementations, be configured to not be displayed. Bubblegram 504 only shows only dimensions that contain values to filter out non-applicable dimensions from review.
Each bubble of the bubblegram 504 illustrates a particular count/distribution of dimensions/attributes in the path 503, which are not zero. For example, a bubble displaying the dimension “Service, Program Name 71(2)” means that there is a count of 71 event values with 2 distinct values in the overall 71 log event count (in this case, some of the 2 distinct values are repeated which results in only 2 distinct overall values).
In some implementations, ranking of bubbles in the bubblegram 504 can be used as a measure of relevance (or relatedness) of the dimension/attribute associated with the particular bubble to the dimension/attribute represented by the overall bubblegram 504. In typical implementations, the size of each of the displayed bubbles is mapped to a ranking value of the distribution. Accordingly, smaller sized bubbles can be either ignored or considered less important unless the associated dimension/attribute is of particular known relevance.
In some implementations, the bubbles are positioned according to their rankings, typically with the highest ranking bubbles in the middle of the bubblegram and the lowest ranking bubbles positioned outwardly from the middle toward the edges of the bubblegram as ranking decreases. Larger-sized bubbles are placed more toward the middle of the bubblegram and are more easily seen and considered to be of higher relevance to the associated dimension/attribute represented by the overall bubblegram 504.
The bubbles of the bubblegram 504 illustrate, within the set time range (for example, approximately 1 hour), a distribution of event types based on normalized log data that meets filter criteria (based on raw or normalized log data).
For example, the associated log data in the applicable time range contains “355,466” events 506. In the example of
The LIKE operator (for example, 509) is a standard SQL operator allowing a search with parts of words to find desired sentences (for example, if “use” is searched for in a document, the search may return “use,” “user,” and “confused.” In some implementations, operators other than LIKE can also be used.
After the three filters 508, 510, and 512 are applied to the raw data, 72 events (514) remain for analysis. Event types 516 pertaining to the 72 remaining events (514) are available for selection by an analyst. As stated above, the event types are based on normalized log data. The other event types are considered “Null” fields and are unique for the filter criteria.
At this point, an analyst can choose to select a Null file bubble to add a filter to the filter criteria (here “Event Code 72(1)” 518). As noted, this filter is based on normalized log data. As this value is unique for the current filter criteria, the filter results of 72 events and the same distribution represented by the bubbles will not change.
On selection, bubble 518 expands and is places approximately in the middle of the bubblegram 504. In some implementations, hovering over the expanded bubble 518 can display a pop-up dialog 610 that can provide additional information related to the event type associated with bubble 518 (here, “Event Source Type: ABAP; Count 72”).
At this point, an analyst can choose to select another “gray” bubble (for example, “Event, Severity Code 72(1)” 610. Note that this filter will also be based on normalized log data. Upon selection, another filter will be added to the existing filter list.
Accordingly, some of the filters can be deleted for efficiency reasons. In some implementations, various GUI mechanisms can be provided to indicate and select a specific filter for deletion. Here, an analyst chooses to delete the first three filters (508, 510, and 512). Referring to
By deleting filters 508, 510, and 512, filter criteria based on the raw data was semantically translated into filter criteria based on the normalized data. As seen here, the sematic attributes “Event Code” and “Event, Severity Code” and values “R4F” and “ERROR” for these filters are not obvious for a user that does not have expertise in ETD. Even less obvious for a non-expert of ETD is the finding that the “Fail write file” action can be found over values “R4F” and “ERROR” of sematic attributes “Event Code” and “Event, Severity Code.” The use of the bubblegram permits this understanding and definition of filters for an ETD pattern.
While it is possible to keep filters based on raw data (that is, same results occur (as previously explained)), SQL queries based on normalized data are many times faster (for example, 8 times faster) to execute than those performed over raw data. This is because in a free text search (over raw data), the entire amount of data is touched (for example, each complete log entry). AS a result, entries in each raw data log is parsed and interpreted (for example, using the LIKE operator). This processing is very time, processor, and memory intensive.
In contrast, SQL SELECTS on normalized data are usually performed with an “=” operator, where comparisons are made using already-parsed and interpreted data elements. To complete such queries, only content of particular fields-of-interest need be processed. As a result, time, processor, and memory usage is less in comparison to executing queries over raw data. As a result, free text searching is generally only reasonable using raw data and SQL queries used by ETD patterns are only reasonable using normalized data. Using ETD patterns with filters based on normalized log data can, for example, increase database processing speed, alert generation speed, and result in significantly less processing resources to execute SQL queries.
As illustrated in
At 1002, a free text search is initiated using search terms entered into a free text search GUI to search log data for the free text. Search results are received based on the free text search. From 1002, method 1000 proceeds to 1004.
At 1004, search results are analyzed to determine if they correspond to user expectations. For example, a result is found or results are found within an expected time frame. If it is determined that the search results do not correspond to user expectations, method 1000 proceeds to 1006. If it is determined that the search results correspond to user expectations, method 1000 proceeds to 1008.
At 1006, the search terms are refined. From 1006, method 1000 proceeds to 1007.
At 1007, the free text search is reinitiated using the refined search terms. From 1007, method 1000 proceeds back to 1004.
At 1008, a set of at least one or more search results (events) is selected from the search results (for example, listed log entries) that should be identified by a completed ETD pattern. The user indicates that the selected set of search results should be processed (for example, by selecting a GUI button). From 1008, method 1000 proceeds to 1010.
At 1010, once the selected search results are indicated for processing, the GUI switches to a rendered forensic lab application with rendered event filters and a bubblegram view (for example,
At 1012, a new event filter containing a distinct value for an event type is added from the bubblegram to a path (or to a pre-existing path) by selecting a bubblegram bubble (for example, bubble 518 in
At 1014, the user analyzes a search result to determine whether the search result has changed. The search result includes the count of events corresponding to the selected filters (for example, 608 in
At 1016 on
Note that case B can only happen, if a filter with non-distinct value was added to the path. It is unusual, but possible, that no bubbles with distinct values exist in the bubblegram. In this case, a compromise should be made and some other filter from the bubblegram should be added that changes the results as little as possible. The final goal is to complete an ETD pattern having a set of filters such that the search results cover events defined in 1008 and contain as few-as-possible other events (logs). From 1016, method 1000 proceeds back to 1012 to
At 1018, one raw-data-based filter is removed from the path. From 1018, method 1000 proceeds to 1020.
At 1020, the user analyzes the search result to determine if the search result has changed. If it is determined that the search result has changed, method 1000 proceeds to 1022. If it is determined that the search result has not changed, method 1000 proceeds to 1024. From 1020, method 1000 proceeds to 1022.
At 1022, the user performs an undo on the prior removal of the filter based on raw data (at 1018). From 1022, method 1000 proceeds back to 1012 on
At 1024, the user determines if another raw-data-based filter exists. If it is determined that another raw-data-based filter exists, method 1000 proceeds back to 1018. If it is determined that another raw-data-based filter does not exist, method 1000 proceeds to 1026.
At 1026, the user sets a relative ETD pattern time range. For example, the user can change the existing static time range to a dynamic time range (for example, “Last Hour,” “Last Minute,” and “last10Minutes”). By switching to the forensic lab application in 1010, the time stamps from the particular selected events were used. These timestamps build a time range covering all events selected in 1008 (that is, past events). A desired ETD pattern should include events that will happen in the future. Therefore a relative time range should be set that will cover the recent events. After adjusting the time range on the ETD pattern from an absolute time range to relative time range, the ETD pattern will find most recent data similar to that set in 1008 (due to the same search criteria). After 1026, method 1000 proceeds to 1028.
At 1028, the user completes an ETD pattern for use based on the relative ETD pattern time range. The ETD pattern defines how to present the search results selected by filters from a path (for example, the search results are grouped by users or systems). The ETD pattern also defines on what condition an alert is created (for example, if after grouping the search results by systems there are two groups (corresponding to two systems), then two alerts are created (one for each system). In addition, some metadata attributes are set for an ETD pattern to complete it (for example, a description of the ETD pattern, name, and security impact). After 1028, method 1000 stops.
The computer 1102 can serve in a role as a client, network component, a server, a database or other persistency, or any other component (or a combination of roles) of a computer system for performing the subject matter described in the instant disclosure. The illustrated computer 1102 is communicably coupled with a network 1130. In some implementations, one or more components of the computer 1102 may be configured to operate within environments, including cloud-computing-based, local, global, or other environment (or a combination of environments).
At a high level, the computer 1102 is an electronic computing device operable to receive, transmit, process, store, or manage data and information associated with the described subject matter. According to some implementations, the computer 1102 may also include or be communicably coupled with an application server, e-mail server, web server, caching server, streaming data server, or other server (or a combination of servers).
The computer 1102 can receive requests over network 1130 from a client application (for example, executing on another computer 1102) and respond to the received requests by processing the received requests using an appropriate software application(s). In addition, requests may also be sent to the computer 1102 from internal users (for example, from a command console or by other appropriate access method), external or third-parties, other automated applications, as well as any other appropriate entities, individuals, systems, or computers.
Each of the components of the computer 1102 can communicate using a system bus 1103. In some implementations, any or all of the components of the computer 1102, hardware or software (or a combination of both hardware and software), may interface with each other or the interface 1104 (or a combination of both), over the system bus 1103 using an application programming interface (API) 1112 or a service layer 1113 (or a combination of the API 1112 and service layer 1113). The API 1112 may include specifications for routines, data structures, and object classes. The API 1112 may be either computer-language independent or dependent and refer to a complete interface, a single function, or even a set of APIs. The service layer 1113 provides software services to the computer 1102 or other components (whether or not illustrated) that are communicably coupled to the computer 1102. The functionality of the computer 1102 may be accessible for all service consumers using this service layer. Software services, such as those provided by the service layer 1113, provide reusable, defined functionalities through a defined interface. For example, the interface may be software written in JAVA, C++, or other suitable language providing data in extensible markup language (XML) format or other suitable format. While illustrated as an integrated component of the computer 1102, alternative implementations may illustrate the API 1112 or the service layer 1113 as stand-alone components in relation to other components of the computer 1102 or other components (whether or not illustrated) that are communicably coupled to the computer 1102. Moreover, any or all parts of the API 1112 or the service layer 1113 may be implemented as child or sub-modules of another software module, enterprise application, or hardware module without departing from the scope of this disclosure.
The computer 1102 includes an interface 1104. Although illustrated as a single interface 1104 in
The computer 1102 includes a processor 1105. Although illustrated as a single processor 1105 in
The computer 1102 also includes a database 1106 that can hold data for the computer 1102 or other components (or a combination of both) that can be connected to the network 1130 (whether illustrated or not). For example, database 1106 can be an in-memory, conventional, or other type of database storing data consistent with this disclosure. In some implementations, database 1106 can be a combination of two or more different database types (for example, a hybrid in-memory and conventional database) according to particular needs, desires, or particular implementations of the computer 1102 and the described functionality. Although illustrated as a single database 1106 in
The computer 1102 also includes a memory 1107 that can hold data for the computer 1102 or other components (or a combination of both) that can be connected to the network 1130 (whether illustrated or not). For example, memory 1107 can be random access memory (RAM), read-only memory (ROM), optical, magnetic, and the like, storing data consistent with this disclosure. In some implementations, memory 1107 can be a combination of two or more different types of memory (for example, a combination of RAM and magnetic storage) according to particular needs, desires, or particular implementations of the computer 1102 and the described functionality. Although illustrated as a single memory 1107 in
The application 1108 is an algorithmic software engine providing functionality according to particular needs, desires, or particular implementations of the computer 1102, particularly with respect to functionality described in this disclosure. For example, application 1108 can serve as one or more components, modules, or applications. Further, although illustrated as a single application 1108, the application 1108 may be implemented as multiple applications 1108 on the computer 1102. In addition, although illustrated as integral to the computer 1102, in alternative implementations, the application 1108 can be external to the computer 1102.
The computer 1102 can also include a power supply 1114. The power supply 1114 can include a rechargeable or non-rechargeable battery that can be configured to be either user- or non-user-replaceable. In some implementations, the power supply 1114 can include power-conversion or management circuits (including recharging, standby, or other power management functionality). In some implementations, the power-supply 1114 can include a power plug to allow the computer 1102 to be plugged into a wall socket or other power source to, for example, power the computer 1102 or recharge a rechargeable battery.
There may be any number of computers 1102 associated with, or external to, a computer system containing computer 1102, each computer 1102 communicating over network 1130. Further, the term “client,” “user,” and other appropriate terminology may be used interchangeably, as appropriate, without departing from the scope of this disclosure. Moreover, this disclosure contemplates that many users may use one computer 1102, or that one user may use multiple computers 1102.
Described implementations of the subject matter can include one or more features, alone or in combination.
For example, in a first implementation, a computer-implemented method, comprising: receiving search results from an initiated free text search of log data from one or more logs, the free text performed using search terms entered into a free text search graphical user interface; selecting a set of at least one search result from the search results containing an event desired to be identified in a completed enterprise threat detection (ETD) pattern; rendering a forensic lab application to complete an ETD pattern; adding an event filter for an event type based on normalized log data to a path; setting a relative ETD pattern time range; and completing an ETD pattern based on the added event filter.
The foregoing and other described implementations can each, optionally, include one or more of the following features:
A first feature, combinable with any of the following features, further comprising: analyzing the search results of the free text search to determine that the search result corresponds to a user expectation; refining the search terms; and reinitiating the free text search using the refined search terms.
A second feature, combinable with any of the previous or following features, wherein the event filter based on normalized log data is added to the path by selecting a bubble corresponding to the event type in a bubblegram view rendered in the forensic lab application.
A third feature, combinable with any of the previous or following features, further comprising analyzing a search result based on the added event filter to determine whether there has been a change in the search result value.
A fourth feature, combinable with any of the previous or following features, further comprising removing previously-added filters from the path upon determining that there has been a change in the search result value.
A fifth feature, combinable with any of the previous or following features, further comprising: removing one raw-data-based filter from the path upon determining that there has not been a change in the search result value; and analyzing the search result based on the added event filter to determine whether there has been a change in the search result value.
A sixth feature, combinable with any of the previous or following features, further comprising: determining if another raw-data-based filter exists upon determining that there has not been a change in the search result value; or undoing the removal of the one raw-data-based filter from the path upon determining that there has been a change in the search result value.
In a second implementation, a non-transitory, computer-readable medium storing one or more instructions executable by a computer system to perform operations comprising: receiving search results from an initiated free text search of log data from one or more logs, the free text performed using search terms entered into a free text search graphical user interface; selecting a set of at least one search result from the search results containing an event desired to be identified in a completed enterprise threat detection (ETD) pattern; rendering a forensic lab application to complete an ETD pattern; adding an event filter for an event type based on normalized log data to a path; setting a relative ETD pattern time range; and completing an ETD pattern based on the added event filter.
The foregoing and other described implementations can each, optionally, include one or more of the following features:
A first feature, combinable with any of the following features, further comprising one or more instructions to: analyze the search results of the free text search to determine that the search result corresponds to a user expectation; refine the search terms; and reinitiate the free text search using the refined search terms.
A second feature, combinable with any of the previous or following features, wherein the event filter based on normalized log data is added to the path by selecting a bubble corresponding to the event type in a bubblegram view rendered in the forensic lab application.
A third feature, combinable with any of the previous or following features, further comprising one or more instructions to analyze a search result based on the added event filter to determine whether there has been a change in the search result value.
A fourth feature, combinable with any of the previous or following features, further comprising one or more instructions to remove previously-added filters from the path upon determining that there has been a change in the search result value.
A fifth feature, combinable with any of the previous or following features, further comprising one or more instructions to: remove one raw-data-based filter from the path upon determining that there has not been a change in the search result value; and analyze the search result based on the added event filter to determine whether there has been a change in the search result value.
A sixth feature, combinable with any of the previous or following features, further comprising one or more instructions to: determine if another raw-data-based filter exists upon determining that there has not been a change in the search result value; or undo the removal of the one raw-data-based filter from the path upon determining that there has been a change in the search result value.
In a third implementation, a computer-implemented system, comprising: a computer memory; and a hardware processor interoperably coupled with the computer memory and configured to perform operations comprising: receiving search results from an initiated free text search of log data from one or more logs, the free text performed using search terms entered into a free text search graphical user interface; selecting a set of at least one search result from the search results containing an event desired to be identified in a completed enterprise threat detection (ETD) pattern; rendering a forensic lab application to complete an ETD pattern; adding an event filter for an event type based on normalized log data to a path; setting a relative ETD pattern time range; and completing an ETD pattern based on the added event filter.
The foregoing and other described implementations can each, optionally, include one or more of the following features:
A first feature, combinable with any of the following features, further configured to: analyze the search results of the free text search to determine that the search result corresponds to a user expectation; refine the search terms; and reinitiate the free text search using the refined search terms.
A second feature, combinable with any of the previous or following features, wherein the event filter based on normalized log data is added to the path by selecting a bubble corresponding to the event type in a bubblegram view rendered in the forensic lab application.
A third feature, combinable with any of the previous or following features, further configured to analyze a search result based on the added event filter to determine whether there has been a change in the search result value.
A fourth feature, combinable with any of the previous or following features, further configured to remove previously-added filters from the path upon determining that there has been a change in the search result value.
A fifth feature, combinable with any of the previous or following features, further configured to: remove one raw-data-based filter from the path upon determining that there has not been a change in the search result value; and analyze the search result based on the added event filter to determine whether there has been a change in the search result value.
A sixth feature, combinable with any of the previous or following features, further configured to: determine if another raw-data-based filter exists upon determining that there has not been a change in the search result value; or undo the removal of the one raw-data-based filter from the path upon determining that there has been a change in the search result value.
Implementations of the subject matter and the functional operations described in this specification can be implemented in digital electronic circuitry, in tangibly embodied computer software or firmware, in computer hardware, including the structures disclosed in this specification and their structural equivalents, or in combinations of one or more of them. Software implementations of the described subject matter can be implemented as one or more computer programs, that is, one or more modules of computer program instructions encoded on a tangible, non-transitory, computer-readable computer-storage medium for execution by, or to control the operation of, data processing apparatus. Alternatively, or additionally, the program instructions can be encoded in/on an artificially generated propagated signal, for example, a machine-generated electrical, optical, or electromagnetic signal that is generated to encode information for transmission to suitable receiver apparatus for execution by a data processing apparatus. The computer-storage medium can be a machine-readable storage device, a machine-readable storage substrate, a random or serial access memory device, or a combination of computer-storage mediums.
The term “real-time,” “real time,” “realtime,” “real (fast) time (RFT),” “near(ly) real-time (NRT),” “quasi real-time,” or similar terms (as understood by one of ordinary skill in the art), means that an action and a response are temporally proximate such that an individual perceives the action and the response occurring substantially simultaneously. For example, the time difference for a response to display (or for an initiation of a display) of data following the individual's action to access the data may be less than 1 ms, less than 1 sec., or less than 5 secs. While the requested data need not be displayed (or initiated for display) instantaneously, it is displayed (or initiated for display) without any intentional delay, taking into account processing limitations of a described computing system and time required to, for example, gather, accurately measure, analyze, process, store, or transmit the data.
The terms “data processing apparatus,” “computer,” or “electronic computer device” (or equivalent as understood by one of ordinary skill in the art) refer to data processing hardware and encompass all kinds of apparatus, devices, and machines for processing data, including by way of example, a programmable processor, a computer, or multiple processors or computers. The apparatus can also be, or further include special purpose logic circuitry, for example, a central processing unit (CPU), an FPGA (field programmable gate array), or an ASIC (application-specific integrated circuit). In some implementations, the data processing apparatus or special purpose logic circuitry (or a combination of the data processing apparatus or special purpose logic circuitry) may be hardware- or software-based (or a combination of both hardware- and software-based). The apparatus can optionally include code that creates an execution environment for computer programs, for example, code that constitutes processor firmware, a protocol stack, a database management system, an operating system, or a combination of execution environments. The present disclosure contemplates the use of data processing apparatuses with or without conventional operating systems, for example LINUX, UNIX, WINDOWS, MAC OS, ANDROID, IOS, or any other suitable conventional operating system.
A computer program, which may also be referred to or described as a program, software, a software application, a module, a software module, a script, or code can be written in any form of programming language, including compiled or interpreted languages, or declarative or procedural languages, and it can be deployed in any form, including as a stand-alone program or as a module, component, subroutine, or other unit suitable for use in a computing environment. A computer program may, but need not, correspond to a file in a file system. A program can be stored in a portion of a file that holds other programs or data, for example, one or more scripts stored in a markup language document, in a single file dedicated to the program in question, or in multiple coordinated files, for example, files that store one or more modules, sub-programs, or portions of code. A computer program can be deployed to be executed on one computer or on multiple computers that are located at one site or distributed across multiple sites and interconnected by a communication network. While portions of the programs illustrated in the various figures are shown as individual modules that implement the various features and functionality through various objects, methods, or other processes, the programs may instead include a number of sub-modules, third-party services, components, libraries, and such, as appropriate. Conversely, the features and functionality of various components can be combined into single components, as appropriate. Thresholds used to make computational determinations can be statically, dynamically, or both statically and dynamically determined.
The methods, processes, or logic flows described in this specification can be performed by one or more programmable computers executing one or more computer programs to perform functions by operating on input data and generating output. The methods, processes, or logic flows can also be performed by, and apparatus can also be implemented as, special purpose logic circuitry, for example, a CPU, an FPGA, or an ASIC.
Computers suitable for the execution of a computer program can be based on general or special purpose microprocessors, both, or any other kind of CPU. Generally, a CPU will receive instructions and data from a read-only memory (ROM) or a random access memory (RAM), or both. The essential elements of a computer are a CPU, for performing or executing instructions, and one or more memory devices for storing instructions and data. Generally, a computer will also include, or be operatively coupled to, receive data from or transfer data to, or both, one or more mass storage devices for storing data, for example, magnetic, magneto-optical disks, or optical disks. However, a computer need not have such devices. Moreover, a computer can be embedded in another device, for example, a mobile telephone, a personal digital assistant (PDA), a mobile audio or video player, a game console, a global positioning system (GPS) receiver, or a portable storage device, for example, a universal serial bus (USB) flash drive, to name just a few.
Computer-readable media (transitory or non-transitory, as appropriate) suitable for storing computer program instructions and data includes all forms of non-volatile memory, media and memory devices, including by way of example semiconductor memory devices, for example, erasable programmable read-only memory (EPROM), electrically erasable programmable read-only memory (EEPROM), and flash memory devices; magnetic disks, for example, internal hard disks or removable disks; magneto-optical disks; and CD-ROM, DVD+/−R, DVD-RAM, and DVD-ROM disks. The memory may store various objects or data, including caches, classes, frameworks, applications, backup data, jobs, web pages, web page templates, database tables, repositories storing dynamic information, and any other appropriate information including any parameters, variables, algorithms, instructions, rules, constraints, or references thereto. Additionally, the memory may include any other appropriate data, such as logs, policies, security or access data, reporting files, as well as others. The processor and the memory can be supplemented by, or incorporated in, special purpose logic circuitry.
To provide for interaction with a user, implementations of the subject matter described in this specification can be implemented on a computer having a display device, for example, a CRT (cathode ray tube), LCD (liquid crystal display), LED (Light Emitting Diode), or plasma monitor, for displaying information to the user and a keyboard and a pointing device, for example, a mouse, trackball, or trackpad by which the user can provide input to the computer. Input may also be provided to the computer using a touchscreen, such as a tablet computer surface with pressure sensitivity, a multi-touch screen using capacitive or electric sensing, or other type of touchscreen. Other kinds of devices can be used to provide for interaction with a user as well; for example, feedback provided to the user can be any form of sensory feedback, for example, visual feedback, auditory feedback, or tactile feedback; and input from the user can be received in any form, including acoustic, speech, or tactile input. In addition, a computer can interact with a user by sending documents to and receiving documents from a device that is used by the user; for example, by sending web pages to a web browser on a user's client device in response to requests received from the web browser.
The term “graphical user interface,” or “GUI,” may be used in the singular or the plural to describe one or more graphical user interfaces and each of the displays of a particular graphical user interface. Therefore, a GUI may represent any graphical user interface, including but not limited to, a web browser, a touch screen, or a command line interface (CLI) that processes information and efficiently presents the information results to the user. In general, a GUI may include a plurality of user interface (UI) elements, some or all associated with a web browser, such as interactive fields, pull-down lists, and buttons. These and other UI elements may be related to or represent the functions of the web browser.
Implementations of the subject matter described in this specification can be implemented in a computing system that includes a back-end component, for example, as a data server, or that includes a middleware component, for example, an application server, or that includes a front-end component, for example, a client computer having a graphical user interface or a Web browser through which a user can interact with an implementation of the subject matter described in this specification, or any combination of one or more such back-end, middleware, or front-end components. The components of the system can be interconnected by any form or medium of wireline or wireless digital data communication (or a combination of data communication), for example, a communication network. Examples of communication networks include a local area network (LAN), a radio access network (RAN), a metropolitan area network (MAN), a wide area network (WAN), Worldwide Interoperability for Microwave Access (WIMAX), a wireless local area network (WLAN) using, for example, 802.11 a/b/g/n or 802.20 (or a combination of 802.11x and 802.20 or other protocols consistent with this disclosure), all or a portion of the Internet, or any other communication system or systems at one or more locations (or a combination of communication networks). The network may communicate with, for example, Internet Protocol (IP) packets, Frame Relay frames, Asynchronous Transfer Mode (ATM) cells, voice, video, data, or other suitable information (or a combination of communication types) between network addresses.
The computing system can include clients and servers. A client and server are generally remote from each other and typically interact through a communication network. The relationship of client and server arises by virtue of computer programs running on the respective computers and having a client-server relationship to each other.
While this specification contains many specific implementation details, these should not be construed as limitations on the scope of any invention or on the scope of what may be claimed, but rather as descriptions of features that may be specific to particular implementations of particular inventions. Certain features that are described in this specification in the context of separate implementations can also be implemented, in combination, in a single implementation. Conversely, various features that are described in the context of a single implementation can also be implemented in multiple implementations, separately, or in any suitable sub-combination. Moreover, although previously described features may be described as acting in certain combinations and even initially claimed as such, one or more features from a claimed combination can, in some cases, be excised from the combination, and the claimed combination may be directed to a sub-combination or variation of a sub-combination.
Particular implementations of the subject matter have been described. Other implementations, alterations, and permutations of the described implementations are within the scope of the following claims as will be apparent to those skilled in the art. While operations are depicted in the drawings or claims in a particular order, this should not be understood as requiring that such operations be performed in the particular order shown or in sequential order, or that all illustrated operations be performed (some operations may be considered optional), to achieve desirable results. In certain circumstances, multitasking or parallel processing (or a combination of multitasking and parallel processing) may be advantageous and performed as deemed appropriate.
Moreover, the separation or integration of various system modules and components in the previously described implementations should not be understood as requiring such separation or integration in all implementations, and it should be understood that the described program components and systems can generally be integrated together in a single software product or packaged into multiple software products.
Accordingly, the previously described example implementations do not define or constrain this disclosure. Other changes, substitutions, and alterations are also possible without departing from the spirit and scope of this disclosure.
Furthermore, any claimed implementation is considered to be applicable to at least a computer-implemented method; a non-transitory, computer-readable medium storing computer-readable instructions to perform the computer-implemented method; and a computer system comprising a computer memory interoperably coupled with a hardware processor configured to perform the computer-implemented method or the instructions stored on the non-transitory, computer-readable medium.
Number | Name | Date | Kind |
---|---|---|---|
5440726 | Fuchs | Aug 1995 | A |
5960170 | Chen | Sep 1999 | A |
6173418 | Fujino et al. | Jan 2001 | B1 |
6629106 | Narayanaswamy | Sep 2003 | B1 |
6779001 | Kanai et al. | Aug 2004 | B1 |
7376969 | Njemanze | May 2008 | B1 |
7380205 | Bezrukov et al. | May 2008 | B2 |
7441197 | Tschiegg et al. | Oct 2008 | B2 |
7457792 | Weigt et al. | Nov 2008 | B2 |
7457793 | Weigt et al. | Nov 2008 | B2 |
7457794 | Weigt et al. | Nov 2008 | B2 |
7545969 | Bennett | Jun 2009 | B2 |
7624092 | Lieske et al. | Nov 2009 | B2 |
7627544 | Chkodrov | Dec 2009 | B2 |
7756808 | Weigt et al. | Jul 2010 | B2 |
7756809 | Weigt et al. | Jul 2010 | B2 |
7761396 | Weigt et al. | Jul 2010 | B2 |
7783723 | Peng et al. | Aug 2010 | B2 |
7788718 | Fei | Aug 2010 | B1 |
7872982 | Atkins | Jan 2011 | B2 |
7908660 | Bahl | Mar 2011 | B2 |
7934257 | Kienzle | Apr 2011 | B1 |
7961633 | Shankar | Jun 2011 | B2 |
7971209 | Eberlein et al. | Jun 2011 | B2 |
8051034 | Mehta et al. | Nov 2011 | B2 |
8091117 | Williams | Jan 2012 | B2 |
8474047 | Adelstein | Jun 2013 | B2 |
8484726 | Sutton | Jul 2013 | B1 |
8554907 | Chen et al. | Oct 2013 | B1 |
8661103 | Mehta et al. | Feb 2014 | B2 |
8775671 | Rodeck et al. | Jul 2014 | B2 |
8892454 | Rabetge et al. | Nov 2014 | B2 |
8954602 | Seifert et al. | Feb 2015 | B2 |
8973147 | Pearcy | Mar 2015 | B2 |
9037678 | Mehta et al. | May 2015 | B2 |
9075633 | Nos | Jul 2015 | B2 |
9106697 | Capalik et al. | Aug 2015 | B2 |
9116906 | Nos et al. | Aug 2015 | B2 |
9148488 | Rabetge et al. | Sep 2015 | B2 |
9170951 | He | Oct 2015 | B1 |
9251011 | Meier et al. | Feb 2016 | B2 |
9262519 | Saurabh | Feb 2016 | B1 |
9304978 | Bezrukov et al. | Apr 2016 | B2 |
9313421 | Deshpande | Apr 2016 | B2 |
9336385 | Spencer | May 2016 | B1 |
9348665 | Storz et al. | May 2016 | B2 |
9383934 | Likacs | Jul 2016 | B1 |
9419989 | Harris | Aug 2016 | B2 |
9524389 | Roth | Dec 2016 | B1 |
9619984 | Donovan | Apr 2017 | B2 |
9690931 | Anantharaja et al. | Jun 2017 | B1 |
9779147 | Sherman et al. | Oct 2017 | B1 |
9779150 | Sherman | Oct 2017 | B1 |
9843596 | Averbuch | Dec 2017 | B1 |
9979741 | Fuhrman | May 2018 | B2 |
10001389 | Das et al. | Jun 2018 | B1 |
10003605 | Muddu et al. | Jun 2018 | B2 |
10079842 | Brandwine et al. | Sep 2018 | B1 |
10102379 | Seifert et al. | Oct 2018 | B1 |
10140447 | Rahaman | Nov 2018 | B2 |
10148675 | Brandwine et al. | Dec 2018 | B1 |
10681064 | Peng et al. | Jun 2020 | B2 |
10986111 | Peng et al. | Apr 2021 | B2 |
20020070953 | Barg | Jun 2002 | A1 |
20030074471 | Anderson | Apr 2003 | A1 |
20030115484 | Mariconi et al. | Jun 2003 | A1 |
20030217137 | Roese | Nov 2003 | A1 |
20040015481 | Zinda | Jan 2004 | A1 |
20040044912 | Connary | Mar 2004 | A1 |
20040078490 | Anderson | Apr 2004 | A1 |
20040093513 | Cantrell | May 2004 | A1 |
20050027667 | Kroll et al. | Feb 2005 | A1 |
20060037075 | Frattura | Feb 2006 | A1 |
20060059115 | Gulfleisch et al. | Mar 2006 | A1 |
20060106847 | Eckardt et al. | May 2006 | A1 |
20060161816 | Gula et al. | Jul 2006 | A1 |
20060218140 | Whitney | Sep 2006 | A1 |
20060253907 | McConnell | Nov 2006 | A1 |
20070067438 | Goranson et al. | Mar 2007 | A1 |
20070073519 | Long | Mar 2007 | A1 |
20070100905 | Masters et al. | May 2007 | A1 |
20070115998 | McElligott | May 2007 | A1 |
20070136437 | Shankar et al. | Jun 2007 | A1 |
20070150596 | Miller et al. | Jun 2007 | A1 |
20070183389 | Clee | Aug 2007 | A1 |
20070186284 | McConnell | Aug 2007 | A1 |
20070266387 | Henmi | Nov 2007 | A1 |
20070283192 | Shevchenko | Dec 2007 | A1 |
20070300296 | Kudla | Dec 2007 | A1 |
20080033966 | Wahl | Feb 2008 | A1 |
20080034425 | Overcash et al. | Feb 2008 | A1 |
20080080384 | Atkins | Apr 2008 | A1 |
20080091681 | Dwivedi | Apr 2008 | A1 |
20080163085 | Subbu et al. | Jul 2008 | A1 |
20080288889 | Hunt et al. | Nov 2008 | A1 |
20080295173 | Tsvetanov | Nov 2008 | A1 |
20080320552 | Kumar | Dec 2008 | A1 |
20090044277 | Aaron et al. | Feb 2009 | A1 |
20090049518 | Roman | Feb 2009 | A1 |
20090288164 | Adelstein | Nov 2009 | A1 |
20090293046 | Cheriton | Nov 2009 | A1 |
20090300045 | Chaudhry et al. | Dec 2009 | A1 |
20090312026 | Parameswar | Dec 2009 | A1 |
20100011031 | Huang | Jan 2010 | A1 |
20100114832 | Lillibridge | May 2010 | A1 |
20100180325 | Golobay | Jul 2010 | A1 |
20110098928 | Hoffman et al. | Apr 2011 | A1 |
20110213741 | Shama | Sep 2011 | A1 |
20110277034 | Hanson | Nov 2011 | A1 |
20110320816 | Yao | Dec 2011 | A1 |
20120005542 | Petersen | Jan 2012 | A1 |
20120158653 | Shaffer et al. | Jun 2012 | A1 |
20120167161 | Kim et al. | Jun 2012 | A1 |
20120191660 | Hoog | Jul 2012 | A1 |
20120210434 | Curtis et al. | Aug 2012 | A1 |
20120271748 | DiSalvo | Oct 2012 | A1 |
20120271790 | Lappas et al. | Oct 2012 | A1 |
20120317078 | Zhou et al. | Dec 2012 | A1 |
20130086023 | Tsukamoto et al. | Apr 2013 | A1 |
20130106830 | de Loera | May 2013 | A1 |
20130198840 | Drissi et al. | Aug 2013 | A1 |
20130212709 | Tucker | Aug 2013 | A1 |
20130262311 | Buhrmann | Oct 2013 | A1 |
20130298243 | Kumar et al. | Nov 2013 | A1 |
20130304665 | Rodeck et al. | Nov 2013 | A1 |
20130304666 | Rodeck et al. | Nov 2013 | A1 |
20130305369 | Karta | Nov 2013 | A1 |
20130326079 | Seifert et al. | Dec 2013 | A1 |
20130347111 | Karta | Dec 2013 | A1 |
20140047413 | Sheive et al. | Feb 2014 | A1 |
20140201836 | Amsler | Jul 2014 | A1 |
20140223283 | Hancock | Aug 2014 | A1 |
20140244623 | King | Aug 2014 | A1 |
20140317681 | Shende | Oct 2014 | A1 |
20150007325 | Eliseev | Jan 2015 | A1 |
20150067880 | Ward | Mar 2015 | A1 |
20150073868 | Garman | Mar 2015 | A1 |
20150106867 | Liang | Apr 2015 | A1 |
20150109305 | Black | Apr 2015 | A1 |
20150143521 | Eliseev | May 2015 | A1 |
20150154524 | Borodow | Jun 2015 | A1 |
20150180891 | Seward | Jun 2015 | A1 |
20150215329 | Singla | Jul 2015 | A1 |
20150237065 | Roytman | Aug 2015 | A1 |
20150264011 | Liang | Sep 2015 | A1 |
20150278371 | Anand | Oct 2015 | A1 |
20150281278 | Gooding | Oct 2015 | A1 |
20150310070 | Stefik et al. | Oct 2015 | A1 |
20150319185 | Kirti | Nov 2015 | A1 |
20150341389 | Kurakami | Nov 2015 | A1 |
20150347751 | Card et al. | Dec 2015 | A1 |
20150355957 | Steiner | Dec 2015 | A1 |
20150358344 | Mumcuoglu | Dec 2015 | A1 |
20150381646 | Lin | Dec 2015 | A1 |
20160057166 | Chesla | Feb 2016 | A1 |
20160057167 | Bach et al. | Feb 2016 | A1 |
20160065594 | Srivastava et al. | Mar 2016 | A1 |
20160092535 | Kuchibhotla et al. | Mar 2016 | A1 |
20160127391 | Kobres | May 2016 | A1 |
20160164891 | Satish et al. | Jun 2016 | A1 |
20160202893 | Mustonen et al. | Jul 2016 | A1 |
20160226905 | Baikalov et al. | Aug 2016 | A1 |
20160248798 | Cabrera | Aug 2016 | A1 |
20160291982 | Mizrahi | Oct 2016 | A1 |
20160292061 | Marron | Oct 2016 | A1 |
20160337384 | Jansson | Nov 2016 | A1 |
20160359886 | Yadav et al. | Dec 2016 | A1 |
20160364315 | Lee | Dec 2016 | A1 |
20160364571 | Lee | Dec 2016 | A1 |
20160364740 | Parker | Dec 2016 | A1 |
20160366174 | Cherinin | Dec 2016 | A1 |
20160373476 | Dell'anno et al. | Dec 2016 | A1 |
20160378978 | Singla | Dec 2016 | A1 |
20160381049 | Lakhani | Dec 2016 | A1 |
20170004005 | Elliott | Jan 2017 | A1 |
20170026400 | Adams et al. | Jan 2017 | A1 |
20170031002 | Newton | Feb 2017 | A1 |
20170034023 | Nickolov | Feb 2017 | A1 |
20170070415 | Bell et al. | Mar 2017 | A1 |
20170091008 | Cherbakov | Mar 2017 | A1 |
20170093902 | Roundy et al. | Mar 2017 | A1 |
20170123875 | Craik et al. | May 2017 | A1 |
20170148060 | Showers | May 2017 | A1 |
20170169217 | Rahaman | Jun 2017 | A1 |
20170185273 | Marache-Francisco et al. | Jun 2017 | A1 |
20170185439 | Chiueh et al. | Jun 2017 | A1 |
20170195439 | Dixon et al. | Jul 2017 | A1 |
20170251365 | Burchard | Aug 2017 | A1 |
20170270006 | Kankylas | Sep 2017 | A1 |
20170279837 | Dasgupta | Sep 2017 | A1 |
20170287179 | Tibshirani et al. | Oct 2017 | A1 |
20170302685 | Ladnai et al. | Oct 2017 | A1 |
20170308602 | Raghunathan et al. | Oct 2017 | A1 |
20170310690 | Mestha et al. | Oct 2017 | A1 |
20170316026 | Kanthak et al. | Nov 2017 | A1 |
20170318036 | Movsisyan | Nov 2017 | A1 |
20170322993 | Brodt et al. | Nov 2017 | A1 |
20170324766 | Gonzalez | Nov 2017 | A1 |
20170329972 | Brisebois et al. | Nov 2017 | A1 |
20180027002 | Rodeck et al. | Jan 2018 | A1 |
20180027010 | Pritzkau et al. | Jan 2018 | A1 |
20180059876 | Peng et al. | Mar 2018 | A1 |
20180063167 | Rodeck | Mar 2018 | A1 |
20180091535 | Chrosziel | Mar 2018 | A1 |
20180091536 | Chrosziel et al. | Mar 2018 | A1 |
20180157835 | Nos | Jun 2018 | A1 |
20180173872 | Lam et al. | Jun 2018 | A1 |
20180173873 | Hassforther et al. | Jun 2018 | A1 |
20180176234 | Kunz et al. | Jun 2018 | A1 |
20180176235 | Lam et al. | Jun 2018 | A1 |
20180176238 | Nos et al. | Jun 2018 | A1 |
20180234447 | Mueen | Aug 2018 | A1 |
20190005423 | Pritzkau et al. | Jan 2019 | A1 |
20190007435 | Pritzkau et al. | Jan 2019 | A1 |
20190190927 | Peng et al. | Jun 2019 | A1 |
20190190935 | Peng et al. | Jun 2019 | A1 |
20200195680 | Pritzkau et al. | Jun 2020 | A1 |
Entry |
---|
Office Action issued in U.S. Appl. No. 15/216,046 dated Aug. 21, 2019, 23 pages. |
Non-Final Office Action issued in U.S. Appl. No. 15/639,863 dated Jun. 24, 2019,37 pages. |
Office Action issued in U.S. Appl. No. 15/847,478, dated Aug. 6, 2019, 36 pages. |
U.S. Office Action in related U.S. Appl. No. 15/383,771 dated Jul. 5, 2019, 16 pages. |
Final Office Action issued in U.S. Appl. No. 15/381,567 dated May 22, 2019, 28 pages. |
Non-Final Office Action issued in U.S. Appl. No. 15/216,046 dated Apr. 29, 2019, 23 pages. |
Non-Final Office Action issued in U.S. Appl. No. 15/246,053 dated May 17, 2019, 28 pages. |
Non-Final Office Action issued in U.S. Appl. No. 15/274,569 dated Mar. 22, 2019, 11 pages. |
Final Office Action issued in U.S. Appl. No. 15/370,084 dated Feb. 4, 2019, 16 pages. |
Final Office Action issued in U.S. Appl. No. 15/274,693 dated Feb. 11, 2019, 19 pages. |
Schumacher, “An effective way to bring SAP Security Notes under control,” Virtual Forge GmbH, Feb. 2017, https://blog.virtualforge.com/en/an-effective-way-to-bring-sap-security-notes-under-control, 4 pages. |
U.S. Office Action in related U.S. Appl. No. 15/274,693 dated Jul. 26, 2018, 14 pages. |
U.S. Office Action in related U.S. Appl. No. 15/370,084 dated Aug. 27, 2018, 14 pages. |
U.S. Office Action in related U.S. Appl. No. 15/383,771 dated Aug. 3, 2018, 12 pages. |
U.S. Office Action in related U.S. Appl. No. 15/216,201 dated Mar. 7, 2018, 14 pages. |
U.S. Office Action in related U.S. Appl. No. 15/216,201 dated Jul. 20, 2018, 15 pages. |
U.S. Office Action in related U.S. Appl. No. 15/246,053 dated May 21, 2018, 14 pages. |
U.S. Office Action in related U.S. Appl. No. 15/246,053 dated Sep. 24, 2018, 14 pages. |
U.S. Office Action in related U.S. Appl. No. 15/253,438 dated Sep. 19, 2018, 17 pages. |
U.S. Office Action in related U.S. Appl. No. 15/274,569 dated Apr. 16, 2018, 11 pages. |
U.S. Office Action in related U.S. Appl. No. 15/274,569 dated Nov. 14, 2018, 10 pages. |
U.S. Office Action in related U.S. Appl. No. 15/380,379 dated Jul. 19, 2018, 9 pages. |
U.S. Office Action in related U.S. Appl. No. 15/380,450 dated Aug. 27, 2018, 19 pages. |
U.S. Office Action in related U.S. Appl. No. 15/380,450 dated Jan. 23, 2019, 21 pages. |
U.S. Office Action in related U.S. Appl. No. 15/380,450 dated Nov. 2, 2018, 19 pages. |
U.S. Office Action in related U.S. Appl. No. 15/381,567 dated Nov. 2, 2018, 17 pages. |
U.S. Office Action in related U.S. Appl. No. 15/383,771 dated Jan. 23, 2019, 14 pages. |
Final office action issued in U.S. Appl. No. 15/847,478 dated Nov. 25, 2019, 14 pages. |
Final Office Action issued in U.S. Appl. No. 15/383,771 dated Oct. 31, 2019, 23 pages. |
Non-final office action issued in U.S. Appl. No. 15/381,567 dated Oct. 7, 2019, 17 pages. |
Non-final office action issued in U.S. Appl. No. 15/847,478 dated Jun. 9, 2020, 13 pages. |
Final Office Action issued in U.S. Appl. No. 15/847,478 dated Sep. 21, 2020, 14 pages. |
Non-Final Office Action issued in U.S. Appl. No. 16/741,071 dated Sep. 1, 2020, 32 pages. |
Number | Date | Country | |
---|---|---|---|
20200145444 A1 | May 2020 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 15639907 | Jun 2017 | US |
Child | 16734986 | US |