Embodiments described herein generally relate to cloud computing and in particular to bulk service mapping.
Cloud computing relates to the sharing of computing resources that are generally accessed via the Internet. In particular, the cloud computing infrastructure allows users, such as individuals and/or enterprises, to access a shared pool of computing resources, such as servers, storage devices, networks, applications, and/or other computing based services. By doing so, users are able to access computing resources on demand that are located at remote locations in order to perform a variety computing functions that include storing and/or processing computing data. For enterprise and other organization users, cloud computing provides flexibility in accessing cloud computing resources without accruing up-front costs, such as purchasing network equipment, and investing time in establishing a private network infrastructure. Instead, by utilizing cloud computing resources, users are able redirect their resources to focus on core enterprise functions.
In today's communication networks, examples of cloud computing services a user may utilize include software as a service (SaaS) and platform as a service (PaaS) technologies. SaaS is a delivery model that provides software as a service rather than an end product. Instead of utilizing a local network or individual software installations, software is typically licensed on a subscription basis, hosted on a remote machine, and accessed as needed. For example, users are generally able to access a variety of enterprise and/or information technology (IT) related software via a web browser. PaaS acts an extension of SaaS that goes beyond providing software services by offering customizability and expandability features to meet a user's needs. For example, PaaS can provide a cloud based developmental platform for users to develop, modify, and/or customize applications and/or automating enterprise operations without maintaining network infrastructure and/or allocating computing resources normally associated with these functions.
Within the context of automating enterprise, IT, and/or other organization-related functions (e.g., human resources (HR)), PaaS often provides users an array of tools to implement complex behaviors, such as enterprise rules, scheduled jobs, events, and scripts, to build automated processes and to integrate with third party systems. Although the tools for PaaS generally offer users a rich set of facilities for building automated processes for various enterprise, IT, and/or other organization-related functions, users typically implement custom scripts to perform the automated process. Requiring customized script to build automated processes may pose a challenge when attempting to address abstraction (e.g., providing domain-appropriate building blocks), code reuse (e.g., having defined application program interface (API) semantics), and/or codeless development. As such, continually improving the technology of developmental platforms that simplify the process for a user to design and run automated processes remains valuable in enhancing clouding computing services.
The following presents a simplified summary of the disclosed subject matter in order to provide a basic understanding of some aspects of the subject matter disclosed herein. This summary is not an exhaustive overview of the technology disclosed herein. It is not intended to identify key or critical elements of the invention or to delineate the scope of the invention. Its sole purpose is to present some concepts in a simplified form as a prelude to the more detailed description that is discussed later.
One embodiment includes a system for bulk mapping network services. The system determines a set of service entry point candidates for an organization based, at least in part, on network information received from one or more load balancing servers. The system receives user input including a selection of a set of service entry points selected from the set of service entry point candidates. The system collects network information related to two or more service entry points of the set of service entry point candidates in parallel, wherein the network information includes returned errors associated with a respective service entry point. The system groups the returned errors into one or more category groups based on characteristics of the returned errors. The system displays the one or more category groups on a display device for bulk resolution.
Another embodiment includes a method for bulk mapping network service. The method includes determining a set of service entry point candidates for an organization based, at least in part, on network information received from one or more load balancing servers. The method also includes receiving user input including a selection of a set of service entry points selected from the set of service entry point candidates. The method also includes collecting network information related to two or more service entry points of the set of service entry point candidates in parallel, wherein the network information includes returned errors associated with a respective service entry point. The method also includes grouping the returned errors into one or more category groups based on characteristics of the returned errors. The method also includes displaying the one or more category groups on a display device for bulk resolution.
Another embodiment includes a non-transitory computer readable medium containing instructions that, when executed by a processor causes a programmable device to perform a technique for bulk mapping. The instructions cause the programmable device to determine a set of service entry point candidates for an organization based, at least in part, on network information received from one or more load balancing servers. The instructions also cause the programmable device to receive user input including a selection of a set of service entry points selected from the set of service entry point candidates. The instructions also cause the programmable device to collect network information related to two or more service entry points of the set of service entry point candidates in parallel, wherein the network information includes returned errors associated with a respective service entry point. The instructions also cause the programmable device to group the returned errors into one or more category groups based on characteristics of the returned errors. The instructions also cause the programmable device to display the one or more category groups on a display device for bulk resolution.
For a more complete understanding of this disclosure, reference is now made to the following brief description, taken in connection with the accompanying drawings and detailed description, wherein like reference numerals represent like parts.
In the following description, for purposes of explanation, numerous specific details are set forth in order to provide a thorough understanding of the embodiments disclosed herein. It will be apparent, however, to one skilled in the art that the disclosed embodiments may be practiced without these specific details. In other instances, structure and devices are shown in block diagram form in order to avoid obscuring the disclosed embodiments. References to numbers without subscripts or suffixes are understood to reference all instance of subscripts and suffixes corresponding to the referenced number. Moreover, the language used in this disclosure has been principally selected for readability and instructional purposes, and may not have been selected to delineate or circumscribe the inventive subject matter, resort to the claims being necessary to determine such inventive subject matter. Reference in the specification to “one embodiment” or to “an embodiment” means that a particular feature, structure, or characteristic described in connection with the embodiments is included in at least one embodiment.
The terms “a,” “an,” and “the” are not intended to refer to a singular entity unless explicitly so defined, but include the general class of which a specific example may be used for illustration. The use of the terms “a” or “an” may therefore mean any number that is at least one, including “one,” “one or more,” “at least one,” and “one or more than one.” The term “or” means any of the alternatives and any combination of the alternatives, including all of the alternatives, unless the alternatives are explicitly indicated as mutually exclusive. The phrase “at least one of” when combined with a list of items, means a single item from the list or any combination of items in the list. The phrase does not require all of the listed items unless explicitly so defined.
As used herein, the term “computing system” refers to a single electronic computing device that includes, but is not limited to a single computer, virtual machine, virtual container, host, server, laptop, and/or mobile device or to a plurality of electronic computing devices working together to perform the function described as being performed on or by the computing system.
As used herein, the term “medium” refers to one or more non-transitory physical media that together store the contents described as being stored thereon. Embodiments may include non-volatile secondary storage, read-only memory (ROM), and/or random-access memory (RAM).
As used herein, the term “application” refers to one or more computing modules, programs, processes, workloads, threads and/or a set of computing instructions executed by a computing system. Example embodiments of an application include software modules, software objects, software instances and/or other types of executable code.
Service Management (SM), sometimes referred to as Enterprise Service Management, refers to a field of information technology (IT) administration that ties particular portions of an IT infrastructure (referred to as Configuration Items (CIs)) to a function provided by an enterprise. For example, an enterprise may be divided functionally into a set of services that the enterprise provides internally and externally. In a large enterprise there may be several Enterprise Units (EUs) and each EU may include one or more services. Examples of internal functions that support an enterprise include but are not limited to, human resources and finance. Examples of external functions that support an enterprise include but are not limited to, product delivery, logistics, e-commerce, and customer support. Obviously, some of these functions services have both internal and external components. One goal of SM is to identify what functions make up a service and, in turn, which CIs support those functions within the IT infrastructure. Having such a view may allow prioritization of IT tasks such as trouble shooting, upgrade, and maintenance in a manner that best supports overall goals of the enterprise. The view tying CIs to service functions may, for example, be represented in a visual graph or a mathematical model referred to as a service map or enterprise service map.
As mentioned above, CIs represent portions of an IT infrastructure. A portion of an IT infrastructure may be confined to an individual server, router, data storage unit, or other physical IT component. Alternatively, a portion of an IT infrastructure may be considered as a logical abstraction of a group of physical IT components providing a function, for example, a data base or web service. A data base containing employee records could then be included in a service map for one or more human resources functions of the enterprise (e.g., payroll, benefits, etc.). Similarly, a web service supporting e-commerce could be a CI included in a service map representing a customer facing web sales function.
In
To utilize computing resources within the developmental platform network 110, network operators may choose to configure the data centers 112 using a variety of computing infrastructures. In one embodiment, one or more of the data centers 112 are configured using a multi-tenant cloud architecture such that a single server instance 114, which can also be referred to as an application instance, handles requests and serves multiple customers. In other words, data centers with multi-tenant cloud architecture commingle and store data from multiple customers, where multiple customer instances are assigned to a single server instance 114. In a multi-tenant cloud architecture, the single server instance 114 distinguishes between and segregates data and other information of the various customers. For example, a multi-tenant cloud architecture could assign a particular identifier for each customer in order to identify and segregate the data from each customer. Generally, implementing a multi-tenant cloud architecture suffer drawbacks, such as a failure to single server instance 114 causing outages for all customers allocated to the single server instance 114.
In another embodiment, one or more of the data centers 112 are configured using a multi-instance cloud architecture to provide every customer its own unique customer instance. For example, a multi-instance cloud architecture could provide each customer instance with its own dedicated application server and dedicated database server. In other examples, the multi-instance cloud architecture could deploy a single server instance 114 and/or other combinations of server instances 114, such as one or more dedicated web server instances, one or more dedicated application server instances, and one or more database server instances, for each customer instance. In a multi-instance cloud architecture, multiple customer instances could be installed on a single physical hardware server where each customer instance is allocated certain portions of the physical server resources, such as computing memory, storage, and processing power. By doing so, each customer instance has its own unique software stack that provides the benefit of data isolation, relatively less downtime for customers to access the developmental platform network 110, and customer-driven upgrade schedules. An example of implementing a customer instance within a multi-instance cloud architecture will be discussed in more detail below when describing
To facilitate higher availability of the customer instance 208, the application server instances 210a-210d and database server instances 212a and 212b are allocated to two different data centers 206a and 206b, where one of the data centers 206 acts as a backup data center. In reference to
Although
The load balancers allow for a single unified outward facing enterprise service to be supported by one or more devices, real or logical, in order to distribute service loads. Distribution of service loads helps maintain reliability of the enterprise service under load and allows a measure of redundancy in case of, for example, hardware failure. According to certain aspects, enterprises having multiple existing enterprise services may route these services through one or more load balancers and the load balancers may be used as a source of information for discovering enterprise service entry points. These entry points are properties of a connection to a CI and these entry points may be used as a starting point for service mapping. Common entry point attributes include host/hostname (including real, alias, internet protocol (IP) address, or virtual IP address), port number, service name, and uniform resource locator (URL) of the service. Once entry points are defined, service mapping may start the discovery and mapping process for the enterprise service from the entry point.
A readiness check may be performed as a part of or prior to bulk service mapping to determine whether a particular cloud computing infrastructure and service mapping instance is prepared for bulk service mapping. For example, the readiness check may verify whether load balancers and mid servers have been discovered by a discovery process. The discovery process generally finds (e.g., obtaining network addresses, host information, and/or IP addresses) devices, such as computers, servers, printers, load balancers, and a variety of IP-enabled devices, along with the applications that run on these device, for a cloud computing infrastructure. Data collected by the discovery process may be stored in the CMDB. As an example the readiness check may verify that a discovered MID server is provisioned with the relevant applications, capabilities, and has IP ranges defined. Credentials may be checked to verify that the service mapping instance and other relevant services have certain credentials or permissions. Load balancers and hosts may be checked to verify that there are a sufficient number of load balancers and hosts for bulk service mapping to be performant and that discovery has occurred or refreshed within a certain number of days. Where there are less load balancers and hosts than necessary, bulk mapping may be less useful, for example, as there may not be enough server resources to absorb the load of bulk mapping or enough services to take advantage of bulk mapping useful. Status for services may also be checked. For example, the readiness checks may verify that netflow services have been configured on the router and that cloud discovery is configured on the MID server. According to certain aspects, an infrastructure may be determined to be ready for bulk service mapping if the readiness check determines that there is at least one MID server configured with a service mapping instance over an IP range, that the service mapping instance has at least one active credential, at least three operational load balancer discovered within the last two weeks, at least 100 hosts discovered within the last two weeks, a netflow connector is configured, active, and refreshed in the last day, and cloud discovery is configured, active, and refreshed in the last two weeks. Readiness issues may be surfaced when user interventions may be needed.
After the readiness checks, service entry point candidates may be determined based, at least in part, from the network information received from one or more load balancers. A service mapping instance may connect to a load balancer using standard protocols, such as simple network management protocol (SNMP), secure shell (SSH), etc. to obtain information related to the configuration of the load balancer. As an example, the service mapping instance may query the load balancer via SNMP to discover devices (such as servers, switches, routers, etc.) and host information. Network traffic data may also be used to determine service entry point candidates. For example, the service mapping instance may use network tools, such as netstat or Splunk on Splunk (S.o.S.), to obtain net flow or traffic data for the server the service mapping instance is running on. These network tools may return, for example, transmission control protocol (TCP) connections, routing tables, or other similar network traffic data.
According to certain aspects, service maps may be generated on multiple selected service entry point candidates 504 in parallel. This parallelization may, for example, be performed using multiple threads, processes, or instances, either of the service mapping process or directed by the service mapping process. In certain cases, the service mapping process may process multiple service entry point candidate as background processes and notify the user when the processing of the service entry point candidates are complete.
For a selected service entry point, a discovery process may be run to identify the host and a mapping procedure to map applications running on the host. A service entry point generally represents a point of access for an application CI. For the discovery process, the service mapping instance may check the CMDB to determine if a device hosting the application CI exists. If the device does not exist, the service mapping instance may probe, via the MID server, the host URL or ports to identify the host and update the CMDB. Once a host is found in the CMDB, the service mapping process discovers the applications running on the host. This may be performed in several ways. Two methods of discovery include pattern matching and network monitoring. Pattern matching may include analysis of applications and their configuration files. To perform pattern matching domain specific knowledge may be required to know how to parse configuration files so as to identify the desired information. Network monitoring includes analysis of network traffic between different CIs within a given network. By analyzing how CIs communicate with each other it can be possible to identify dependencies of CIs on each other and make educated guesses to build a service map.
According to certain aspects, the errors may be encountered while attempting to create a service map for one or more service entry point candidates. These errors may be categorized into category groups based on the characteristics of the error across multiple service entry point candidates. For example, authentication errors and user errors may be grouped together in a single category group as the errors are related to credential issues and may be resolved together, for example, by propagating a service mapping user as needed with a particular set of permissions. In certain cases, these categories provide a meaningful flow and indicate a level of impact on services.
Category group cards 702 may also include error status information based on the underlying errors of the group. For example, status information may also indicate that the category group is in discovery. A category group may be in discover while mapping is still being performed or where a rediscover all has been initiated on the group, such as to verify a corrective action after the corrective action was taken. Status information may also include a comment which may be assigned to the group, for example, by the user. Status information may also indicate that active errors are present and that no corrective actions have been taken on those errors. Generally category group cards may display a single status information for the group. For example, if a rediscovery operation is still in progress, the in discovery status information may be displayed in the card even if new active errors have been found.
Errors may be resolved by drilling down into a particular category group card by clicking on the particular category group card. A list of the errors in the group associated with the particular category group card may be displayed. In certain cases, the list of errors may be arranged into subgroups of the category group of errors and each subgroup represented by an error card. The subgroups may be based on a specific type of error of the group of errors. For example, invalid user errors and insufficient permission errors may be grouped into separate subgroups.
Actions may be recommended to resolve certain error cards and these actions may be performed as against a single error, multiple selected errors or error cards, or all error cards. Recommended actions may include internal actions, which may be performed within the service mapping UI, such as creating new credentials or changing MID server IP ranges. Other recommended actions may be external, which may include actions that need to be performed outside of the service mapping UI, such as installing a new MID server, or opening forwarding ports, etc. In certain cases, recommended actions may also include best practice recommendations which are recommended, but optional. Errors may also be flagged to be ignored or rediscovery may be run for selected error cards. For internal actions, rediscovery may be automatically run after the recommended action is applied. For error resolution using recommended external actions, a list of options for pushing the this task to external resources be displayed, such as exporting a file with the task and sending the file as an email attachment, opening a task or a task application for a user, or marking the external task as resolved.
Referring now to
The computing device 900 includes a processing element 902 that contains one or more hardware processors, where each hardware processor may have a single or multiple processor cores. In one embodiment, the processing element 902 may include at least one shared cache that store data (e.g., computing instructions) that are utilized by one or more other components of processing element 902. For example, the shared cache may be locally cache data stored in a memory for faster access by components of the processing elements 902. In one or more embodiments, the shared cache may include one or more mid-level caches, such as level 2 (L2), level 3 (L3), level 4 (L4), or other levels of cache, a last level cache (LLC), or combinations thereof. Examples of processors include, but are not limited to a central processing unit (CPU) a microprocessor. Although not illustrated in
Persons of ordinary skill in the art are aware that software programs may be developed, encoded, and compiled in a variety computing languages for a variety software platforms and/or operating systems and subsequently loaded and executed by processing element 902. In one embodiment, the compiling process of the software program may transform program code written in a programming language to another computer language such that the processing element 902 is able to execute the programming code. For example, the compiling process of the software program may generate an executable program that provides encoded instructions (e.g., machine code instructions) for processor 902 to accomplish specific, non-generic, particular computing functions.
After the compiling process, the encoded instructions may then be loaded as computer executable instructions or process steps to processing element 902 from storage (e.g., memory 904) and/or embedded within the processing element 902 (e.g., cache). Processing element 902 can execute the stored instructions or process steps in order to perform instructions or process steps to transform the computing device into a non-generic, particular, specially programmed machine or apparatus. Stored data, e.g., data stored by a storage device, can be accessed by processing element 902 during the execution of computer executable instructions or process steps to instruct one or more components within the computing device 900.
A user interface 910 can include a display, positional input device (such as a mouse, touchpad, touchscreen, or the like), keyboard, or other forms of user input and output devices. The user interface 910 can be coupled to processor element 902. Other output devices that permit a user to program or otherwise use the computing device can be provided in addition to or as an alternative to network communication unit 908. When the output device is or includes a display, the display can be implemented in various ways, including by a liquid crystal display (LCD) or a cathode-ray tube (CRT) or light emitting diode (LED) display, such as an organic LED (OLED) display. Persons of ordinary skill in the art are aware that the computing device 900 may comprise other components well known in the art, such as sensors, powers sources, and/or analog-to-digital converters, not explicitly shown in
At least one embodiment is disclosed and variations, combinations, and/or modifications of the embodiment(s) and/or features of the embodiment(s) made by a person having ordinary skill in the art are within the scope of the disclosure. Alternative embodiments that result from combining, integrating, and/or omitting features of the embodiment(s) are also within the scope of the disclosure. Where numerical ranges or limitations are expressly stated, such express ranges or limitations may be understood to include iterative ranges or limitations of like magnitude falling within the expressly stated ranges or limitations (e.g., from about 1 to about 10 includes, 2, 3, 4, etc.; greater than 0.10 includes 0.11, 0.12, 0.13, etc.). The use of the term “about” means ±10% of the subsequent number, unless otherwise stated.
Use of the term “optionally” with respect to any element of a claim means that the element is required, or alternatively, the element is not required, both alternatives being within the scope of the claim. Use of broader terms such as comprises, includes, and having may be understood to provide support for narrower terms such as consisting of, consisting essentially of, and comprised substantially of. Accordingly, the scope of protection is not limited by the description set out above but is defined by the claims that follow, that scope including all equivalents of the subject matter of the claims. Each and every claim is incorporated as further disclosure into the specification and the claims are embodiment(s) of the present disclosure.
It is to be understood that the above description is intended to be illustrative, and not restrictive. For example, the above-described embodiments may be used in combination with each other. Many other embodiments will be apparent to those of skill in the art upon reviewing the above description. The scope of the invention therefore should be determined with reference to the appended claims, along with the full scope of equivalents to which such claims are entitled. It should be noted that the discussion of any reference is not an admission that it is prior art to the present invention, especially any reference that may have a publication date after the priority date of this application.
This application claims priority to U.S. Provisional Patent Application Ser. No. 62/568,087 filed Oct. 4, 2017, and entitled “Platform Computing Environment and Functionality Thereof,” the contents of which are incorporated herein by reference.
Number | Name | Date | Kind |
---|---|---|---|
5978594 | Bonnell | Nov 1999 | A |
6321229 | Goldman | Nov 2001 | B1 |
6609122 | Ensor | Aug 2003 | B1 |
6658650 | Bates | Dec 2003 | B1 |
6799189 | Huxoll | Sep 2004 | B2 |
6816898 | Scarpelli | Nov 2004 | B1 |
6895586 | Brasher | May 2005 | B1 |
7027411 | Pulsipher | Apr 2006 | B1 |
7028301 | Ding | Apr 2006 | B2 |
7062683 | Warpenburg | Jun 2006 | B2 |
7131037 | LeFaive | Oct 2006 | B1 |
7170864 | Matharu | Jan 2007 | B2 |
7328260 | Muthiyan | Feb 2008 | B1 |
7392300 | Anantharangachar | Jun 2008 | B2 |
7610512 | Gerber | Oct 2009 | B2 |
7617073 | Trinon | Nov 2009 | B2 |
7689628 | Garg | Mar 2010 | B2 |
7716353 | Golovinsky | May 2010 | B2 |
7769718 | Murley | Aug 2010 | B2 |
7783744 | Garg | Aug 2010 | B2 |
7877783 | Cline | Jan 2011 | B1 |
7890802 | Gerber | Feb 2011 | B2 |
7925981 | Pourheidari | Apr 2011 | B2 |
7930396 | Trinon | Apr 2011 | B2 |
7941506 | Bonal | May 2011 | B2 |
7945860 | Vambenepe | May 2011 | B2 |
7966398 | Wiles | Jun 2011 | B2 |
8051164 | Peuter | Nov 2011 | B2 |
8082222 | Rangarajan | Dec 2011 | B2 |
8200803 | Benfield | Jun 2012 | B2 |
8224683 | Manos | Jul 2012 | B2 |
8266096 | Navarrete | Sep 2012 | B2 |
8346752 | Sirota | Jan 2013 | B2 |
8380645 | Kowalski | Feb 2013 | B2 |
8457928 | Dang | Jun 2013 | B2 |
8478569 | Scarpelli | Jul 2013 | B2 |
8554750 | Rangaranjan | Oct 2013 | B2 |
8646093 | Myers | Feb 2014 | B2 |
8674992 | Poston | Mar 2014 | B2 |
8683032 | Spinelli | Mar 2014 | B2 |
8689241 | Naik | Apr 2014 | B2 |
8706667 | Stanfill et al. | Apr 2014 | B2 |
8743121 | De Peuter | Jun 2014 | B2 |
8745040 | Kowalski | Jun 2014 | B2 |
8799609 | Bauer et al. | Aug 2014 | B1 |
8812539 | Milousheff | Aug 2014 | B2 |
8818994 | Kowalski | Aug 2014 | B2 |
8887133 | Behnia | Nov 2014 | B2 |
8907988 | Poston | Dec 2014 | B2 |
9015188 | Behne | Apr 2015 | B2 |
9037536 | Vos | May 2015 | B2 |
9065783 | Ding | Jun 2015 | B2 |
9098322 | Apte | Aug 2015 | B2 |
9122552 | Whitney | Sep 2015 | B2 |
9137115 | Mayfield | Sep 2015 | B2 |
9261372 | Cline | Feb 2016 | B2 |
9317327 | Apte | Apr 2016 | B2 |
9323801 | Morozov | Apr 2016 | B2 |
9363252 | Mueller | Jun 2016 | B2 |
9412084 | Kowalski | Sep 2016 | B2 |
9467344 | Gere | Oct 2016 | B2 |
9497136 | Ramarao | Nov 2016 | B1 |
9534903 | Cline | Jan 2017 | B2 |
9535737 | Joy | Jan 2017 | B2 |
9557969 | Sharma | Jan 2017 | B2 |
9613070 | Kumar | Apr 2017 | B2 |
9631934 | Cline | Apr 2017 | B2 |
9645833 | Mueller | May 2017 | B2 |
9654473 | Miller | May 2017 | B2 |
9659051 | Hutchins | May 2017 | B2 |
9766935 | Kelkar | Sep 2017 | B2 |
9792387 | George | Oct 2017 | B2 |
9805322 | Kelkar | Oct 2017 | B2 |
9852165 | Morozov | Dec 2017 | B2 |
9967162 | Spinelli | May 2018 | B2 |
10002203 | George | Jun 2018 | B2 |
20060184410 | Ramamurthy | Aug 2006 | A1 |
20130325789 | Krishnan | Dec 2013 | A1 |
20140047274 | Lumezanu | Feb 2014 | A1 |
20160014073 | Reddy | Jan 2016 | A1 |
20160034334 | Sadovsky et al. | Feb 2016 | A1 |
20160162825 | Dan | Jun 2016 | A1 |
20160359680 | Parandehgheibi | Dec 2016 | A1 |
20180123939 | Raman | May 2018 | A1 |
20180176261 | Bansal | Jun 2018 | A1 |
Number | Date | Country |
---|---|---|
2013112288 | Aug 2013 | WO |
Entry |
---|
IBM, “Handling errors in BPDs,” IBM Knowledgecenter, Retrieved from the Internet: URL: https://www.ibm.com/support/knowledgecenter/en/SSFTBX_8.5.5/com.ibm.wbpm.wle.editor.doc/topics/cauth_handling_errors_bpds.html [retrieved on Aug. 9, 2017]. |
Extended European Search Report for European Patent Application No. 18198156.4 dated Feb. 26, 2019; 10 pgs. |
First Examination Report for European Patent Application No. 18198156.4 dated Feb. 18, 2020; 8 pgs. |
Number | Date | Country | |
---|---|---|---|
20190104024 A1 | Apr 2019 | US |
Number | Date | Country | |
---|---|---|---|
62568087 | Oct 2017 | US |