Robotic process automation (RPA) systems enable automation of repetitive and manually intensive computer-based tasks. In an RPA system, a computer software, namely a software robot (often referred to as a “bot”), may mimic the actions of a human being in order to perform various computer-based tasks. For instance, an RPA system can be used to interact with one or more software applications through user interfaces, as a human being would do. Therefore, RPA systems typically do not need to be integrated with existing software applications at a programming level, thereby eliminating the difficulties inherent to integration, namely bringing together diverse components. Advantageously, RPA systems permit the automation of application level repetitive tasks via software robots that are coded to repeatedly and accurately perform the repetitive task.
RPA systems have their own graphical user interface to create, manage and execute software robots. Unfortunately, however, these graphical user interfaces serve a lot of functions that are not immediately accessible or readily understood by many users. Therefore, there is a need for improved approaches to access capabilities of RPA systems with increased efficiency and less domain knowledge.
Embodiments disclosed herein concern improved access to robotic process automation (RPA) systems. A user may interact with an RPA system by way of a communication platform. In one embodiment, the communication platform supports text messaging and/or natural language communications (i.e., voice or speech communications) with a virtual agent that interfaces with the RPA system. The user can communicate with the virtual agent using a conversational-based user interface. In this way, a user of the communication platform is able to conveniently interact with the RPA system, such as in a conversational manner. For example, a user can induce an action by the RPA system through use of one or more text messages or through use of natural language communications. By analyzing and interpreting the conversation, the user's intent or desire can be determined and carried out by the RPA system. Thereafter, results from the RPA system can be formatted and returned to the user via the conversational-based user interface or other means.
In one embodiment, to better understand the user's intent or desire from text-based messages or natural language communications (i.e., voice or speech communications), artificial intelligence can be used. Once the user's intent or desire with respect to an RPA system is estimated, then an appropriate software automation process supported by the RPA system can be determined and utilized.
The invention can be implemented in numerous ways, including as a method, system, device, apparatus (including computer readable medium and graphical user interface). Several embodiments of the invention are discussed below.
As a computer-implemented method for providing software automation, one embodiment can, for example, include at least: examining a conversation provided in a communication platform for a software automation objective; retrieving a virtual agent automation dialog relevant to the software automation objective; initiating presentation of the virtual agent automation dialog within the communication platform to acquire configuration parameters; and activating a software automation process in accordance with the configuration parameters to provide the software automation objective.
Optionally, the communication platform can be a unified communication platform supporting text and voice/speech communications. The communication platform can also be a unified communication and collaboration platform. The virtual agent automation dialog can comprise a chat session between the virtual agent and a user, wherein the chat session includes a plurality of text messages. Furthermore, the computer-implemented method can also identify the software automation process to be activated from a plurality of available software automation processes based on the software automation objective or the virtual agent automation dialog.
As a computer-implemented method for providing software automation, another embodiment can, for example, include at least: examining a conversation provided in a communication platform for a software automation objective; identifying a software automation process associated with the software automation objective; retrieving a parameter set for the identified software automation process; forming at least one dialog message to request parameter data for the parameter set for the identified software automation process; inserting the at least one dialog message into the conversation provided in the communication platform; examining the conversation provided in the communication platform for a response to the at least one dialog message; extracting the parameter data from the response to the at least one dialog message; and requesting execution of the identified software automation process in accordance with the parameter data from the response to the at least one dialog message.
As a computer-implemented system for facilitating conversational user interaction between a communication platform and a robotic process automation system, one embodiment can, for example, include at least: a communication platform interface to receive communication within the communication platform associated with a virtual digital assistant; an artificial intelligence (AI) platform interface to provide the received communication to an AI evaluation platform and to receive evaluation feedback from the AI evaluation platform; a plurality of dialogs used with or by the virtual digital assistant to support user access to the robotic process automation system; and a conversational control module. The conversational control module can be configured to: provide the received communication, received via the communication platform interface, to the AI platform interface; receive the evaluation feedback from the AI platform interface; select at least one dialog from the plurality of dialogs based on evaluation feedback; identify a software automation process associated with the received communication, the evaluation feedback and/or the selected dialog; invoke the selected dialog with or by the virtual digital assistant to acquire user parameter data for use with the identified software automation process; and activate the identified software automation process based on at least a portion of the acquired user parameter data.
As a non-transitory computer readable medium including at least computer program code tangible stored thereon for providing software automation, one embodiment can, for example, include at least: computer program code for examining a conversation provided in a communication platform for a software automation objective; computer program code for retrieving a virtual agent automation dialog relevant to the software automation objective; computer program code for initiating presentation of the virtual agent automation dialog within the communication platform to acquire configuration parameters; and computer program code for activating a software automation process in accordance with the configuration parameters to provide the software automation objective.
As a non-transitory computer readable medium including at least computer program code tangible stored thereon for providing software automation, another embodiment can, for example, include at least: computer program code for examining a conversation provided in a communication platform for a software automation objective; computer program code for identifying a software automation process associated with the software automation objective; computer program code for retrieving a parameter set for the identified software automation process; computer program code for forming at least one dialog message to request parameter data for the parameter set for the identified software automation process; computer program code for inserting the at least one dialog message into the conversation provided in the communication platform; computer program code for examining the conversation provided in the communication platform for a response to the at least one dialog message; computer program code for extracting the parameter data from the response to the at least one dialog message; and computer program code for requesting execution of the identified software automation process in accordance with the parameter data from the response to the at least one dialog message.
Other aspects and advantages of the invention will become apparent from the following detailed description taken in conjunction with the accompanying drawings which illustrate, by way of example, the principles of the invention.
The invention will be readily understood by the following detailed description in conjunction with the accompanying drawings, wherein like reference numerals designate like elements, and in which:
Embodiments disclosed herein concern improved access to RPA systems. A user may interact with an RPA system by way of a communication platform. In one embodiment, the communication platform supports text messaging and/or natural language communications (i.e., voice or speech communications) with a virtual agent that interfaces with the RPA system. The user can communicate with the virtual agent using a conversational-based user interface. In this way, a user of the communication platform is able to conveniently interact with the RPA system, such as in a conversational manner. For example, a user can induce an action by the RPA system through use of one or more text messages or through use of natural language communications. By analyzing and interpreting the conversation, the user's intent or desire can be determined and carried out by the RPA system. Thereafter, results from the RPA system can be formatted and returned to the user via the conversational-based user interface or other means.
In one embodiment, to better understand the user's intent or desire from the text messages or natural language communications (i.e., voice or speech communications), artificial intelligence can be used. Once the user's intent or desire with respect to an RPA system is estimated, then an appropriate software automation process supported by the PRA system can be determined and utilized. Advantageously, users are able to interact with RPA systems using a conversational style that is readily understood by users.
Generally speaking, RPA systems use computer software to emulate and integrate the actions of a human interacting within digital systems. In an enterprise environment, these RPA systems are often designed to execute a business process. In some cases, the RPA systems use AI and/or other machine learning capabilities to handle high-volume, repeatable tasks that previously required humans to perform. The RPA systems support a plurality of software automation processes (SAPs). The RPA systems also provide for creation, configuration, management, execution, monitoring, and performance of software automation processes.
A software automation process can also be referred to as a software robot, software agent, or a bot. A software automation process can interpret and execute tasks on your behalf. Software automation processes are particularly well suited for handling a lot of the repetitive tasks that humans perform every day. Software automation processes can perform a task or workflow they are tasked with once or 10,000 times and do it accurately every time. As one example, a software automation process can locate and read data in a document, email, file, or window. As another example, a software automation process can connect with one or more Enterprise Resource Planning (ERP), Customer Relations Management (CRM), core banking, and other business systems to distribute data where it needs to be in whatever format is necessary. As another example, a software automation process can perform data tasks, such as reformatting, extracting, balancing, error checking, moving, copying, etc. As another example, a software automation process can grab data desired from a webpage, application, screen, file, or other data source. As still another example, a software automation process can be trigger based on time or an event, and can serve to take files or data sets and move them to another location, whether it is to a customer, vendor, application, department or storage. These various capabilities can also be used in any combination. As an example of an integrated software automation process, the software automation process can start a task or workflow based on a trigger, such as a file being uploaded to an FTP system. The integrated software automation process can then download that file, scrape relevant data from it, upload the relevant data to a database, and then send an email to inform the recipient that the data has been successfully processed.
Embodiments of various aspects of the invention are discussed below with reference to
Additionally, the conversational control module 102 can interact with a robotic process automation system 112. The robotic process automation system 112 supports a plurality of different robotic processes, which are denoted software automation processes 114. The software automation processes 114 can also be referred to as “bots.” The robotic process automation system 112 can maintain, execute, and/or monitor software automation processes 114. The robotic process automation system 112 can also report status or results of software automation processes 114.
Further, the conversational control module 102 can interact with a dialog storage 116. The dialog storage 116 can store a plurality of different dialogs. Each different dialog can pertain to a portion of a conversation to be had with a requestor. The requestor is a user that uses the communication platform 104 to request some action from the RPA control system 100. The requestor uses a computing device to interact with the communication platform 104 in a wired or wireless manner. The computing device can, for example, be a mobile phone, tablet computer, desktop computer, and the like. The dialogs serve to elicit responses from the requestor. The dialogs can be text-based or speech-based. In one implementation, the dialogs are associated with a particular one of the software automation processes 114 maintained and operated by the robotic process automation system 112. The dialogs can serve to allow the requestor interacting with the communication platform 104 to effectively interact with the robotic process automation system 112 with the assistance of the conversational control module 102 and related components of the conversational RPA control system 100.
Still further, the conversational control module 102 can interact with a database 118. The database 118 can store structured or relational data for the conversational RPA control system 100. In one embodiment, the database 118 can be used to record conversational data pertaining to one or more active conversations at the communication platform 104 between a requestor and a virtual agent. The recorded conversation data can include requestor data (account, profile, access level, and any other data), conversation state, active dialog, and/or dialog state. For example, in one implementation, the database 118 can provide storage of one or more conversation states, a dialog state, and a user account.
The conversation state can record the state of an on-going structured conversation provide between a user (e.g., requestor) and virtual agent. The dialog state can record the state of an on-going dialog between the virtual agent and a user of the communication platform 104. The user account, is an account for a user and can contain access rights to the robotic process automation system 112.
The conversational software automation process 200 can initially examine 202 a conversation in a communication platform. The conversation in the communication platform can be ongoing or can be initiated by the user or the virtual agent. A decision 204 can then determine whether a software automation objective has been detected. Here, the conversation is examined 202 to determine whether a software automation objective is being sought by the conversation. When the decision 204 determines that a software automation objective has not been detected, then the conversational software automation process 200 can return to repeat block 202 so that the conversation can be further examined.
On the other hand, when decision 204 determines that the software automation objective has been detected, a virtual agent automation dialog that is relevant to the software automation objective can be identified 206. Next, presentation of the virtual agent automation dialogue can be initiated 208 to acquire configuration parameters. Then, a software automation process can be activated 210 in accordance with the configuration parameters to provide the software automation objective. Here, the software automation process being activated 210 can be identified based on the software automation objective and/or the virtual agent automation dialog. For example, the software automation process being activated 210 can be mapped to the software automation objective and/or the virtual agent automation dialog. Following block 210, the conversational software automation process 200 can end.
Typically, the software automation objective is to provide an interaction with a robotic process automation system, such as the robotic process automation system 112 illustrated in
Instead, the other interactions might request: (i) report(s), (ii) usage information, (iii) performance information, (iv) search of available software automation processes, and/or (v) various other interactions that are supported by the robotic process automation system (e.g., robotic process automation system 112 illustrated in
The conversational automation process 300 can begin with a decision 302 that determines whether a virtual assistant (VA) message has been received. The VA message is a message that is received via a communication platform, such as the communication platform 104 illustrated in
When the decision 304 determines that the VA message is not for an RPA system, then the VA message is responded 306 to by other processing. This other processing depends on other system capabilities but is unrelated to the RPA system. Here, the virtual assistant may, but need not, offer assistance with other areas besides an RPA system. Following the block 306, the conversational automation process 300 can end.
On the other hand, when the virtual assistant message is for an RPA system, then a particular software automation process (SAP) associated with the VA message can be determined 308. Here, in one implementation, based on the VA message, the particular software automation process can be determined 308. For example, if the focus of the VA message is to generate and receive a CRM report, the particular software automation process capable of doing so can be identified and utilized. In one embodiment, if there are multiple candidate software automation processes that are capable of doing the focus of the VA message, then additional messaging can be provided via the VA to determine which of the multiple candidate software automation processes is best suited.
After the particular software automation process has been determined 308, a parameter set identifying parameters for the particular software automation process can be retrieved 310. Then, one or more dialog messages can be formed 312 to request parameter data for the parameters in the parameter set. Next, as illustrated in
Alternatively, when the decision 318 determines that the conversational automation process 300 should no longer await such a response, then the conversational automation process 300 can fail 320 the RPA system request. For example, after a predetermined period of time waiting, the wait for a dialog response can be deemed to have timed-out and thus terminate the dialog and fail (or end) processing of the RPA system request. After the RPA system request has failed 320, the conversational automation process 300 can end.
On the other hand, when the decision 316 determines that a dialog response has been received, then parameter data can be extracted 322 from the dialog response. Thereafter, a decision 324 can determine whether there are any more dialog messages to be sent. When the decision 324 determines that there are one or more dialog messages to be sent, the processing can return to repeat the block 314 and subsequent blocks so that a next dialog message can be sent to the requestor and a response thereto can be processed in a similar fashion. Alternatively, when the decision 324 determines that no more dialog messages are to be sent, execution of the particular software automation process in accordance with the parameter data can be requested 326. Thereafter, the conversational automation process 300 can end.
The status message process 400 can determine whether a particular software automation process has completed 402. When the decision 402 determines that the particular software automation process has not completed, the status message process 400 can await completion of the particular software automation process. On the other hand, when the decision 402 determines that the particular software automation process has completed, status data from the particular software automation process can be retrieved 404. A status message can then be formed 406. Thereafter, the status message process 400 can initiate sending 408 of the status message to the requestor. The status message can inform the requestor that the particular software automation process has completed its processing and provide further information, such as execution time, time and date of execution, and whether a report was generated. If a report was generated by the particular software automation process, then the status message can also facilitate access to the report. For example, the status message can enable the requestor to download the report or to electronically save the report. The status message can be sent 408 in a variety of ways. In one example, the status message can be sent to the requestor via electronic mail, text message, or other electronic means. In another example, the status message can be sent to the requestor via a virtual agent used with a communication platform, such as the communication platform 104. Following the block 408, the status message process 400 can end.
The conversational automation process 500 can launch 502 a communication platform. The communication platform can support one or both of text-based communication and speech-based communication. A virtual agent can also be enabled 504 for use with the communication platform. The virtual agent is a digital agent that provides user assistance through conversations. In this embodiment, the virtual agent is configured to participate in communications or interactions using the communication platform.
Next, a user can be permitted 506 to interact with the virtual agent via the communication platform. The user and the virtual agent can participate in a conversation via the communication platform. In doing so, the conversational automation process 500 can recognize 508 a user's desire to utilize a software automation process (SAP) available from an RPA system. By examining (e.g., parsing) the conversation, the conversational automation process 500 is able to recognize 508 the user's desire to utilize the software automation process. Here, in one implementation, artificial intelligence can be used based on recognizing 508 the user's desire from examination of the conversation between the virtual agent and the user via the communication platform. Since the RPA system supports a plurality of software automation processes, the appropriate software automation process can be identified by the user's desire based on examination of a conversation. However, if there is not clear identification of the appropriate software automation process from the user's desire, then additional dialog can be provided via the communication platform to request and receive additional input from the user using the virtual agent so that the appropriate software automation process can be properly identified.
After the user's desire to utilize the software automation process has been recognized 508, parameters of the software automation process can be accessed 510. Here, each software automation process typically requires one or more input parameters (or variables) for its proper execution. Thus, a message query can be formed 512. The message query can, for example, request parameter input from the user. After the message query has been formed 512, the message query can be presented 514 to the user via the communication platform, such as via the virtual agent. Here, each software automation process typically requires one or more input parameters (or variables) for its proper execution. The message query that is formed 512 and presented 514 can seek the one or more input parameters from the user in a conversational fashion via the communication platform.
Referring now to
Next, execution of the software automation process and can be requested 520 in accordance with the parameter data. After the software automation process has started its execution, the software automation process will eventually complete its tasks. The completion of the software automation process can be discovered 522. After the completion of the software automation process has been discovered 522, status data for the software automation process can be obtained 524. For example, the conversational automation process 500 can request and receive the status data from the software automation process itself or its robotic process automation system, such as the robotic process automation system 112 shown in
Software automation processes can be used in a wide range of computing environments or situations. One example of a particular software automation process provides interaction with a CRM software system. Typically, CRM software systems are used to provide various reports, such as sales report.
The conversational control module, such as the conversational control module 102, can decipher a requestor's desire from a conversation happening in a communication platform. The conversational control module can be assisted by an AI platform, such as the AI platform 108. In such case, the AI platform can serve to decipher the requestor's desire from the conversation in the communication platform. In this example, the requestor desires to get a CRM report. The AI platform can, for this example, retrieving reports for CRM reports (e.g., Salesforce reports) can be trained to understand this specific desire of the requestor. For instance, the AI platform can be trained to recognize such a specific desire by being provided some examples of what the user might say:
1. Get my Salesforce report
2. Retrieve my sales report
3. Download my Salesforce report for me?
4. Can you get my sales report?
These examples are fed into the AI platform, which will create a machine learning module that will understand new speech that are not in the examples. In this case, the desire (or intent) can be specifically mapped to the “Get Report” intent which have been defined. One suitable AI platform is LUIS available from Microsoft Corporation at www.luis.ai. LUIS is a machine learning-based service designed to identify valuable information in conversations. LUIS can interpret user goals (intents) and distills valuable information from sentences (entities), for a high quality, nuanced language model.
In one embodiment, the AI platform can extract the “subject” entity. In LUIS, entities are any pieces of information that can be extracted from speech. In this embodiment, the “subject” entity is being extracted by the AI platform, which in our example, is input “Salesforce report.” The “subject” entity is useful because the user might also seek other reports from other systems, for example, Facebook, Google, Oracle, etc. By extracting “subject,” the conversational control module can then determine which specific report to receive or obtain.
The following examples are simply for illustration purposes only and not intended to be limiting in any way.
As an example, an exemplary response to an Application Programming Interface (API) call to the AI platform to evaluate a communication (e.g., query) received via the communication platform 104 can be as follows.
In this example API call, the query is what the requestor (or user) entered via the communication platform. The conversational control module can examine the response from the AI platform and extract (i) intent, which is “Get Report”, and (ii) subject, which is “Salesforce report”. The “top_intent” is the intent used as the requestor's desire, and the entities contain the subject of the conversation (e.g., query).
Hence, an exemplary software automation process of a CRM software system might then be used to generate a sales report, as being understood as the requestor's desire. However, typically, the exemplary software automation process requires one or more parameters (or variables), namely input parameters (or variables). For example, the parameters for the exemplary software automation process might include: Region, Date and Report. The Region can be a character string identifying a geographic region for the report. The Date can be a character string identifying a date or date range for the report. The Report can be a number identifying a type or format for the report.
In one embodiment, the robotic process automation system supporting the exemplary software automation process, can be accessed to request and receive the parameters that are needed for the exemplary software automation process. In one implementation, the robotic process automation system can include an API that programmatically permits access to the parameter information for the various software automation processes it supports. For example, for the exemplary software automation process that provides a sales report from a CRM system, an API can be used to request and receive a set of parameters (or variables). The set of parameters (or variables) for the exemplary software automation process provided via the API, might be denoted as:
Hence, in this example, the set of parameters (or variables) for the exemplary software automation process to provide a sales report for the CRM system includes: Region, Date and Report, which are input parameters (or variables).
Thereafter, when the conversational control module is ready to execute the exemplary software automation process using the robotic process automation system, such as the robotic process automation system 112, the conversational control module can issue an API call to the robotic process automation system requesting execution of the exemplary software automation process and provide any needed parameters (or variables) therefor. An exemplary API call to the robotic process automation system for a sale report from the CRM system may be as follows.
When the exemplary software automation process completes its execution, the desired sales report may be produced. Hence, at this point the requestor can be alerted to the availability of the sales report. More particularly, on completion, the exemplary software automation process returns a result back to the conversational control module that can be used to cause a message to be provided back to the requestor via the communication platform 104. An example of a result callback from the exemplary software automation process after it has completed execution can be as follows.
In this example, the exemplary software automation process on completion returns two output variables, “ReportName” and “FileDownloadURL”. These variables can be extracted from the returned data at the conversational control module. The conversational control module can then interact with the communication platform 104 to cause the virtual agent to return result information to the requestor. In this example, the communication provided to the requestor can be information that the requested report, denoted “Daily Report”, is ready and available for download from at a denoted Universal Resource Locator (URL).
As noted previously, the conversational-based user interface allows a requestor (or user) to interface with a robotic process automation system in a conversational fashion. In such cases, the communication provided in the conversational fashion may not occur in real-time but can be extended over a period of time, such as several minutes, hours or days. Typically, such communication may also be unstructured. The conversational-based user interface can cause a virtual agent to present graphical user interface screens to the requestor (or user). Exemplary graphical user interface screens are described below with reference to
Dialogs can also be customized to users or enterprises. Dialogs can be different depending on the conversation type, such as text, natural language, or form-based. Inputs sought and/or input default values for dialogs can also be customized. Validation to be performed can be customized. Dialogs can make use of pre-defined questions. Dialogs can also be dependent on status of a relevant software automation process (e.g., bot), such as active or disabled. The customization can be on all conversations of a user or enterprise, or can differ for each conversation of a user or enterprise.
As noted above a database, such as the database 118 illustrated in
In this exemplary structure, various variables are stored to assist with management of dialogs. In this example, the dialog management data can store identifiers, documents, states, and user data. For example, the variable “DialogState” can be used to record and thus track where in a given dialog the communication interchange is with a particular user. A variable “dialogStack” can be used to record a stack position (or level) within a dialog. As another example, the variable “DialogStateDictionary” can be used to record dialog variables that have be acquired. As still another example, the “UserStateModel” can be used to record and thus track information about the user. The information about the user recorded in the “UserStateModel” can, for example, include “AuthContext” (with URL, token and Username) to track user login to a RPA system; “ClientId” to track the particular user in the conversation; “TenantId” to track a particular company to which the particular user is affiliated; and “RunHistories” to track a history of all software automation processes (e.g., bots) the particular user has previously run.
The RPA system 1000 can also include a control room 1008. The control room 1008 is operatively coupled to the data storage 1002 and is configured to execute instructions that, when executed, cause the RPA system 1000 to respond to a request from a client device 1010 that is issued by a user 1012.1. The control room 1008 can act as a server to provide to the client device 1010 the capability to perform an automation task to process a work item from the plurality of work items 1006. The RPA system 1000 is able to support multiple client devices 1010 concurrently, each of which will have one or more corresponding user session(s) 1018, which provides a context. The context can, for example, include security, permissions, audit trails, etc. to define the permissions and roles for bots operating under the user session 1018. For example, a bot executing under a user session, cannot access any files or use any applications that the user, under whose credentials the bot is operating, does not have permission to do so. This prevents any inadvertent or malicious acts from a bot under which bot 1004 executes.
The control room 1008 can provide, to the client device 1010, software code to implement a node manager 1014. The node manager 1014 executes on the client device 1010 and provides a user 1012 a visual interface via browser 1013 to view progress of and to control execution of automation tasks. It should be noted that the node manager 1014 can be provided to the client device 1010 on demand, when required by the client device 1010, to execute a desired automation task. In one embodiment, the node manager 1014 may remain on the client device 1010 after completion of the requested automation task to avoid the need to download it again. In another embodiment, the node manager 1014 may be deleted from the client device 1010 after completion of the requested automation task. The node manager 1014 can also maintain a connection to the control room 1008 to inform the control room 1008 that device 1010 is available for service by the control room 1008, irrespective of whether a live user session 1018 exists. When executing a bot 1004, the node manager 1014 can impersonate the user 1012 by employing credentials associated with the user 1012.
The control room 1008 initiates, on the client device 1010, a user session 1018 (seen as a specific instantiation 1018.1) to perform the automation task. The control room 1008 retrieves the set of task processing instructions 1004 that correspond to the work item 1006. The task processing instructions 1004 that correspond to the work item 1006 can execute under control of the user session 1018.1, on the client device 1010. The node manager 1014 can provide update data indicative of status of processing of the work item to the control room 1008. The control room 1008 can terminate the user session 1018.1 upon completion of processing of the work item 1006. The user session 1018.1 is shown in further detail at 1019, where an instance 1024.1 of user session manager 1024 is seen along with a bot player 1026, proxy service 1028, and one or more virtual machine(s) 1030, such as a virtual machine that runs Java® or Python®. The user session manager 1024 provides a generic user session context within which a bot 1004 executes.
The bots 1004 execute on a player, via a computing device, to perform the functions encoded by the bot. Some or all of the bots 1004 may in certain embodiments be located remotely from the control room 1008. Moreover, the devices 1010 and 1011, which may be conventional computing devices, such as for example, personal computers, server computers, laptops, tablets and other portable computing devices, may also be located remotely from the control room 1008. The devices 1010 and 1011 may also take the form of virtual computing devices. The bots 1004 and the work items 1006 are shown in separate containers for purposes of illustration but they may be stored in separate or the same device(s), or across multiple devices. The control room 1008 can perform user management functions, source control of the bots 1004, along with providing a dashboard that provides analytics and results of the bots 1004, performs license management of software required by the bots 1004 and manages overall execution and management of scripts, clients, roles, credentials, security, etc. The major functions performed by the control room 1008 can include: (i) a dashboard that provides a summary of registered/active users, tasks status, repository details, number of clients connected, number of scripts passed or failed recently, tasks that are scheduled to be executed and those that are in progress; (ii) user/role management—permits creation of different roles, such as bot creator, bot runner, admin, and custom roles, and activation, deactivation and modification of roles; (iii) repository management—to manage all scripts, tasks, workflows and reports etc.; (iv) operations management—permits checking status of tasks in progress and history of all tasks, and permits the administrator to stop/start execution of bots currently executing; (v) audit trail—logs creation of all actions performed in the control room; (vi) task scheduler—permits scheduling tasks which need to be executed on different clients at any particular time; (vii) credential management—permits password management; and (viii) security: management—permits rights management for all user roles. The control room 1008 is shown generally for simplicity of explanation. Multiple instances of the control room 1008 may be employed where large numbers of bots are deployed to provide for scalability of the RPA system 1000.
In the event that a device, such as device 1011 (e.g., operated by user 1012.2) does not satisfy the minimum processing capability to run a node manager 1014, the control room 1008 can make use of another device, such as device 1015, that has the requisite capability. In such case, a node manager 1014 within a Virtual Machine (VM), seen as VM 1016, can be resident on the device 1015. The node manager 1014 operating on the device 1015 can communicate with browser 1013 on device 1011. This approach permits RPA system 1000 to operate with devices that may have lower processing capability, such as older laptops, desktops, and portable/mobile devices such as tablets and mobile phones. In certain embodiments the browser 1013 may take the form of a mobile application stored on the device 1011. The control room 1008 can establish a user session 1018.2 for the user 1012.2 while interacting with the control room 1008 and the corresponding user session 1018.2 operates as described above for user session 1018.1 with user session manager 1024 operating on device 1010 as discussed above.
In certain embodiments, the user session manager 1024 provides five functions. First is a health service 1038 that maintains and provides a detailed logging of bot execution including monitoring memory and CPU usage by the bot and other parameters such as number of file handles employed. The bots 1004 can employ the health service 1038 as a resource to pass logging information to the control room 1008. Execution of the bot is separately monitored by the user session manager 1024 to track memory, CPU, and other system information. The second function provided by the user session manager 1024 is a message queue 1040 for exchange of data between bots executed within the same user session 1018. The third function is a deployment service (also referred to as a deployment module) 1042 that connects to the control room 1008 to request execution of a requested bot 1004. The deployment service 1042 can also ensure that the environment is ready for bot execution, such as by making available dependent libraries. The fourth function is a bot launcher 1044 which can read metadata associated with a requested bot 1004 and launch an appropriate container and begin execution of the requested bot. The fifth function is a debugger service 1046 that can be used to debug bot code.
The bot player 1026 can execute, or play back, a sequence of instructions encoded in a bot. The sequence of instructions can, for example, be captured by way of a recorder when a human performs those actions, or alternatively the instructions are explicitly coded into the bot. These instructions enable the bot player 1026, to perform the same actions as a human would do in their absence. In one implementation, the instructions can compose of a command (action) followed by set of parameters, for example: Open Browser is a command, and a URL would be the parameter for it to launch a web resource. Proxy service 1028 can enable integration of external software or applications with the bot to provide specialized services. For example, an externally hosted artificial intelligence system could enable the bot to understand the meaning of a “sentence.”
The user 1012.1 can interact with node manager 1014 via a conventional browser 1013 which employs the node manager 1014 to communicate with the control room 1008. When the user 1012.1 logs in from the client device 1010 to the control room 1008 for the first time, the user 1012.1 can be prompted to download and install the node manager 1014 on the device 1010, if one is not already present. The node manager 1014 can establish a web socket connection to the user session manager 1024, deployed by the control room 1008 that lets the user 1012.1 subsequently create, edit, and deploy the bots 1004.
In the embodiment shown in
Turning to the bots Bot 1 and Bot 2, each bot may contain instructions encoded in one or more programming languages. In the example shown in
The control room 1008 operates to compile, via compiler 1208, the sets of commands generated by the editor 1202 or the recorder 1204 into platform independent executables, each of which is also referred to herein as a bot JAR (Java ARchive) that perform application level operations captured by the bot editor 1202 and the bot recorder 1204. In the embodiment illustrated in
As noted in connection with
An entry class generator 1308 can create a Java class with an entry method, to permit bot execution to be started from that point. For example, the entry class generator 1308 takes, as an input, a parent bot name, such “Invoice-processing.bot” and generates a Java class having a contract method with a predefined signature. A bot class generator 1310 can generate a bot class and orders command code in sequence of execution. The bot class generator 1310 can take, as input, an in-memory bot structure and generates, as output, a Java class in a predefined structure. A Command/Iterator/Conditional Code Generator 1312 wires up a command class with singleton object creation, manages nested command linking, iterator (loop) generation, and conditional (If/Else If/Else) construct generation. The Command/Iterator/Conditional Code Generator 1312 can take, as input, an in-memory bot structure in JSON format and generates Java code within the bot class. A variable code generator 1314 generates code for user defined variables in the bot, maps bot level data types to Java language compatible types, and assigns initial values provided by user. The variable code generator 1314 takes, as input, an in-memory bot structure and generates Java code within the bot class. A schema validator 1316 can validate user inputs based on command schema and includes syntax and semantic checks on user provided values. The schema validator 1316 can take, as input, an in-memory bot structure and generates validation errors that it detects. The attribute code generator 1318 can generate attribute code, handles the nested nature of attributes, and transforms bot value types to Java language compatible types. The attribute code generator 1318 takes, as input, an in-memory bot structure and generates Java code within the bot class. A utility classes generator 1320 can generate utility classes which are used by an entry class or bot class methods. The utility classes generator 1320 can generate, as output, Java classes. A data type generator 1322 can generate value types useful at runtime. The data type generator 1322 can generate, as output, Java classes. An expression generator 1324 can evaluate user inputs and generates compatible Java code, identifies complex variable mixed user inputs, inject variable values, and transform mathematical expressions. The expression generator 1324 can take, as input, user defined values and generates, as output, Java compatible expressions.
The JAR generator 1328 can compile Java source files, produces byte code and packs everything in a single JAR, including other child bots and file dependencies. The JAR generator 1328 can take, as input, generated Java files, resource files used during the bot creation, bot compiler dependencies, and command packages, and then can generate a JAR artifact as an output. The JAR cache manager 1330 can put a bot JAR in cache repository so that recompilation can be avoided if the bot has not been modified since the last cache entry. The JAR cache manager 1330 can take, as input, a bot JAR.
In one or more embodiment described herein command action logic can be implemented by commands 1201 available at the control room 1008. This permits the execution environment on a device 1010 and/or 1015, such as exists in a user session 1018, to be agnostic to changes in the command action logic implemented by a bot 1004. In other words, the manner in which a command implemented by a bot 1004 operates need not be visible to the execution environment in which a bot 1004 operates. The execution environment is able to be independent of the command action logic of any commands implemented by bots 1004. The result is that changes in any commands 1201 supported by the RPA system 1000, or addition of new commands 1201 to the RPA system 1000, do not require an update of the execution environment on devices 1010, 1015. This avoids what can be a time and resource intensive process in which addition of a new command 1201 or change to any command 1201 requires an update to the execution environment to each device 1010, 1015 employed in a RPA system. Take, for example, a bot that employs a command 1201 that logs into an on-online service. The command 1201 upon execution takes a Uniform Resource Locator (URL), opens (or selects) a browser, retrieves credentials corresponding to a user on behalf of whom the bot is logging in as, and enters the user credentials (e.g. username and password) as specified. If the command 1201 is changed, for example, to perform two-factor authentication, then it will require an additional resource (the second factor for authentication) and will perform additional actions beyond those performed by the original command (for example, logging into an email account to retrieve the second factor and entering the second factor). The command action logic will have changed as the bot is required to perform the additional changes. Any bot(s) that employ the changed command will need to be recompiled to generate a new bot JAR for each changed bot and the new bot JAR will need to be provided to a bot runner upon request by the bot runner. The execution environment on the device that is requesting the updated bot will not need to be updated as the command action logic of the changed command is reflected in the new bot JAR containing the byte code to be executed by the execution environment.
The embodiments herein can be implemented in the general context of computer-executable instructions, such as those included in program modules, being executed in a computing system on a target, real or virtual, processor. Generally, program modules include routines, programs, libraries, objects, classes, components, data structures, etc. that perform particular tasks or implement particular abstract data types. The program modules may be obtained from another computer system, such as via the Internet, by downloading the program modules from the other computer system for execution on one or more different computer systems. The functionality of the program modules may be combined or split between program modules as desired in various embodiments. Computer-executable instructions for program modules may be executed within a local or distributed computing system. The computer-executable instructions, which may include data, instructions, and configuration parameters, may be provided via an article of manufacture including a computer readable medium, which provides content that represents instructions that can be executed. A computer readable medium may also include a storage or database from which content can be downloaded. A computer readable medium may further include a device or product having content stored thereon at a time of sale or delivery. Thus, delivering a device with stored content, or offering content for download over a communication medium, may be understood as providing an article of manufacture with such content described herein.
The exemplary computing environment 1400 may have additional features such as, for example, tangible storage 1410, one or more input devices 1414, one or more output devices 1412, and one or more communication connections 1416. An interconnection mechanism (not shown) such as a bus, controller, or network can interconnect the various components of the exemplary computing environment 1400. Typically, operating system software (not shown) provides an operating system for other software executing in the exemplary computing environment 1400, and coordinates activities of the various components of the exemplary computing environment 1400.
The tangible storage 1410 may be removable or non-removable, and includes magnetic disks, magnetic tapes or cassettes, CD-ROMs, DVDs, or any other medium which can be used to store information in a non-transitory way, and which can be accessed within the computing system 1400. The tangible storage 1410 can store instructions for the software implementing one or more features of a PRA system as described herein.
The input device(s) or image capture device(s) 1414 may include, for example, one or more of a touch input device such as a keyboard, mouse, pen, or trackball, a voice input device, a scanning device, an imaging sensor, touch surface, or any other device capable of providing input to the exemplary computing environment 1400. For multimedia embodiment, the input device(s) 1414 can, for example, include a camera, a video card, a TV tuner card, or similar device that accepts video input in analog or digital form, a microphone, an audio card, or a CD-ROM or CD-RW that reads audio/video samples into the exemplary computing environment 1400. The output device(s) 1412 can, for example, include a display, a printer, a speaker, a CD-writer, or any another device that provides output from the exemplary computing environment 1400.
The one or more communication connections 1416 can enable communication over a communication medium to another computing entity. The communication medium conveys information such as computer-executable instructions, audio or video input or output, or other data. The communication medium can include a wireless medium, a wired medium, or a combination thereof.
The various aspects, features, embodiments or implementations of the invention described above can be used alone or in various combinations.
Embodiments of the invention can, for example, be implemented by software, hardware, or a combination of hardware and software. Embodiments of the invention can also be embodied as computer readable code on a computer readable medium. In one embodiment, the computer readable medium is non-transitory. The computer readable medium is any data storage device that can store data which can thereafter be read by a computer system. Examples of the computer readable medium generally include read-only memory and random-access memory. More specific examples of computer readable medium are tangible and include Flash memory, EEPROM memory, memory card, CD-ROM, DVD, hard drive, magnetic tape, and optical data storage device. The computer readable medium can also be distributed over network-coupled computer systems so that the computer readable code is stored and executed in a distributed fashion.
Numerous specific details are set forth in order to provide a thorough understanding of the present invention. However, it will become obvious to those skilled in the art that the invention may be practiced without these specific details. The description and representation herein are the common meanings used by those experienced or skilled in the art to most effectively convey the substance of their work to others skilled in the art. In other instances, well-known methods, procedures, components, and circuitry have not been described in detail to avoid unnecessarily obscuring aspects of the present invention.
In the foregoing description, reference to “one embodiment” or “an embodiment” means that a particular feature, structure, or characteristic described in connection with the embodiment can be included in at least one embodiment of the invention. The appearances of the phrase “in one embodiment” in various places in the specification are not necessarily all referring to the same embodiment, nor are separate or alternative embodiments mutually exclusive of other embodiments. Further, the order of blocks in process flowcharts or diagrams representing one or more embodiments of the invention do not inherently indicate any particular order nor imply any limitations in the invention.
The many features and advantages of the present invention are apparent from the written description. Further, since numerous modifications and changes will readily occur to those skilled in the art, the invention should not be limited to the exact construction and operation as illustrated and described. Hence, all suitable modifications and equivalents may be resorted to as falling within the scope of the invention.