This U.S. patent application claims priority under 35 U.S.C. § 119 to: India application Ser. No. 20/232,1087106, filed on Dec. 20, 2023. The entire contents of the aforementioned application are incorporated herein by reference
The disclosure herein generally relate to the field of Artificial Intelligence (AI) model development and, more particularly, to a method and system for selection-optimization-fine-tuning of Artificial Intelligence (AI) models for nuanced task by analyzing the nuanced task using Logic and Rule Integrated Decoder Neural Network (LRI-DecNN)
Leveraging Generative Artificial Intelligence (GenAI) to automate complex, nuanced tasks within enterprises is an open area of research. These tasks, often similar but not identical, require a degree of understanding and adaptability that surpasses traditional automation tools. Nuanced tasks can be defined as tasks that are complex, ambiguous, heterogenous, and the ones that require highly skilled human beings to understand and execute successfully. The current landscape of automating complex, highly skilled tasks using Generative AI presents several significant challenges for enterprises such as ensuring data quality, model training complexities, balancing creativity with accuracy, or integrating the model into existing workflows. Mere focus on task automation using AI, or GenAI may not be always the right approach for effective automation with AI augmentation. The effort that has gone into automation should be outperformed by results or effectiveness of the automation. Thus understanding the task complexity, the task in entirety and as a sequence of subtasks plays a major role in task automation effectiveness. The right AI model selection according to the task is another major factor.
Some recent works have proposed task automation, wherein they attempt to determine a task automation score by analyzing the task using Natural Language Processing and later decide on effectiveness of task automation based on the score. However, once identified for task automation the task is classified into predefined automation categories for further AI based automation. The automation category for AI usage here provides a static approach for automation. Furthermore, as mentioned tasks for automation may be better executed in segments, and if generalized under one automation approach may not provide effective automation solution. Furthermore, continual learning incorporating human feedback is a key to enhance efficiency and effectiveness of automation processes.
Thus, task analysis of nuanced task is challenging an open area of research. Furthermore execution of the analyzed task with AI augmentation, specifically GenAI augmentation, to bring in effective, efficient automation is critical and needs to be explored.
Embodiments of the present disclosure present technological improvements as solutions to one or more of the above-mentioned technical problems recognized by the inventors in conventional systems.
For example, in one embodiment, a method for selection-optimization-fine-tuning of Artificial Intelligence (AI) models for performing a task is provided. The method includes analyzing, by using a Logic and Rule Integrated-Decoder Neural Network (LRI-DecNN), a task received in natural language to obtain a contextual output for the task comprising a complexity (C), a clarity of objectives (O) and an augmentation potential (A) of the task. A self-attention mechanism of the LRI-DecNN is modified to integrate a logic and rule factor (B) enabling the LRI-DecNN to pay attention to a set of predefined logic and rules.
Further, the method includes assessing by the LRI-DecNN executed, suitability of the task for Artificial Intelligence (AI) augmentation based on value of a score function(S) derived from the complexity (C), the clarity of objectives (O) and the augmentation potential (A). Further, the method includes translating the task into an agent chain comprising a set of chained subtasks using a meta-mixture of experts technique, wherein the meta-mixture of experts technique utilizes a routing function (R) that evaluates requirements of the task and matches the requirements with the strengths and specializations of each expert model amongst the set of expert models of the meta-mixture of experts technique. Furthermore, the method includes determining an AI model among a set of AI models for each subtask among the set of chained subtasks based on a statistical function for Blackbox estimation in AI augmentation. Further, the method includes fine-tuning each AI model identified for each subtask among the set of chained subtasks using a KL divergence to align output distribution of each AI model with an expected distribution of the task, wherein fine-tuning involves adjusting training data, modifying the AI model architecture, or tweaking the learning process of the AI model to minimize the KL divergence.
Furthermore, the method includes determining an effective optimization technique for each AI model in accordance with the subtask using a Q-learning model to optimize each AI model. Further, the method includes tuning a set of hyperparameters of each AI model among the set of AI models. An Actor-Critic framework is integrated to refine Q-values, in real time, obtained from previous reinforcement learning steps of the Q-learning, and wherein an actor proposes actions based on a current policy, while a critic evaluates the actions by estimating a value function, leading to more precise Q-value adjustments. During inferencing, the hyperparameter tuned set of AI models identified for each subtask combinedly perform the task for an enterprise, and wherein the task is a nuanced task.
In another aspect, a system for selection-optimization-fine-tuning of Artificial Intelligence (AI) models for performing a task is provided. The system comprises a memory storing instructions; one or more Input/Output (I/O) interfaces; and one or more hardware processors coupled to the memory via the one or more I/O interfaces, wherein the one or more hardware processors are configured by the instructions to analyze, by using a Logic and Rule Integrated-Decoder Neural Network (LRI-DecNN), a task received in natural language to obtain a contextual output for the task comprising a complexity (C), a clarity of objectives (O) and an augmentation potential (A) of the task. A self-attention mechanism of the LRI-DecNN is modified to integrate a logic and rule factor (B) enabling the LRI-DecNN to pay attention to a set of predefined logic and rules.
Further, the one or more hardware processors are configured to assess by the LRI-DecNN executed, suitability of the task for Artificial Intelligence (AI) augmentation based on value of a score function(S) derived from the complexity (C), the clarity of objectives (O) and the augmentation potential (A). Further, the one or more hardware processors are configured to translate the task into an agent chain comprising a set of chained subtasks using a meta-mixture of experts technique, wherein the meta-mixture of experts technique utilizes a routing function (R) that evaluates requirements of the task and matches the requirements with the strengths and specializations of each expert model amongst the set of expert models of the meta-mixture of experts technique. Furthermore, the one or more hardware processors are configured to determine an AI model among a set of AI models for each subtask among the set of chained subtasks based on a statistical function for Blackbox estimation in AI augmentation. Further, the one or more hardware processors are configured to fine-tune each AI model identified for each subtask among the set of chained subtasks using a KL divergence to align output distribution of each AI model with an expected distribution of the task, wherein fine-tuning involves adjusting training data, modifying the AI model architecture, or tweaking the learning process of the AI model to minimize the KL divergence.
Furthermore, Further, the one or more hardware processors are configured to determine an effective optimization technique for each AI model in accordance with the subtask using a Q-learning model to optimize each AI model. Further, Further, the one or more hardware processors are configured to tune a set of hyperparameters of each AI model among the set of AI models. An Actor-Critic framework is integrated to refine Q-values, in real time, obtained from previous reinforcement learning steps of the Q-learning, and wherein an actor proposes actions based on a current policy, while a critic evaluates the actions by estimating a value function, leading to more precise Q-value adjustments. During inferencing, the hyperparameter tuned set of AI models identified for each subtask combinedly perform the task for an enterprise, and wherein the task is a nuanced task.
In yet another aspect, there are provided one or more non-transitory machine-readable information storage mediums comprising one or more instructions, which when executed by one or more hardware processors causes a method for selection-optimization-fine-tuning of Artificial Intelligence (AI) models for performing a task.
The method includes analyzing, by using a Logic and Rule Integrated-Decoder Neural Network (LRI-DecNN), a task received in natural language to obtain a contextual output for the task comprising a complexity (C), a clarity of objectives (O) and an augmentation potential (A) of the task. A self-attention mechanism of the LRI-DecNN is modified to integrate a logic and rule factor (B) enabling the LRI-DecNN to pay attention to a set of predefined logic and rules.
Further, the method includes assessing by the LRI-DecNN executed, suitability of the task for Artificial Intelligence (AI) augmentation based on value of a score function(S) derived from the complexity (C), the clarity of objectives (O) and the augmentation potential (A). Further, the method includes translating the task into an agent chain comprising a set of chained subtasks using a meta-mixture of experts technique, wherein the meta-mixture of experts technique utilizes a routing function (R) that evaluates requirements of the task and matches the requirements with the strengths and specializations of each expert model amongst the set of expert models of the meta-mixture of experts technique. Furthermore, the method includes determining an AI model among a set of AI models for each subtask among the set of chained subtasks based on a statistical function for Blackbox estimation in AI augmentation. Further, the method includes fine-tuning each AI model identified for each subtask among the set of chained subtasks using a KL divergence to align output distribution of each AI model with an expected distribution of the task, wherein fine-tuning involves adjusting training data, modifying the AI model architecture, or tweaking the learning process of the AI model to minimize the KL divergence.
Furthermore, the method includes determining an effective optimization technique for each AI model in accordance with the subtask using a Q-learning model to optimize each AI model. Further, the method includes tuning a set of hyperparameters of each AI model among the set of AI models. An Actor-Critic framework is integrated to refine Q-values, in real time, obtained from previous reinforcement learning steps of the Q-learning, and wherein an actor proposes actions based on a current policy, while a critic evaluates the actions by estimating a value function, leading to more precise Q-value adjustments. During inferencing, the hyperparameter tuned set of AI models identified for each subtask combinedly perform the task for an enterprise, and wherein the task is a nuanced task.
It is to be understood that both the foregoing general description and the following detailed description are exemplary and explanatory only and are not restrictive of the invention, as claimed.
The accompanying drawings, which are incorporated in and constitute a part of this disclosure, illustrate exemplary embodiments and, together with the description, serve to explain the disclosed principles:
It should be appreciated by those skilled in the art that any block diagrams herein represent conceptual views of illustrative systems and devices embodying the principles of the present subject matter. Similarly, it will be appreciated that any flow charts, flow diagrams, and the like represent various processes which may be substantially represented in computer readable medium and so executed by a computer or processor, whether or not such computer or processor is explicitly shown.
Exemplary embodiments are described with reference to the accompanying drawings. In the figures, the left-most digit(s) of a reference number identifies the figure in which the reference number first appears. Wherever convenient, the same reference numbers are used throughout the drawings to refer to the same or like parts. While examples and features of disclosed principles are described herein, modifications, adaptations, and other implementations are possible without departing from the scope of the disclosed embodiments.
Nuanced enterprise tasks can be defined as tasks that are complex, ambiguous, heterogenous, and the ones that require highly skilled human beings to understand and execute successfully. The current landscape of automating complex, highly skilled tasks using Generative Artificial Intelligence (AI) presents several significant challenges for enterprises as mentioned below:
Embodiments of the present disclosure provide a method and system for selection-optimization-fine-tuning of Artificial Intelligence (AI) models for nuanced task by analyzing the nuanced task using Logic and Rule Integrated Decoder Neural Network (LRI-DecNN). The system and method alleviate the issues of knowledge concentration in enterprises by automating tasks characterized by high variability. It employs advanced generative AI algorithms to interpret, manage, and execute a range of tasks that traditionally demanded human discretion due to their inherent variability and complexity.
The system receives a natural language task and performs task analysis using the LRI-DecNN or the Decoder-only Transformer (GenAI) model, which has an architecture that integrates Business requirements of an enterprise into the task analysis by the LRI-DecNN. Further, suitability of the analyzed task for AI augmentation, the task is analyzed using a score function based on its characteristics and requirements of the task. Accordingly, the system automatically constructs an agent chain using a meta-mixture technique, where each agent is responsible for a specific part of the task (subtask), such as data gathering, analysis, or decision-making. The system then selects the most appropriate generative AI model from its repository, considering factors like task complexity, data type, and desired outcome. It fine-tunes the model using advanced techniques to ensure optimal performance. Actor-Critic framework is integrated to refine Q-values, in real time, obtained from previous reinforcement learning steps of Q-learning applied during model optimization, wherein an actor proposes actions based on the current policy, while a critic evaluates the actions by estimating a value function, leading to more precise Q-value adjustments.
Once set up, the agent chain executes the task, with each agent performing its designated function. The system continuously monitors performance, making real-time adjustments to the agent chain and model parameters for maximum efficiency and accuracy.
The AI models identified for each subtask can be Generative AI models such as Generative Adversarial Networks (GANs) that can create visual and multimedia artifacts from both imagery and textual input data, Transformer-based models such as Generative Pre-Trained (GPT) language models that can use information gathered and create textual content, and the like.
Listed below are example tasks that are effectively and efficiently augmented with AI or GenAI via the system.
Referring now to the drawings, and more particularly to
In an embodiment, the system 100 includes a processor(s) 104, communication interface device(s), alternatively referred as input/output (I/O) interface(s) 106, and one or more data storage devices or a memory 102 operatively coupled to the processor(s) 104. The system 100 with one or more hardware processors is configured to execute functions of one or more functional blocks of the system 100.
Referring to the components of system 100, in an embodiment, the processor(s) 104, can be one or more hardware processors 104. In an embodiment, the one or more hardware processors 104 can be implemented as one or more microprocessors, microcomputers, microcontrollers, digital signal processors, central processing units, state machines, logic circuitries, and/or any devices that manipulate signals based on operational instructions. Among other capabilities, the one or more hardware processors 104 are configured to fetch and execute computer-readable instructions stored in the memory 102. In an embodiment, the system 100 can be implemented in a variety of computing systems including laptop computers, notebooks, hand-held devices such as mobile phones, workstations, mainframe computers, servers, and the like.
The I/O interface(s) 106 can include a variety of software and hardware interfaces, for example, a web interface, a graphical user interface for interactive exchange and human feedback to automatically continually fine tune the AL models and re-selection of AI models for subtasks. The user interface also enables receiving the natural language task for which the AI models have to be selected and fine-tuned. The I/O interface 106 can facilitate multiple communications within a wide variety of networks N/W and protocol types, including wired networks, for example, LAN, cable, etc., and wireless networks, such as WLAN, cellular and the like. In an embodiment, the I/O interface(s) 106 can include one or more ports for connecting to a number of external devices or to another server or devices.
The memory 102 may include any computer-readable medium known in the art including, for example, volatile memory, such as static random access memory (SRAM) and dynamic random access memory (DRAM), and/or non-volatile memory, such as read only memory (ROM), erasable programmable ROM, flash memories, hard disks, optical disks, and magnetic tapes.
In an embodiment, the memory 102 includes a plurality of modules 110 such as the LRI-DecNN, the scoring module implementing a scoring function, the meta-mixture of experts module implementing a meta-mixture of experts technique, Blackbox estimation module, and the like as depicted in system architectural overview of
The plurality of modules 110 include programs or coded instructions that supplement applications or functions performed by the system 100 for executing different steps involved in the process of selection-optimization-fine-tuning of Artificial Intelligence (AI) models for nuanced task by analyzing the nuanced task using Logic and Rule Integrated Decoder Neural Network (LRI-DecNN), being performed by the system 100. The plurality of modules 110, amongst other things, can include routines, programs, objects, components, and data structures, which perform particular tasks or implement particular abstract data types. The plurality of modules 110 may also be used as, signal processor(s), node machine(s), logic circuitries, and/or any other device or component that manipulates signals based on operational instructions. Further, the plurality of modules 110 can be used by hardware, by computer-readable instructions executed by the one or more hardware processors 104, or by a combination thereof. The plurality of modules 110 can include various sub-modules (not shown).
Further, the memory 102 may comprise information pertaining to input(s)/output(s) of each step performed by the processor(s) 104 of the system 100 and methods of the present disclosure.
Further, the memory 102 includes a database 108. The database (or repository) 108 may include a plurality of abstracted pieces of code for refinement and data that is processed, received, or generated as a result of the execution of the plurality of modules in the module(s) 110. Further, the database can have a repository of AI/GENAI models, from which the system can select the most appropriate model for a specific subtask.
Although the database 108 is shown internal to the system 100, it will be noted that, in alternate embodiments, the database 108 can also be implemented external to the system 100, and communicatively coupled to the system 100. The data contained within such an external database may be periodically updated. For example, new data may be added into the database (not shown in
In an embodiment, the system 100 comprises one or more data storage devices or the memory 102 operatively coupled to the processor(s) 104 and is configured to store instructions for execution of steps of the method 200 by the processor(s) or one or more hardware processors 104. The steps of the method 200 of the present disclosure will now be explained with reference to the components or blocks of the system 100 as depicted in
Referring to the steps of the method 200, at step 202 of the method 200, the one or more hardware processors 104 are configured by the instructions to analyze, by the LRI-DecNN, a task received in natural language to obtain a contextual output for the task comprising a complexity (C), a clarity of objectives (O) and an augmentation potential (A) of the task. A self-attention mechanism of the LRI-DecNN is modified to integrate a logic and rule factor (B) enabling the LRI-DecNN to pay attention to a set of predefined logic and rules. The task is received through the UI (I/O interface 106) or an Application Programming Interface (API). This could be a complex business problem, or a process automation requirement provided in natural language. The received natural language input is parsed, with understanding of structured and unstructured data, with help of robust data ingestion and preprocessing capabilities, known in the art. The task analysis is performed by the LRI-DecNN.
The LRI-DecNN model is distinct from traditional Decoder Transformer models, which typically use a combination of encoder and decoder components. The decoder-only architecture is specifically designed to generate outputs based on a given context, making it highly effective for tasks that require a deep understanding of language and the generation of human-like responses. As depicted in the architecture of the decoder disclosed (LRI-DecNN) in
The LRI-DecNN operates by processing the input sequence (the task description) and generating a contextually relevant output. It does this by predicting the next word or sequence of words based on the input it has received. Unlike encoder-decoder models, which first encode an input sequence into a fixed representation and then decode it into an output, the decoder-only model continuously refines its understanding of the input as it generates the output. This allows for a more dynamic and nuanced interpretation of the task. The model leverages advanced techniques such as attention mechanisms, which enable it to focus on different parts of the input text and understand how these parts relate to each other. This results in a more coherent and contextually accurate understanding of the task.
Integration of Business Logic and Rules: The LRI-DecNN decoder-only model is enhanced by integrating business logic and rules as an additional input to the attention mechanism. This integration allows the model to align its outputs with specific business practices and regulatory requirements, providing a more nuanced understanding of complex enterprise tasks.
Revised Mathematical Notation for the attention mechanism of the LRI-DecNN: As depicted in
As mentioned earlier, it is critical to analyze whether the effort, cost, time etc. that goes into automation with AI augmentation, is worth the returns. Thus, at step 204 of the method 200, the one or more hardware processors 104 are configured by the instructions to assess suitability of the task for Artificial Intelligence (AI) augmentation based on value of a score function(S) derived from the complexity (C), the clarity of objectives (O) and the augmentation potential (A).
Understanding of task analysis performed by the system 100 for a Nuance task example: Automation of Quality Control in a Manufacturing Plant:
In this scenario, the LRI-DecNN with the self-attention mechanism integrating (B) would be tailored to analyze images of manufactured products, paying attention to predefined criteria and rules of the enterprise for quality assessment. The integration of these elements—Complexity (C), Clarity of Objectives (O), Augmentation Potential (A), and Logic and Rule Factor (B)—ensures a comprehensive and effective approach to automating the quality control process in a manufacturing setting.
Task Suitability Assessment: The step 204 above involves evaluating the complexity, clarity of objectives, and the feasibility of automating the task. The assessment leverages the LRI-DEC NN's analysis, using Natural Language Processing (NLP) techniques to gauge the task's intricacies and alignment with AI capabilities. In the context of personalized augmentation, the task suitability assessment determines whether a task is appropriate for augmentation with GenAI, specifically focusing on enhancing human capabilities rather than full automation. This involves evaluating the task's complexity, the clarity of its objectives, and the potential for GenAI to augment human performance. The assessment leverages the decoder-only model's analysis, using advanced NLP techniques to gauge the task's intricacies and its alignment with GenAI capabilities for augmentation. The decoder-only model examines factors such as the level of human expertise required, the potential for GenAI to provide meaningful support, and the integration of GenAI into human workflows.
The score function(S) evaluates various aspects of the task in terms of augmentation potential. For instance, it can assign scores based on the potential for GenAI to enhance human decision-making, the complexity of the task that can be simplified by GenAI, and the integration of GenAI insights into human-driven processes. The scoring function can be represented as S=f(C, O, A) as mentioned above. Each of these factors is evaluated and scored, with the total score(S) determining the task's suitability for augmentation with AI or GenAI models.
Example of Task that can be ideal candidate for Personalized Augmentation with GenAI: A task like high-level strategic decision-making, such as developing a new business strategy based on emerging market trends, might score high in augmentation potential due to the complexity and the value GenAI can add in providing data-driven insights and scenario analysis.
Example of Task that may not be suitable candidate for Personalized Augmentation with GenAI: A rule-based task such as data entry, which is straightforward and does not benefit significantly from GenAI augmentation, would likely score low in augmentation potential.
The range for high and low values of S, the augmentation suitability score in the context of assessing a natural language task for AI or GenAI augmentation, typically depends on how the scoring function is calibrated. However, without specific calibration details, a general approach can be considered:
The meta-learner, a sophisticated component of the system 100, is responsible for learning and adjusting the weights applied to complexity (C), clarity of objectives (O), and augmentation potential (A) in the scoring formula. This learning process is conducted through the analysis of a wide range of tasks and their outcomes post-GenAI augmentation, allowing the meta-learner to discern patterns and relationships between task characteristics and successful augmentation.
At step 206 of the method 200, the one or more hardware processors 104 are configured by the instructions to translate the task into an agent chain comprising set of chained subtasks using a meta-mixture of experts technique. The meta-mixture of experts technique utilizes a routing function (R) that evaluates requirements of the task and matches the requirements with the strengths and specializations of each expert model AI model among the set of expert models of the meta-mixture of experts technique.
Task translation into agent chain (chained or linked subtasks) using meta-mixture of experts Technique: As mentioned above, this technique involves expert choice and routing among various task-centric LLMs to create optimized agent chain for the specific task. It is a modified form of ensemble learning where instead of multiple neuron groupings within a large neural network, multiple LLMs (experts) as a whole are used as experts, and decisions on which expert to use at what time are based on the task requirements. This ensures a more robust and flexible approach to task execution. The assessed tasks are translated into a series of agent chain using the meta-mixture of experts technique, specifically for the purpose of augmenting human capabilities with GenAI. This technique leverages the expertise of multiple task-centric LLMs to support and enhance human tasks with GenAI. It represents a modified form of ensemble learning tailored for augmentation with GenAI. Instead of using multiple neuron groupings within a single large neural network, the system 100 employs multiple LLMs (experts) in their entirety. Decisions on which expert to use and at what time are based on the specific requirements of the task and how best GenAI can augment human performance. The meta-mixture of experts Technique involves dynamically routing tasks to the most suitable LLMs based on their expertise in augmenting human tasks with GenAI. This is achieved through a routing algorithm that evaluates the task requirements and matches them with the strengths and specializations of available LLMs in terms of augmentation with GenAI.
The choice routing algorithm can be represented as be a routing function®, mathematically expressed as: (R(T, Ei where i=1 to n)), where (T) is the task, and (E) is the set of available experts {1 . . . to n} (decoder only Large Language Models (LLMs)) for augmentation with GenAI. The function (R) evaluates the suitability of each expert for augmenting the task with GenAI and routes the task accordingly. The meta-mixture of experts technique for augmentation with GenAI thus uses multiple decoder only LLMs are used as distinct experts, each contributing their specialized knowledge to augment human tasks with GenAI. This approach not only enhances the flexibility and effectiveness of human task execution but also demonstrates a unique and non-obvious solution to the challenge of augmenting complex, nuanced tasks in an enterprise setting with GenAI.
The combination of task sustainability assessment using the score function(S) and the meta-mixture of experts technique applied for task disintegration into meaningful subtasks provides a deep integration of GenAI technology with human workflows, setting a new standard in the augmentation of skilled tasks with GenAI and representing a clear leap over existing solutions in the market.
Upon obtaining agent chain for the task at step 206, at step 208 of the method 200, the one or more hardware processors 104 are configured by the instructions to determine an AI model among a set of AI models (available in the repository or database 108) for each subtask among the set of chained subtasks based on a statistical function for Blackbox estimation in AI augmentation.
Best AI Model Recommendation in accordance with the subtask: Model analysis with Blackbox estimation: Blackbox performance estimation is a technique where the internal workings of the model are not necessarily understood in detail, but the output performance is analyzed. This method is used to estimate how different AL models or GenAI models might perform on the given task. It involves statistical analysis, predictive modeling, and sometimes machine learning algorithms to forecast model performance without delving into the model's internal mechanisms. Based on the Blackbox analysis, the system 100 selects the most suitable AI models (includes Gen AI models). This decision is crucial as it determines the effectiveness and efficiency of the task automation. The selection process may involve matching the task's requirements with the models' known capabilities, analyzing historical performance data, and considering the computational efficiency of each model. The system 100 collects data on various GenAI models' past performances, including metrics like accuracy, efficiency, and context relevance. This data forms the basis of the Blackbox estimation. Machine learning algorithms, particularly those adept at pattern recognition and prediction, analyze this data to forecast the performance of each GenAI model on the new task. The algorithms might use features like task complexity, required expertise level, and historical success rates of models in similar tasks. The system 100 then generates a performance score or probability estimate for each GenAI model, indicating its suitability for the task at hand.
Role of GenAI: In the system 100, the GenAI is not limited to task suitability assessment but plays a pivotal role in the entire process of task analysis, model selection, and execution. GenAI models are central to the system's capability to handle complex, nuanced tasks that traditional AI or ML models might struggle with. The importance of GenAI lies in its ability to generate, understand, and process human-like language and concepts, making it ideal for tasks that require a high degree of cognitive understanding and creativity. This is in contrast to typical ML models, which might be more suited for tasks with clear-cut patterns and predictable outcomes.
Creating a statistical function for Blackbox estimation, that is unique to GenAI task augmentation effectiveness estimation involves designing a model that can predict the performance of a GenAI model on a specific task. This function would need to consider various factors that influence the effectiveness of GenAI in augmenting tasks. Provided below is an example of how such a function might be structured:
Function definition: Let a function ‘F’ be defined that estimates the effectiveness of a GenAI model for a given task. This function takes a set of input parameters and outputs a score representing the predicted effectiveness.
Function Structure: The function ‘F’ can be represented as a weighted sum of these parameters, each with its own coefficient that signifies its importance:
where ‘w_1’, ‘w_2’, ‘w_3’, ‘w_4’, and ‘w_5’ are the weights assigned to each parameter.
Example Calculation: Suppose a task with the following parameter values is present:
Then the effectiveness score is calculated as:
Interpretation: A score of 0.79 (out of 1) indicates a high level of predicted effectiveness for the GenAI model on this particular task. The higher the score, the more suitable the GenAI model is deemed for the task. This statistical function represents an approach to estimating the effectiveness of GenAI models in task augmentation. It uniquely combines various critical factors, including task complexity, historical performance data, data availability, model adaptability, and task-model alignment, to provide a comprehensive and predictive score. The use of weighted parameters allows for customization based on the specific requirements and priorities of different tasks and industries. This method offers a quantifiable, data-driven way to assess the suitability of GenAI models for complex tasks, enhancing decision-making processes in enterprise environments.
The use of Blackbox performance estimation by the system 100 for AI model or the GenAI models identified for performing each of the subtasks is a significant advancement in the field of enterprise task automation and augmentation. Unlike traditional approaches that rely heavily on understanding the internal mechanisms of each model, the Blackbox optimization disclosed herein allows for a rapid, efficient, and highly effective estimation of a GenAI model's suitability for complex tasks. The integration of advanced statistical and machine learning techniques for predictive modeling ensures that the system can accurately forecast the performance of GenAI models, even in scenarios where the tasks are highly nuanced and require a sophisticated level of understanding. This approach not only streamlines the model selection process but also ensures that the chosen AL/GenAI model is optimally aligned with the task's specific requirements. The incorporation of GenAI across the system, far beyond mere task assessment, sets the system apart from existing solutions and underscoring its potential to revolutionize how enterprises approach task automation and augmentation with GenAI.
Once best AI/GenAI models are identified in accordance with the subtask at step 206, then at step 210 of the method 200, the one or more hardware processors 104 are configured by the instructions to fine tune each AI model identified for each subtask among the set of chained subtasks using a Kullback-Leibler (KL) divergence to align output distribution of each AI model with an expected distribution of the task. As known in the art KL divergence is measure the difference between two probability distributions over the same variable x, which the system applies in the domain of AI model fine tuning for the task. Fine tuning involves adjusting training data, modifying the AI model architecture, or tweaking the learning process of the AI model to minimize the KL divergence.
Data and task estimation for fine-tuning: This sub-step involves analyzing the task and available data to determine the best approach for fine-tuning the AI model. It uses the concept of KL divergence, a statistical measure used in information theory, to estimate the difference between two probability distributions. In this context, it helps in understanding how the model's output distribution deviates from the expected results.
Fine-Tuning strategy formulation using KL Divergence: Objective: The goal here is to align the GenAI model's output distribution as closely as possible with the expected distribution for a given task. This alignment is crucial for the model to perform effectively in the specific context of the task.
Inform fine-tuning strategy: Based on the KL divergence results, a strategy for fine-tuning is developed for the model. This might involve adjusting the training data, modifying the model architecture, or tweaking the learning process. The strategy aims to minimize the KL divergence, thereby aligning the model's output more closely with the expected distribution.
Example Scenario: Consider a scenario where a GenAI model is being fine-tuned for financial insights generation. The expected distribution emphasizes accuracy in financial terminology, coherence in economic analysis, and a formal tone. The current model's output, however, shows a tendency towards general news style and lacks depth in financial analysis. By calculating the KL divergence, this misalignment is quantified. The fine-tuning strategy then focuses on incorporating more financial data, adjusting the model to prioritize financial terms, and retraining it to reduce the KL divergence, thereby aligning the model's output with the desired financial news characteristics.
This approach represents a significant advancement in GenAI model fine-tuning, offering a systematic, data-driven method to align models with specific task requirements. The use of KL divergence provides a quantifiable measure of alignment, enabling precise adjustments and improvements in model performance. This method is particularly valuable in domains where accuracy and specificity are critical, and it demonstrates an application of information theory in the realm of GenAI.
Fine-tuning strategy decision point: The system 100 then decides on the fine-tuning strategies. This could involve choosing between various approaches like supervised learning, unsupervised learning, or transfer learning based on the task requirements and data characteristics. This decision is informed by the previous KL divergence analysis, ensuring that the fine-tuning strategy aligns the model's outputs closely with the desired outcomes.
Fine-tuning strategy decision point with focus on autoencoders: In the context of fine-tuning Generative AI (GenAI) models for tasks with scarce or unavailable labeled data, autoencoders emerge as a powerful unsupervised learning strategy. Autoencoders are neural networks designed to learn efficient representations (encodings) of the input data, typically for the purpose of dimensionality reduction or feature learning.
At step 212 of the method 200, the one or more hardware processors 104 are configured by the instructions to determining an effective optimization technique for each AI model in accordance with the subtask using a Q-learning model to optimize each AI model.
At step 214 of the method 200, the one or more hardware processors 104 are configured by the instructions to tune a set of hyperparameters of each AI model among the set of AI models. This involves using advanced algorithms to suggest the best settings for these parameters. The decision is based on an analysis of how different hyper-parameter settings impact the model's performance. Techniques like grid search, random search, or Bayesian optimization are commonly used in this process.
An Actor-Critic framework is integrated to refine Q-values, in real time, obtained from previous reinforcement learning steps of the Q-learning, wherein an actor proposes actions based on the current policy, while a critic evaluates the actions by estimating a value function, leading to more precise Q-value adjustments.
The AI model, with its fine-tuning strategies, optimization methods, and hyper-parameters, is deployed for the task. Technical aspect: This step involves setting up the model in a real-world environment and ensuring it integrates smoothly with existing systems and workflows.
Performance monitoring and data collection: The system continuously monitors the model's performance during task execution. This involves collecting data on various performance metrics, which can be used for further optimization and improvements.
The Meta Learner Reinforcement of the system 100 employs an Actor-Critic (AAC) approach to optimize the Q-values learned in earlier steps, such as through Q-learning in model optimization processes. This advanced method is instrumental in refining and enhancing the decision-making capabilities of the system, particularly in optimizing Generative AI (GenAI) models for complex tasks.
The implementation of the Actor-Critic architecture for optimizing Q-values in the system 100 is a sophisticated advancement in GenAI model optimization. This method allows for dynamic adaptation of strategies based on real-world performance feedback. The integration of a pre-learned Q-table with the Actor-Critic model enables the system to utilize prior learning while continuously refining its approach through the Critic's evaluations and the Actor's policy updates. The use of Temporal Difference (TD) error as a basis for the Critic's assessment and Actor's policy gradient calculations ensures that the model not only leverages historical data but also adapts to new scenarios, making the optimization process more effective and robust. This application of reinforcement learning principles in optimizing Q-values reflects a significant technological advancement, offering a nuanced, effective, and adaptive approach to GenAI model optimization.
This approach not only optimizes the GenAI model's performance for specific tasks but also ensures continual learning and adaptation, making it a robust and dynamic solution for complex AI applications.
Ongoing Adaptation: The system 100 continually refines its analyses, model recommendations, and optimization strategies based on real-time learning and feedback. This step embodies the concept of continuous learning in AI, where the system evolves and improves over time, adapting to new data, changing conditions, and feedback. The system represents a significant leap in the field of enterprise AI, offering a flexible, and highly effective solution for automating complex, nuanced tasks, and democratizing SME knowledge within organizations. Its unique combination of automated agent chain generation, advanced model recommendation, and fine-tuning, along with an API-driven configurator, positions it as a ground-breaking tool in the realm of generative AI applications in enterprise settings.
Thus, the method and system herein provide:
Each of these advancements in the system 100 disclosed herein represents a significant progression in the field of enterprise GenAI automation. They not only enhance the efficiency and effectiveness of GenAI applications in complex tasks but also expand the realm of possibilities for automation in areas previously considered beyond the reach of existing GenAI technologies.
The written description describes the subject matter herein to enable any person skilled in the art to make and use the embodiments. The scope of the subject matter embodiments is defined by the claims and may include other modifications that occur to those skilled in the art. Such other modifications are intended to be within the scope of the claims if they have similar elements that do not differ from the literal language of the claims or if they include equivalent elements with insubstantial differences from the literal language of the claims.
It is to be understood that the scope of the protection is extended to such a program and in addition to a computer-readable means having a message therein; such computer-readable storage means contain program-code means for implementation of one or more steps of the method, when the program runs on a server or mobile device or any suitable programmable device. The hardware device can be any kind of device which can be programmed including e.g. any kind of computer like a server or a personal computer, or the like, or any combination thereof. The device may also include means which could be e.g. hardware means like e.g. an application-specific integrated circuit (ASIC), a field-programmable gate array (FPGA), or a combination of hardware and software means, e.g. an ASIC and an FPGA, or at least one microprocessor and at least one memory with software processing components located therein. Thus, the means can include both hardware means, and software means. The method embodiments described herein could be implemented in hardware and software. The device may also include software means. Alternatively, the embodiments may be implemented on different hardware devices, e.g. using a plurality of CPUs.
The embodiments herein can comprise hardware and software elements. The embodiments that are implemented in software include but are not limited to, firmware, resident software, microcode, etc. The functions performed by various components described herein may be implemented in other components or combinations of other components. For the purposes of this description, a computer-usable or computer readable medium can be any apparatus that can comprise, store, communicate, propagate, or transport the program for use by or in connection with the instruction execution system, apparatus, or device.
The illustrated steps are set out to explain the exemplary embodiments shown, and it should be anticipated that ongoing technological development will change the manner in which particular functions are performed. These examples are presented herein for purposes of illustration, and not limitation. Further, the boundaries of the functional building blocks have been arbitrarily defined herein for the convenience of the description. Alternative boundaries can be defined so long as the specified functions and relationships thereof are appropriately performed. Alternatives (including equivalents, extensions, variations, deviations, etc., of those described herein) will be apparent to persons skilled in the relevant art(s) based on the teachings contained herein. Such alternatives fall within the scope of the disclosed embodiments. Also, the words “comprising,” “having,” “containing,” and “including,” and other similar forms are intended to be equivalent in meaning and be open ended in that an item or items following any one of these words is not meant to be an exhaustive listing of such item or items, or meant to be limited to only the listed item or items. It must also be noted that as used herein and in the appended claims, the singular forms “a,” “an,” and “the” include plural references unless the context clearly dictates otherwise.
Furthermore, one or more computer-readable storage media may be utilized in implementing embodiments consistent with the present disclosure. A computer-readable storage medium refers to any type of physical memory on which information or data readable by a processor may be stored. Thus, a computer-readable storage medium may store instructions for execution by one or more processors, including instructions for causing the processor(s) to perform steps or stages consistent with the embodiments described herein. The term “computer-readable medium” should be understood to include tangible items and exclude carrier waves and transient signals, i.e., be non-transitory. Examples include random access memory (RAM), read-only memory (ROM), volatile memory, nonvolatile memory, hard drives, CD ROMs, DVDs, flash drives, disks, and any other known physical storage media.
It is intended that the disclosure and examples be considered as exemplary only, with a true scope of disclosed embodiments being indicated by the following claims.
| Number | Date | Country | Kind |
|---|---|---|---|
| 202321087106 | Dec 2023 | IN | national |