The present disclosure generally relates to network computing and more particularly to real-time searches over the network.
Large language models (LLM) are rapidly transforming the state of the art for computing, including network searching. However, the amount of processing complexity for and LLM is staggering and can lead to hallucination and/or extraordinary intensive computational and network requirements.
An aspect of the specification provides a method for real-time search including: configuring a large language model (LLM) engine with a context shift using a plurality of specific contextualization objects within a hierarchy dependent from a general contextualization object; receiving, at a natural language processing (NLP) engine, an input message; forwarding the input message from the collaboration platform to the LLM engine; determining, at the LLM engine, that the input message includes an unstructured query corresponding to one of the specific contextualization objects based on the general context shift and the unstructured query; the number of tokens in the determining being less than if the determining is based on a non-hierarchical contextualization object; preparing, at the LLM engine, a draft response message including a structured query to at least one of a plurality of search engines corresponding to the one of the specific contextualization objects; forwarding the draft response message from the LLM engine to the collaboration platform; sending the structured query from the collaboration platform to a management engine for routing and processing by the at least one of a plurality of search engines; receiving, at the collaboration platform, a response to the structured query, from the search engine via the management engine; and, generating an output message responsive to the input message including the draft response message that substitutes the response for the structured query.
An aspect of the specification provides a method wherein the general contextualization object is based on home renovation planning and configures the LLM engine to classify the unstructured query into at least one of a materials query, a labour query, a delivery query and a general query.
An aspect of the specification provides a method wherein the general contextualization object is based on travel searching and configures the LLM engine to classify the unstructured query into at least one of an air-search query, an air-policy query, a general query, an events query and a ground transportation itinerary search query; each of the queries corresponding to one or more of the search engines.
An aspect of the specification provides a method wherein a subsequent unstructured query builds on a previous response; and wherein a different specific contextualization object is chosen for the subsequent query than for the original unstructured query.
An aspect of the specification provides a method wherein the real-time search is a travel query and the search engines are travel actor engines; the travel query includes a transportation-actor component and a hospitality-actor component and the transportation-actor component is respective to at least one travel actor engine and the hospitality-actor component is respective to another at least one travel actor engine.
An aspect of the specification provides a method wherein the travel query includes a transportation-actor component that is restricted by an employer policy component.
An aspect of the specification provides a method wherein the employer policy component corresponds to an employer policy search engine that maintains restrictions as to types of queries to the transportation-actor search engines and the hospitality-actor search engines; the restrictions based on an account from which the input message originates.
An aspect of the specification provides a method wherein the travel query implies a coordination between travel-actors such that the results are responsively filtered by the coordination.
An aspect of the specification provides a method wherein the coordination is based on aligning a flight schedule with an availability of a ground-transportation service and accommodation.
An aspect of the specification provides a method wherein the travel query includes one or more travel-actors including: transportation-actors including airlines, rail services, bus lines and ferry lines; hospitality-actors including hotels, resorts and bed and breakfasts; for-hire ground-transportation actors including car-rentals, taxis and car sharing; and dining-actors including restaurants, bistros and bars.
An aspect of the specification provides a method wherein the input message and output message are incorporated into a collaboration tool executing on a collaboration platform that hosts the NLP engine.
An aspect of the specification provides a method wherein the collaboration tool is a social media platform.
An aspect of the specification provides a method wherein the travel query includes an account profile of the user generating the input message.
An aspect of the specification provides a collaboration platform including a real time network search function based on natural language processing queries; the platform including a processor and a memory; the processor executing programming instructions for: configuring a large language model (LLM) engine with a context shift using a plurality of specific contextualization objects within a hierarchy dependent from a general contextualization object; receiving, at a natural language processing (NLP) engine, an input message; forwarding the input message from the collaboration platform to the LLM engine; determining, at the LLM engine, that the input message includes an unstructured query corresponding to one of the specific contextualization objects based on the general context shift and the unstructured query; the number of tokens in the determining being less than if the determining is based on a non-hierarchical contextualization object; preparing, at the LLM engine, a draft response message including a structured query to at least one of a plurality of search engines corresponding to the one of the specific contextualization objects; forwarding the draft response message from the LLM engine to the collaboration platform; sending the structured query from the collaboration platform to a management engine for routing and processing by the at least one of a plurality of search engines; receiving, at the collaboration platform, a response to the structured query, from the search engine via the management engine; and, generating an output message responsive to the input message including the draft response message that substitutes the response for the structured query.
An aspect of the specification provides a collaboration platform wherein the management engine is incorporated into the collaboration platform.
An aspect of the specification provides a collaboration platform wherein the LLM engine is incorporated into the collaboration platform.
An aspect of the specification provides a collaboration platform wherein the NLP engine is incorporated into the collaboration platform.
An aspect of the specification provides a collaboration platform wherein the NLP engine and LLM engine are combined into a single engine. An aspect of the present specification provides a method for real-time travel search comprising:
An aspect of the specification provides a method for real-time travel search including: configuring a large language model (LLM) engine with a real-time travel query context shift; receiving, at a natural language processing (NLP) engine, an input message; forwarding the input message from the collaboration platform to the LLM engine; determining, at the LLM engine, that the input message includes an unstructured travel query; preparing, at the LLM engine, a draft response message including a structured travel query based on the unstructured travel query; forwarding the draft response message from the LLM engine to the collaboration platform; sending the structured travel query from the collaboration platform to a travel management engine; receiving, at the collaboration platform, a travel itinerary responsive to the structured travel query, from the travel management engine; and, generating an output message responsive to the input message including the draft response message that substitutes the travel itinerary for the structured travel query.
An aspect of the specification provides a method wherein the travel query includes a transportation-actor component and a hospitality-actor component.
An aspect of the specification provides a method wherein the travel query includes a transportation-actor component that is restricted by an employer policy component.
An aspect of the specification provides a method wherein travel query implies a coordination between travel-actors such that the results are responsively filtered by the coordination.
An aspect of the specification provides a method wherein the coordination is based on aligning a flight schedule with an availability of a ground-transportation service and accommodation.
An aspect of the specification provides a method wherein the travel query includes one or more travel-actors including: transportation-actors including airlines, rail services, bus lines and ferry lines; hospitality-actors including hotels, resorts and bed and breakfasts; for-hire ground-transportation actors including car-rentals, taxis and car sharing; and dining-actors including restaurants, bistros and bars.
An aspect of the specification provides a method wherein the travel query includes an employer travel policy.
An aspect of the specification provides a method wherein the input message and output message are incorporated into a collaboration tool.
An aspect of the specification provides a method wherein the travel query includes an account profile of the user generating the input message.
The present specification also provides methods, apparatuses and computer-readable media according to the foregoing.
Thus, according to the illustrative embodiment, system 100 comprises a collaboration platform 104 connected to a network 108 such as the Internet. Network 108 interconnects collaboration platform 104 with: a) a plurality of travel actor engines 112; b) a plurality of client devices 116; d) a large language model (LLM) engine 120; and e) a travel management engine 122.
(Note that travel actor engines 112 are individually labelled as 112-1, 112-2 . . . 112-n. Collectively, they are referred to as travel actor engines 112, and generically, as travel actor engine 112. The nomenclature is used elsewhere such as for devices 116.)
Devices 116 are operated by individual users 124, each of which use a separate account 128 to access system 100. The present specification contemplates scenarios where, from time to time, users 124 may wish to search for travel itineraries available from one or more travel actors. Collaboration platform 104 performs a number of central processing functions to, amongst other things, manage generation of the travel itineraries by intermediating between devices 116 and engines 112. Collaboration platform 104 will be discussed in greater detail below.
Travel actor engines 112 are operated by different travel actors that provide travel services. Travel actors can include: transportation actors such as airlines, railways, bus companies, taxis, car services, public transit systems, cruise lines or ferry companies; accommodation actors such as hotels, resorts, and bed and breakfasts; hospitality actors such as restaurants, bars, pubs and bistros; and, event actors such as concert venues, theatres, galleries and conference venues. Other examples of travel actors will occur to those of skill in the art. Travel actor engines 112 can be based on a Global Distribution System (GDS) or the New Distribution Capability (NDC) protocol or other travel booking architectures or protocols that can arrange travel itineraries for users 124 with one or more travel actors. Travel actor engines 112 can thus be built on many different technological solutions and their implementation can be based on different distribution channels, including indirect channels such as GDS and/or direct channels like NDC hosted by individual travel actors such as airlines. Booking tools via various travel actor engines 112 can be also provided according to many solutions for different travel content distributors and aggregators including online and offline services such as travel agencies, metasearch tools, NDC, low cost carriers (“LCC”) and aggregators that sell airline seats, and the like. Travel actor engines 112 can be “white label” in that they are powered by travel technology companies such as Amadeus™ but branded by other entities, or they can be hosted directly by the travel operator such as an airline operating a particular airline transportation actor or a railway operating a particular railway transportation actor. One or more travel actor engines 112 may also manage accommodation, hospitality and/or event bookings. Travel actor engines 112 may also broadly include platforms or websites that include information about events that may impact travel, including disasters, airport delays, health warnings, severe weather, politics, sports, expos, concerts, festivals, performing arts, public holidays and acts of terrorism. Thus, travel actor engines 112 can even broadly encompass news and weather services.
Client devices 116 can be any type of human-machine interface for interacting with platforms 104. For example, client devices 116 can include traditional laptop computers, desktop computers, mobile phones, tablet computers and any other device that can be used to send and receive communications over network 108 and its various nodes that complement the input and output hardware devices associated with a given client device 116. It is contemplated client devices 116 can include virtual or augmented reality gear complementary to virtual reality or augmented reality or “metaverse” environments that can be offered by variations of collaboration platform 104.
Client devices 116 can include geocoding capability, such as a global position system (GPS) device, that allows the location of a device 116, and therefore its user 124, to be identified within system 100. Other means of implementing geocoding capabilities to ascertain the location of users 124 are contemplated, but in general system 100 can include the functionality to identify the location of each device 116 and/or its respective user 124. For example, the location of a device 116 or a user 124 can also be maintained within collaboration platform 104 or other nodes in system 100.
Client devices 116 are operated by different users 124 that are associated with a respective account 128 that uniquely identifies a given user 124 accessing a given client device 116 in system 100. A person of skill in the art is to recognize that the electronic structure of each account 128 is not particularly limited, and in a simple example embodiment, can be a unique identifier comprising an alpha-numeric sequence that is entirely unique in relation to other accounts 128 in system 100. Accounts 128 can also be based on more complex structures that may include combinations of account credentials (e.g. user name, password, Two-factor authentication token, etc.) that further securely and uniquely identify a given user 124. Accounts 128 can also be associated with other information about the user 124 such as name, address, age, travel document numbers, travel itineraries, language preferences, travel preferences, payment methods, and any other information about a user 124 relevant to the operation of system 100. Accounts 128 themselves may also point to additional accounts (not shown in the Figures) for each user 124, as a plurality of accounts may be uniquely provided for each user 124, with each account being associated with different nodes in system 100. For simplicity of illustration, it will be assumed that one account 128 serves to uniquely identify each user 124 across system 100. Indeed, the salient point is that accounts 128 make each user 124 uniquely identifiable within system 100.
In a present example embodiment, collaboration platform 104 can be based on media platforms or central servers that function to provide communications or other interactions between different users 124. Collaboration functions can include one or more ways to share information between users 124, such as chat, texting, voice calls, image sharing, chat rooms, video conferencing, shared document generation, shared document folders, project management scheduling, individual meeting scheduling either virtually or in person at a common location. Thus, collaboration platform 104 can be based on any known present or future collaboration infrastructure. Non-limiting examples of collaboration platforms 104 include enterprise chat platforms that host collaboration tools such as Microsoft Teams, or Slack, or can be based on business social media platforms such as Linked-In™. To expand on the possibilities, collaboration platform 104 can be based on social media ecosystems such as TikTok™, Instagram™, Facebook™ or the like. Collaboration platform 104 can also be based on multiplayer gaming environments such as Fortnite™ or metaverse environments such as Roblox™. Collaboration platform 104 can also be based on entire office suites such as Microsoft Office™ or suites of productivity applications that include email, calendaring, to-do lists, and contact management such as Microsoft Outlook™. Collaboration platform 104 can also include geo-code converters such as Google Maps™ or Microsoft Bing™ that can translate or resolve GPS coordinates from devices 116 (or other location sources of users 124) into physical locations. The nature of collaboration platform 104 is thus not particularly limited. Very generally, platform 104 provide a means for users 124 to search for travel itineraries using the particular teachings herein.
Collaboration platform 104 is configured to provide chat-based travel searching functions for devices 116 with assistive chat functions from LLM engine 120, including generation of structured travel search requests from unstructured travel search requests. LLM engine 120 can be based on any large language model platform such as ChatGPT from OpenAI. Notably, LLM engine 120 is limited to a static dataset that is difficult to update, and therefore unable to respond to real-time travel queries on its own.
Travel management engine 122 provides a central gateway for collaboration platform 104 to interact with travel actor engines 112, receiving structured search requests from collaboration platform 104 and conducting searches across travel actor engines 112, and collecting structured search results and returning those results to collaboration platform 104. Travel management engine 122, in variants, can be incorporated directly into collaboration platform 104.
Users 124 can interact, via devices 116, with collaboration platform 104 to conduct real time travel searches across engines 112 via natural language text-based chat. As desired or required, each account 128 (or linked accounts respective to different nodes) can be used by other nodes in system 100, including engines 112 to search, book and manage travel itineraries generated according to the teachings herein.
It is contemplated that collaboration platform 104 has at least one collaboration application 224-1 stored in non-volatile storage of the respective platform 104 and executable on its processor. (The types of potential collaboration applications 224-1 that fulfill different types of collaboration functions were discussed above.) Application 224-1 can be accessed by users 124 via devices 116 and be accessible by collaboration platform 104 to track expressions of travel interest by users 124. The expressions of interest may be direct (e.g. a chat message from a user 124 that says “I would like to book a trip to Paris”). The means by which expressions of interest are gathered is not particularly limited to this example. Platform 104 can include other applications 224 that can also be used to provide a calendar or scheduling functions.
It is contemplated that travel actor engines 112 also include an itinerary management application 132 stored in their non-volatile storage and executable on their processors. Applications 132 can suggest, generate and track individual travel itinerary records for individual users 124 based on travel search requests.
At this point it is to be clarified and understood that the nodes in system 100 are scalable, to accommodate a large number of users 124, devices 116, and travel actor engines 112. Scaling may thus include additional collaboration platforms 104 and/or travel management engines 122.
Having described an overview of system 100, it is useful to comment on the hardware infrastructure of system 100.
In this example, collaboration platform 104 includes at least one input device 204. Input from device 204 is received at a processor 208 which in turn controls an output device 212. Input device 204 can be a traditional keyboard and/or mouse to provide physical input. Likewise output device 212 can be a display. In variants, additional and/or other input devices 204 or output devices 212 are contemplated or may be omitted altogether as the context requires.
Processor 208 may be implemented as a plurality of processors or one or more multi-core processors. The processor 208 may be configured to execute different programing instructions responsive to the input received via the one or more input devices 204 and to control one or more output devices 212 to generate output on those devices.
To fulfill its programming functions, processor 208 is configured to communicate with one or more memory units, including non-volatile memory 216 and volatile memory 220. Non-volatile memory 216 can be based on any persistent memory technology, such as an Erasable Electronic Programmable Read Only Memory (“EEPROM”), flash memory, solid-state hard disk (SSD), other type of hard-disk, or combinations of them. Non-volatile memory 216 may also be described as a non-transitory computer readable media. Also, more than one type of non-volatile memory 216 may be provided.
Volatile memory 220 is based on any random access memory (RAM) technology. For example, volatile memory 220 can be based on a Double Data Rate (DDR) Synchronous Dynamic Random-Access Memory (SDRAM). Other types of volatile memory 220 are contemplated.
Processor 208 also connects to network 108 via a network interface 232. Network interface 232 can also be used to connect another computing device that has an input and output device, thereby obviating the need for input device 204 and/or output device 212 altogether.
Programming instructions in the form of applications 224 are typically maintained, persistently, in non-volatile memory 216 and used by the processor 208 which reads from and writes to volatile memory 220 during the execution of applications 224. Various methods discussed herein can be coded as one or more applications 224. One or more tables or databases 228 are maintained in non-volatile memory 216 for use by applications 224.
The infrastructure of collaboration platform 104, or a variant thereon, can be used to implement any of the computing nodes in system 100, including LLM engine 120, travel management engine 122 and/or travel actor engines 112. Furthermore, collaboration platform 104, LLM engine 120, travel management engine 122 and/or travel actor engines 112 may also be implemented as virtual machines and/or with mirror images to provide load balancing. They may be combined into a single engine or a plurality of mirrored engines or distributed across a plurality of engines. Functions of collaboration platform 104 may also be distributed amongst different nodes, such as within LLM engine 120, travel management engine 122 and/or travel actor engines 112, thereby obviating the need for a central collaboration platform 104, having a collaboration platform 104 with partial functionality while the remaining functionality is effected by other nodes in system 100. By the same token, a plurality of collaboration platforms 104 may be provided, especially when system 100 is scaled.
Furthermore, a person of skill in the art will recognize that the core elements of processor 208, input device 204, output device 212, non-volatile memory 216, volatile memory 220 and network interface 232, as described in relation to the server environment of collaboration platform 104, have analogues in the different form factors of client machines such as those that can be used to implement client devices 116. Again, client devices 116 can be based on computer workstations, laptop computers, tablet computers, mobile telephony devices or the like.
Block 304 comprises defining travel query types. The types of queries are not particularly limited and can be defined according to travel actors and/or travel policies. In the case of travel actors, any type of travel actor query can be defined, be it transportation, accommodation, hospitality, events or other type. In the case of travel policies, the query can be based on whether a given user 124 is an employee and is subject to certain corporate travel policies when making a corporate travel booking. Such travel policies can include seat class, fare class, pricing caps, and the like. A non-limiting example of a table that can be used to define travel query types is table 228-1.
Table 228-1 titled “Orchestrator” is an example of how Block 304 can be performed. Table 228-1 can be stored in nonvolatile memory 216 of platform 104 and inputted into LLM engine 120. The format of Table 228-1 is designed for the LLM Engine 120 based on ChatGPT from Open AI, but a person skilled in the art will appreciate that Table 228-1 is just an example.
Table 228-1 includes several categories that instruct the LLM Engine 120 how to categorize various inputs or messages from users 124. The example in Table 228-1 is limited and includes “˜GENERAL˜”, “˜AIR_SEARCH˜”, “˜AIR_POLICY˜”, “˜EVENTS_INFO˜”, “˜GROUND TRANSPORTATION ITINERARY_SEARCH˜”, “˜RESTART_SESSION˜” and “˜UNSUPPORTED˜”, each of which are defined in Table 228-1. Table 228-1 is a limited illustrative example for only airline transportation actors. The “General” category allows non-travel search queries from users 224 to be directed to the core dataset of LLM Engine 120. “Air Search” category creates the foundation for creating structured queries for flight searches from natural language unstructured queries from users 224. “Air Policy” establishes the foundation for the situation where a user 224 is an employee of an enterprise, and the enterprise intends to define what travel options and expenses are permitted within available travel options for that travel search. “Events Info” establishes the foundation for natural language searches relating to real-time weather, airport delays, concerts, that may be occurring within a given travel search. “Ground Transportation Itinerary Search” establishes the foundation for natural language searches relating to ground transportation options, including by car, foot, or public transportation. “Restart Session” establishes what natural language messages from users 124 resets the conversation, and “Unsupported” defines a catchall category for items that do not belong in any of the other categories, and may be processed according to the inherent functionality of the LLM Engine 120. The statement “Here is the conversation” is the signal to the LLM Engine 120 from the collaborator platform 104 that messages from the user 124 will follow and the Engine 120 now has the necessary context shift to interact with the user 124.
Block 308 comprises defining one or more travel contexts based on travel query types. Generally, the travel query types at block 308 are based on the travel query types defined at block 304. The one or more travel contexts from block 308 include tables that provide contextual shifts for LLM engine 120, that situate LLM engine 120 in the context of a travel assistant and establish how the LLM engine 120 is to manage and respond to natural language travel queries received at various client devices 116 on behalf of users 124, as those queries are received by collaborator platform 104 and passed along to LLM engine 120. Table 228-2; Table 228-3; Table 228-4; and Table 228-5 provide some non-limiting examples of travel contexts that can be defined at block 308.
Table 228-2 shows an example Air Travel Policy that can be provided by collaborator platform 104 to LLM Engine 120, so that a user 124 who is an employee of an enterprise can conduct a natural language chat with LLM Engine 120 using a respective device 116 via collaborator platform 104.
Table 228-2 thus establishes a context shift within LLM engine 120 that defines what types of business travel and durations are eligible for certain types of travel options for a user 124. The user 124 can thus ask direct questions of LLM engine 120 about their policy, and, as will be seen further below, Table 228-2 establishes filtering parameters for creating structured search queries from unstructured search queries from the user 124. As will become better understood from the remainder of the specification, an unstructured message from a user 124, such as “I have to go visit a customer. What flight options are there from Boston to Paris on Jul. 5, 2023?” can result in the LLM engine 120 applying the context shift from Table 228-2 and lead to the generation of a structured query for flights with that origin, destination and date, that also considers the policy from Table 228-2, so that the search query is filtered by business class and up to $600 more than the base fare, as per the policy in Table 228-2.
Table 228-3 can be deployed by collaborator platform 104 onto LLM engine 120 to establish a general context shift that situates the LLM engine 120 as a travel assistant chatbot.
Table 228-4, like Table 228-3, is another example that can be deployed by collaborator platform 104 onto LLM engine 120 to further establish a context shift that situates the LLM engine 120 as a travel assistant chatbot. (Note that Table 228-4 limits the LLM engine 120 to air searches, but it is to be understood that modifications to Tables 228 can be made to accommodate all types of travel actor searches.)
Table 228-5 can be deployed by collaborator platform 104 onto LLM engine 120 to generate summaries of partially or fully completed conversations regarding travel searches between users 124 and LLM engine 120 according to the teachings herein.
Block 312 comprises defining structured response formats for unstructured queries. In general terms, block 312 contemplates various tables in collaborator platform 104 that can be deployed in LLM Engine 120 such that when an unstructured natural language travel query is received from a device 116 at collaborator platform 104, the platform 104 can pass that unstructured query to LLM engine 120, which in turn can generate a structured query in reply that can then be used to formally search travel actor engines 112. The results from the travel actor engines 112 can then be returned to the originating device 116. Non-limiting example tables include Table 228-6 and Table 228-7.
Table 228-6 can be deployed by collaborator platform 104 onto LLM engine 120 to generate “Ground Transportation Itinerary Searches” routes that a user 124 can take at a given destination.
Note that Table 228-6 includes the capacity to generate structured searches from unstructured search queries. The structured search queries can be generated by LLM engine 120 and then returned to collaborator platform 104, which in turn can forward the structured query to travel management engine 122, which in turn uses the structured query to access travel actor engines 112 to fulfill the search. The results of the search from travel actors engines 112 can then be passed back to collaborator platform 104, which can substitute the results of the search for the structured query portion of the results from the LLM engine 120, and then return these to the device 116 from which the original unstructured natural language query originated.
Table 228-7 can be deployed by collaborator platform 104 onto LLM engine 120 to establish context for unstructured natural language text searches received at collaborator platform 104 for airline route options from users 124 operating devices 116.
Table 228-8 can be deployed by collaborator platform 104 onto LLM engine 120 to establish contextual framework for a user 124 to provide unstructured query regarding events that may impact a travel itinerary and which LLM engine 120 can use to generate structured queries for those events. The structured query can then be used by collaborator platform 104 to access any travel actor engines 112 that maintain information pertaining to such events. Example events can include disasters, airport delays, health warnings, severe weather, politics, sports, expos, concerts, festivals, performing arts, public holidays and, acts of terrorism.
Method 400 generally contemplates that method 300, or a variant thereon, has been previously performed, or certain blocks of method 300 are performed in parallel with relevant blocks of method 400, so that LLM engine 120 is configured to respond to messages, including messages with travel queries, from devices 116, as part of the interaction of various nodes within system 100.
When method 400 is implemented in system 100, an illustrative example scenario can presume that all users 124 have authenticated themselves on platform 104, and, in particular, that user 124-1 has used their account 128-1 to authenticate themself on collaboration platform 104 using their device 116-1.
Block 404 comprises receiving a natural language input message. Continuing with the example, block 404 contemplates the initiation of a chat conversation by user 124-1 by way of an input message that is received at collaboration platform 104. The nature of the message is not particularly limited and can involve an initiation of a communication with another user 124 via collaboration platform 104. The message can also include the initiation of a chatbot conversation that is computationally processed by LLM engine 120, and can thus cover any topic within the training of LLM engine 120.
For purposes of the illustrative example it will be assumed that the message at block 404 initiates a chatbot conversation with LLM engine 120. This example is shown in
At block 412, a determination is made as to whether the message 504-1 includes a travel query. Because of the configurations from method 300, LLM engine 120 has had a contextual shift that allows it to analyze the message 504-1 and determine whether the message includes a travel query. Based on the example message 504-1, “Hey, I need to book a flight to Paris.”, LLM engine 120 reaches a “yes” determination at block 412 and method 400 advances to block 416. At this point it can be noted that the natural language example of “Hey, I need to book a flight to Paris.” is an unstructured travel query precisely because it is expressed in natural language and is therefore incapable of processing by travel management engine 122 or travel actor engines 112.
Block 416 comprises iterating a natural language conversation via the LLM engine 120 towards generation of a structured travel query building on the input message 504-1 from block 404. Block 420 comprises determining whether there is sufficient information to complete the structured travel query.
Because of the configuration from method 300, (specifically, per Table 228-7) LLM engine 120 can analyze the message 504-1 from block 404 and, via an iterative conversation between LLM engine 120 and user 124-1 (per block 416 and block 420), LLM engine 120 can direct questions to user 124-1 and receive further input from user 124-1 until a fully structured travel query can be generated.
Performance of block 416 and block 420 is shown in
It is to be emphasized that the messages 504 in
Block 424 comprises engaging the LLM to prepare a draft response message to the original message from block 404.
(Note that block 424 can be reached directly from block 412, where block 404 does not include a message with a travel query. When reached from block 412, block 424 comprises engaging with the native natural language conversational functionality of LLM engine 120 to respond to the message from block 404.)
According to our illustrative example from
Thus, according to our example in
Sub-message 504-4-SQ thus contains a structured travel query that can be used by travel management engine 122 and/or travel actor engines 112 to effect a search for airline itineraries that fit the unstructured travel query assembled from message 504-1, message 504-2 and message 504-3.
Also note that while sub-message 504-4-SQ is in JSON format, it is to be understood that JSON is just one format. Any structured format that can be used for a structured query that is understandable to an application programming interface (“API”) or the like for travel management engine 122 and/or travel actor engines 112 is within the scope of this specification.
Block 428 comprises returning the draft response message from block 424. Performance Block 428 is also represented in
Block 432 comprises determining if the message from block 428 includes a structured travel query. A “no” determination leads to block 444 and thus the message drafted at block 424 is sent directly to the originating client device 116. A “yes” determination at block 432 leads to block 436, at which point the structured travel query is sent to external sources for fulfillment.
Block 440 comprises receiving a response to the structured travel query from block 436.
Block 444 comprises generating an output message in response to the input message from block 404. Where no travel query was included in the message from block 404, then, as discussed, block 416, block 420, block 436 and block 440 do not occur and thus the output message at block 444 is consistent with the native natural language processing and conversational functions within LLM engine 120.
However, where a travel query was included in the message at block 404, as per our example, then block 444 comprises generating an output message that includes the travel query response from block 440. It is contemplated that the display of device 116 is controlled to generate the output message.
Many variants and extrapolations of the specific example discussed in the Figures are contemplated and will now occur to those of skill in the art. For example, message 504-4-F can additionally include an invitation for further conversation from LLM engine 120 to help further refine the search results. As an example, message 504-4-F could include the additional question “Do you have any airline preferences?”, inviting further natural language conversation between user 124 and LLM engine 120, as intermediated by collaboration platform 104. The inclusion of such an additional question can cause further iterations at block 416 and block 420, to generate further structured queries of the type earlier discussed in relation to sub-message 504-4-SQ, that lead to further searches conducted on travel actor engines 112 in similar fashion to the earlier discussion in relation to block 436 and block 440. Such further structured searches can continue to be narrowed as per responses from the user 124, with LLM engine 120 generated the structured searches and travel management engine 122 fulfilling the searches, with collaboration platform 104 substituting the structured search queries from LLM engine 120 with the user-readable responses obtained by travel management engine 122. User 124 can likewise engage in booking functions via travel management engine 122 that are offered in flight cards such as flight card 904-1.
A person skilled in the art can also appreciate how the structured queries generated by LLM engine 120 can be extremely sophisticated in nature, whereby travel management engine 122 may make a series of structured queries to travel actor engines 112. Here is an example scenario. If user 124-1 generates an unstructured natural language query of the form “I would like to see flights from Nice to Paris on April 12 where the flights must land only during operational hours of the Paris Metro system”, then a first structured query can be made of a first travel actor engine 112 that has information about the operational hours of the Paris Metro system, which can then be returned to LLM engine 120 to generate a second structured query that filters by flights that land during the operational hours returned from the first query. LLM engine 120 may also engage in a series of questions to the user 124 to ultimately arrive at the series of necessary structured queries of different travel actor engines 112 to obtain results that are responsive to the original query.
It will now be apparent just how far the unstructured queries can scale within the scope of the present specification: “I would like to see flights from Nice to Paris on April 12 where the flights must land only during operational hours of the Paris Metro system and on days when the Paris Symphony Orchestra is performing Mozart and hotel prices are less than 300 Euros per night for double occupancy within ten blocks of the symphony venue”. Here additional structured queries are made of travel actor engines 112, which include event actors that ascertain the schedule if the Paris Symphony and accommodation actors that have hotel rooms at the specified price point and a location within the prescribed geographic radius.
The inventors have compared the efficiency of processing messages 504 by ChatGPT™ from OpenAI™ as an example of one single LLM engine 120 using object 1200 and objects 1204, in comparison to processing messages 504 with prior art contextualization techniques using ChatGPT™ that use a non-hierarchical context shift. The inventors have noted that these show efficiency gains of at least 18.67% and up to 24% as shown in the table labelled “Efficiency and Cost Comparison Table”. (See also
To further illustrate the hierarchical use of contextualization objects outlined in this disclosure, a comparative analysis framework to illustrate the computational efficiency and cost-effectiveness of System 100 is shown in
Contrasting with
In sum,
System 100 thus employs a hierarchical model of contextualization object 1200 and objects 1204 that improves classification precision without encountering the prompt size limitations typical in standard large language models. This hierarchical structure allows for the implementation of multiple levels of contextualization objects, including sub-contextualization objects (and potential for sub-sub contextualization objects, and deeper levels), thereby creating a robust tree. Each level can focus on increasingly specific categories, beginning with general topics, (air search, air policy, etc.) and becoming increasingly specific (flight cards, events) which can be refined further as needed.
System 100 allows for extensive scalability and flexibility in managing complex queries but also improves the overall efficiency, by reducing the number of tokens required for each classification by narrowing the focus at each level of the hierarchy, thereby decreasing the computational load and increasing the response speed. This method allows for each query to be handled more effectively, with a systematic narrowing down that avoids the broad sweep of general classifiers. The ability to maintain a large classifier structure, while focusing on detailed topics, enhances the system's capability to deliver precise and contextually appropriate responses.
An incidental benefit is that the hierarchical contextualization objects also introduces significant cost savings by optimizing the number of tokens required to process queries. Prior art systems may require a larger, more expensive model to handle extensive prompts in a single request. In contrast, our system efficiently distributes the load across several smaller requests, each tailored to a specific part of the classifier tree. This approach is illustrated by the substantial cost reductions observed, where the cost for an AIR_SEARCH request, for example, is reduced from $0.45 to $0.108, representing a savings of approximately 76%.
This cost-effectiveness is further highlighted by the practical limitations of token capacities in various GPT models. For instance, the inventors' examples demonstrate that a GPT-4 model with a 32K token capacity is required in prior art. System 100, utilizing a 4K model, avoids this issue by efficiently managing the token distribution across multiple classifiers, thus not only saving costs but also enhancing scalability and adaptability to complex queries.
In view of the above it will now be apparent that variations, combinations, and/or subsets of the foregoing embodiments are contemplated. For example, collaboration platform 104 may be obviated or its function distributed throughout a variant on system 100, by incorporating collaboration platform 104 directly inside LLM engine 120. Furthermore, the present specification is readily extensible into metaverse environments, where devices 116 include virtual or augmented reality hardware and operate avatars within a metaverse platform. The metaverse platform can host virtual travel agents in the form of metaverse avatars, whose speech is driven by the teachings herein. The teachings herein can also be incorporated into physical robots that operate according to the teachings herein. While the present embodiments refer to travel searches, broader e-commerce searches can also be effected in variants, such as for cellular telephone plans, vehicle purchases, home purchases, whereby user messages containing unstructured search requests are received and an LLM engine is used to flesh out the search parameters and generate structured search requests which can then be passed to e-commerce search engines, and the returned results can replace the structured search request in the final result returned to the user.
In other variants, collaboration platform 104 need not provide collaboration or other communication services between users 124, and thus collaboration platform 104 can simply be substituted with a chatbot platform that is used to fulfill the travel search dialogue with a given user 124 according to the teachings herein. Collaboration platform 104 can be incorporated into the LLM Engine 120.
In another variant, the present teachings are applicable to search engines beyond travel. For example, in the application of a home renovation, the general contextualization object can classify queries into materials, labor, delivery schedule, and other or general. The specific contextualization objects can be respective to different search engines for home renovation materials (e.g. building supplies, bathroom fixtures, kitchen fixtures, plumbing, electrical), and for labor (e.g. carpenters, plumbers, electricians) and for delivery services (e.g. courier companies, trucking companies, customs brokers, logistic handlers). The natural conversation with LLM engine 120 can be effected using reduced tokens and thereby allow for more complex conversations, and appropriate responses. Generally, the present specification can facilitate coordination among a variety of different search engine sources through a single LLM engine, thereby providing a natural language conversation environment that is more efficient and has a reduced likelihood of producing hallucinations.
In another variant, machine learning feedback can be used to further improve the context shifts and/or train the LLM Engine 120 in providing its dialogue with the user 124. The conversations between been users 124 and LLM Engine 120 can be archived and fed into a machine learning studio platform. The studio allows to train a machine learning algorithm. The machine learning algorithm, can for example, generate a new version of the orchestrator prompt engineering from Table 228-1, or any of the other Tables 228. Then the updated model can deployed into method 300 via an application via a workflow from the machine learning studio platform to LLM Engine 120.
Accordingly, in this variant, one or more of the applications 224 may include the machine learning studio platform with any desired related machine learning deep-learning based algorithms and/or neural networks, and the like, which are trained to improve the Tables in method 300. (Hereafter machine learning applications 224). Furthermore, in these examples, the machine learning applications 224 may be operated by the processor 208 in a training mode to train the machine learning and/or deep-learning based algorithms and/or neural networks of the machine learning applications 224 in accordance with the teachings herein.
The one or more machine-learning algorithms and/or deep learning algorithms and/or neural networks of the machine learning applications 224 may include, but are not limited to: a generalized linear regression algorithm; a random forest algorithm; a support vector machine algorithm; a gradient boosting regression algorithm; a decision tree algorithm; a generalized additive model; neural network algorithms; deep learning algorithms; evolutionary programming algorithms; Bayesian inference algorithms; reinforcement learning algorithms, and the like. However, generalized linear regression algorithms, random forest algorithms, support vector machine algorithms, gradient boosting regression algorithms, decision tree algorithms, generalized additive models, and the like may be preferred over neural network algorithms, deep learning algorithms, evolutionary programming algorithms, and the like.
A person skilled in the art will now appreciate that the teachings herein can improve the technological efficiency and computational and communication resource utilization across system 100 by making more efficient use of network and processing resources in system 100, as well as more efficient use of transportation actors 224-1. For example, present large language models (LLM) cannot provide real time travel search because they are trained and rely upon on a static dataset that is periodically updated. Enabling real-time access to the internet is generally incompatible with the static nature of large language model datasets, and different architecture and continuous updating, which would be computationally expensive and challenging to manage. At the same time, existing chat functionality does not address the problem of collecting rich and structured travel queries that can be used to provide meaningful searches. It should now also be apparent that LLM Engine 120 can be used as a natural language processing (NLP) engine for system 100 and its variants.
It should be recognized that features and aspects of the various examples provided above can be combined into further examples that also fall within the scope of the present disclosure. In addition, the figures are not to scale and may have size and shape exaggerated for illustrative purposes.
The present specification claims priority from U.S. Provisional Patent Application 63/463,146, filed May 1, 2023, the contents of which are incorporated herein by reference.
Number | Date | Country | |
---|---|---|---|
63463146 | May 2023 | US |