The field of the invention is communication of medical information using an automated system.
The following description includes information that can be useful in understanding the present disclosure. It is not an admission that any of the information provided herein is prior art or relevant to the presently claimed invention, or that any publication specifically or implicitly referenced is prior art.
Variations of Artificial Intelligence (“AI”) have been used in many different fields, including science, gaming industry, statistics, etc. For example, Google's AlphaGo™ is an AI game that mimics human play, and then improves its own play by running large numbers of games against other instances of itself.
AI has also been used to automatically detect human emotional states, using prosody of speech. For example, U.S. Pat. Pub. No. 20060122834 A1 to Bennett discloses a prosody and emotion recognition system that enables a quick and accurate recognition of speech utterance based on literal content and user emotional state information. For another example, U.S. Pat. No. 8,682,666 to Degani discloses method and system to determine current behavioral, psychological and speech styles characteristics of a speaker in a given situation and context by analyzing the speech utterances of the speaker.
AI has also been put to use in automatically and contextually summarizing human communications. For example, U.S. Pat. No. 9,420,227 to Shires discloses a system for differentiating between two or more individuals' voice data during a conversation, and the producing corresponding text for each individual. Shires also discloses AI use of voice data, physical features of the speakers, characteristics of the words utilized, etc. to generate summarized output.
Still further, AI has been used to detect errors in medical communications. For example, U.S. Pat. Pub. No. 2014/0012575 to Ganong discloses a system that can detect speech input in a medical or other field, and evaluate the speech for indications of potential significant errors.
Despite all the work in AI over the years, there doesn't appear to be any work directed to creating de novo communications that have appropriate tone, surface text, and subtext, as might be particularly useful in communicating medical information to different recipients.
All publications identified herein are incorporated by reference to the same extent as if each individual publication or patent application were specifically and individually indicated to be incorporated by reference. Where a definition or use of a term in an incorporated reference is inconsistent or contrary to the definition of that term provided herein, the definition of that term provided herein applies and the definition of that term in the reference does not apply.
The subject matter described herein provides computer enabled apparatus, systems and methods for automatically generating custom communications to recipients. A particular focus is for the generated communications to take into account how different recipients can be expected to respond to the communications, on both intellectual and emotional levels.
Contemplated methods include deciphering what each person is saying during a conversation, making multiple inferences from the words, prosody, and possibly other observable cues of the conversation, and then generating written or other communications summarizing the conversation. In some embodiments, stock phrases are selected and assembled with a goal of achieving surface text, subtext and tone.
In a doctor-patient interaction, for example, the computer generated communication(s) might include guidance based upon inferred diagnostic information, inferred doctor and recipient's respective contexts, and desired impacts on the recipients of the communication(s). Thus, systems and methods contemplated herein would very likely generate different communications for patients, family members, and consulting physicians. Also, systems and methods contemplated herein would very likely generate different communications to patients having similar diagnoses, but different prognoses. Such differences can advantageously result from different tones, surface texts, and subtexts in the communications.
The various inferences can be obtained from suitable AI systems, by submitting text and/or audio through established APIs. Some or all of the contemplated inferencing and other steps can be performed in real time or near real time. Various objects, features, aspects and advantages of the disclosed subject matter will become more apparent from the following detailed description of embodiments, along with the accompanying drawing figures in which like numerals represent like components.
Throughout the following discussion, numerous references will be made regarding servers, services, interfaces, engines, modules, clients, peers, portals, platforms, or other systems formed from computing devices. It should be appreciated that the use of such terms is deemed to represent one or more computing devices having at least one processor (e.g., ASIC, FPGA, DSP, x86, ARM, ColdFire, GPU, multi-core processors, etc) configured to execute software instructions stored on a computer readable tangible, non-transitory medium (e.g., hard drive, solid state drive, RAM, flash, ROM, etc). For example, a server can include one or more computers operating as a web server, database server, or other type of computer server in a manner to fulfill described roles, responsibilities, or functions. One should further appreciate the disclosed computer-based algorithms, processes, methods, or other types of instruction sets can be embodied as a computer program product comprising a non-transitory, tangible computer readable media storing the instructions that cause a processor to execute the disclosed steps. The various servers, systems, databases, or interfaces can exchange data using standardized protocols or algorithms, possibly based on HTTP, HTTPS, AES, public-private key exchanges, web service APIs, known financial transaction protocols, or other electronic information exchanging methods. Data exchanges can be conducted over a packet-switched network, a circuit-switched network, the Internet, LAN, WAN, VPN, or other type of network. The terms “configured to” and “programmed to” in the context of a processor refer to being programmed by a set of software instructions to perform a function or set of functions.
While the inventive subject matter is susceptible of various modification and alternative embodiments, certain illustrated embodiments thereof are shown in the drawings and will be described below in detail. It should be understood, however, that there is no intention to limit the invention to the specific form disclosed, but on the contrary, the invention is to cover all modifications, alternative embodiments, and equivalents falling within the scope of the claims.
The following discussion provides many example embodiments of the inventive subject matter. Although each embodiment represents a single combination of inventive elements, the inventive subject matter is considered to include all possible combinations of the disclosed elements. Thus if one embodiment comprises elements A, B, and C, and a second embodiment comprises elements B and D, then the inventive subject matter is also considered to include other remaining combinations of A, B, C, or D, even if not explicitly disclosed.
In some embodiments, the numbers expressing quantities or ranges, used to describe and claim certain embodiments of the invention are to be understood as being modified in some instances by the term “about.” Accordingly, in some embodiments, the numerical parameters set forth in the written description and attached claims are approximations that can vary depending upon the desired properties sought to be obtained by a particular embodiment. In some embodiments, the numerical parameters should be construed in light of the number of reported significant digits and by applying ordinary rounding techniques. Notwithstanding that the numerical ranges and parameters setting forth the broad scope of some embodiments of the invention are approximations, the numerical values set forth in the specific examples are reported as precisely as practicable. The numerical values presented in some embodiments of the invention can contain certain errors necessarily resulting from the standard deviation found in their respective testing measurements. Unless the context dictates the contrary, all ranges set forth herein should be interpreted as being inclusive of their endpoints and open-ended ranges should be interpreted to include only commercially practical values. Similarly, all lists of values should be considered as inclusive of intermediate values unless the context indicates the contrary.
As used in the description herein and throughout the claims that follow, the meaning of “a,” “an,” and “the” includes plural reference unless the context clearly dictates otherwise. Also, as used in the description herein, the meaning of “in” includes “in” and “on” unless the context clearly dictates otherwise.
All methods described herein can be performed in any suitable order unless otherwise indicated herein or otherwise clearly contradicted by context. The use of any and all examples, or exemplary language (e.g., “such as”) provided with respect to certain embodiments herein is intended merely to better illuminate the invention and does not pose a limitation on the scope of the invention otherwise claimed. No language in the specification should be construed as indicating any non-claimed element essential to the practice of the invention.
Groupings of alternative elements or embodiments of the invention disclosed herein are not to be construed as limitations. Each group member can be referred to and claimed individually or in any combination with other members of the group or other elements found herein. One or more members of a group can be included in, or deleted from, a group for reasons of convenience and/or patentability. When any such inclusion or deletion occurs, the specification is herein deemed to contain the group as modified, thus fulfilling the written description of all Markush groups used in the appended claims.
The following method/system is able to transform recordings of spoken interactions between a doctor and a patient into formatted out-patient letters or natural language which goes into free-form text fields of EMR systems. This invention automates a process which is currently predominantly manual, namely the creation of out-patient letters and entries into EMR systems which consumes a lot of time of medical professionals such as physicians, medical assistants, scribes, etc.
In a preferred embodiment, the system consists of four modules or sets of modules:
A visual representation of a preferred embodiment is shown in
While Modules 1 and 2 can be standard technologies, see the provided citations, Modules 3 and 4 are not standard, and will be described in further detail below. Tagging module 3 has two sub-modules, 3a and 3b. A bucket classification module 4 also has to sub-modules, 4b and 4b.
Module 3—a module to transform diarized spoken language in textual form into a conceptual graph representation. The output of Modules 1 and 2 combined are delivered to Module 3 as a sequence in the form word_1/speaker_1 word_2/speaker_2 . . . , for example
good/P morning/P doctor/P how/D are/D you/D . . . /P
where P means patient and D means doctor.
The task of turning this input into a conceptual graph representation is performed by two sub-modules, one for tagging concepts and one for tagging relations, described in the following:
Module 3a (
well/D [Human-Patient I/P] have/P [Disease-BreastCancer breast/P cancer/P] ok/D
The set of possible semantic concepts can use predefined concepts, such as Human-Father, Human-Mother, Human-Patient, . . . , Location-Hospital, . . . , as derived from medical ontologies such as SNOMED CT or ICD-10, or they can be enhanced by concepts discovered by the annotators throughout the annotation process.
Using the annotated diarized speech input, we can now train a DNN-based tagger (e.g. a BLST with attention mechanism) which uses embeddings (e.g. with a window of 256 words times 128 embeddings). Here, the embeddings should be the concatenation of two vectors, a word vector and a speaker vector at the input. And the output should be one concept per input word, distinguishing words bearing no concept (0), ones which initiate a concept (e.g. Begin_Disease-BreastCancer), and those at the inside of a concept phrase (e.g. Continue_Disease-BreastCancer). This is an example tagging sequence:
well/D:0 I/P:Begin_Human-Patient have/P:0 breast/P:Begin_Disease-BreastCancer
cancer/P:Continue_Disease-BreastCancer ok/D:0
where “:” is the separator between in- and output.
Module 3b (
well/D:0 I/P:Begin_Human-Patient have/P:0 breast/P:Begin_Disease-BreastCancer
cancer/P:Continue_Disease-BreastCancer ok/D:0
to a limited set of relations, e.g.
In order to be able to distinguish concepts of the same type and uniquely define relations, each instance of a concept in the input sequence to Module 3b is automatically assigned an ID (i.e. 16). The next encounter of the same concept would get another ID (e.g. 22). Hence, the input of the annotation internally has the form
well/D:0 I/P:Begin_14 have/P:0 breast/P:Begin_16 cancer/P:Continue_16 ok/D:0
where
ID 14 is Human-Patient (see
ID 16 is Disease-BreastCancer (see
ID 21 is Anomaly-Tumor
and the annotator annotates such an input sequence with relations between the individual concepts, e.g.,
(hasDisease, 14, 16) (see
which stands for
“the patient has breast cancer”
or
(causedBy, 16, 21)
which stands for
“the breast cancer is caused by a tumor”.
With these annotations, DNNs are trained for each relation type (hasDisease, causedBy, . . . ). The input layer of these DNNs consists of the concatenation of input and output layers of the DNN of Module 3a), and the output layer consists of a matrix over all possible parameter combinations the respective relation type can assume. E.g. for a relation with two parameters, such as hasDisease or causedBy, the number of nodes in the output layer is N̂2 with N being the maximum ID in the training data.
At run time, as during the training, the input of the DNN will be the concatenation of input and output layers of the DNN of Module 3a. To determine which relations were found, one needs to find all those output matrix nodes that fired, determining the tagged relations, e.g., there might be two nodes firing for the causedBy tagger, such as
(causedBy, 16, 21)
(causedBy, 34, 64)
Module 4 (
Module 4a (
Module 4b (
select relations for the bucket
sort the relations alphabetically
use them as input of a DNN (this DNN should not be a recurrent neural network), e.g.
(causedBy, 16, 21)
(hasDisease, 14, 16)
( . . . )
0
0
where the zeros at the end are inserted to pad to the fixed width of the input layer (e.g. 256). The output of the bucket-dependent natural language generator has also a fixed number of nodes, e.g. 256, which consists of word indices in a vocabulary list, e.g.
34, 25, . . . , 48, 26, EOS, 87, 89, . . .
where EOS is the end-of-section marker. For example, this list of indices could stand for the natural language section
“the patient has breast cancer” (See
In a preferred embodiment (
Another preferred embodiment (
The initial stages transform the recorded conversation into a text format usable by the natural language processing (NLP) modules that follow: first, a speaker diarization module determines who is speaking when and uses this information to break the audio recording into segments, which are then passed through a medical automatic speech recognition (ASR) stage. Following ASR, the scribe must convert a transcribed spontaneous conversation into a final and fully formatted report. The scribe does not perform this translation directly—this would require enormous amounts of parallel data to solve, end to end, with any single technique. Instead, a two-stage approach is developed in which the scribe mines the conversation for information and saves it in a structured format, then exports this structured data to the final report.
Between these two stages, there is a “reasoning” step that operates directly on the structured data to clean and prepare it for export, if needed. In this way, the bulk of the NLP work is divided into two well-studied problems: knowledge extraction (330) and natural language generation (350). Generating structured data as an intermediate step has other advantages as well; for one, it can be kept in the patient's history for use later by the scribe—or even by other systems, if it is saved in standardized structured data formats.
Speaker diarization (310) is the “who spoke when” problem, also called speaker indexing. The input is audio features sampled at 100 Hz frame rate, and the output is frame-labels indicating speaker identify for each frame. Four labels are possible: speaker 1 (e.g. the doctor), speaker 2 (e.g. the patient), overlap (both speakers), and silence (within-speaker pauses and between-speaker gaps). The great majority of doctor-patient encounters involve exactly two speakers. Although this method is easily generalizable to more speakers, the current embodiment focuses on the two-speaker problem.
The diarization broadly distinguishes “bottom-up” vs. “top-down” approaches. This embodiment uses a top-down approach that utilizes a modified expectation maximization (EM) algorithm at decoding time to learn the current speaker and background silence characteristics in real time. It is coded in plain C for maximum efficiency and currently operates at ˜50× real-time factor.
Diarization requires an expanded set of audio features compared to ASR. In ASR, only phoneme identity is of final interest, and so audio features are generally insensitive to speaker characteristics. By contrast, in diarization, only speaker identity is of final interest. Also, diarization performs a de facto speech activity detection (SAD), since states 1-3 vs. state 4 are speech vs. silence. Therefore features successful for SAD are helpful to diarization as well. Accordingly, an expanded set of gammatone-based audio features are used for the total SAD+diarization+ASR problem.
Speech recognition (320). ASR operates on the audio segments produced by the diarization stage, where each segment contains one conversational turn (1 speaker+possibly a few frames of overlap). Currently, the diarization and ASR stages are strictly separated and the ASR decoding operates by the same neural network (NN) methodology for general medical ASR. (See E Edwards et al, Medical speech recognition: reaching parity with humans. In Proc SPECOM, volume LNCS 10458, pages 512-524. Springer, 2017). In brief, the acoustic model (AM) consists of a NN trained to predict context-sensitive phones from the audio features; and the language model (LM) is a 3- or 4-gram statistical LM prepared with methods of interpolation and pruning that were developed to address the massive medicalvocabulary challenge. Decoding operates in real time by use of weighted finite-state transducer (WFST) methodology coded in C++. Our current challenge is to adapt the AM and LM to medical conversations, which have somewhat different statistics compared to medical dictations.
Knowledge extraction (330). A novel strategy is adopted to simplify the knowledge extraction problem by tagging sentences and turns in the conversation based upon the information they are likely to contain. These classes overlap largely with sections in the final report—chief complaint, medical history, etc. Then, a variety of strategies are applied, depending on the type of information being extracted, on filtered sections of text.
A hierarchical recurrent neural networks (RNNs) is used to tag turns and sentences with their predicted class; each sentence is represented by a single vector encoded by a word-level RNN with an attention mechanism. Sentences are classified individually rather than the entire document. In most cases, a sentence vector is generated from an entire speech turn; for longer turns, however, detection of sentence boundaries is required. This is essentially a punctuation restoration task, which has been undertaken using RNNs with attention. (See W Salloum, et al, Deep learning for punctuation restoration in medical reports. In Proc Workshop BioNLP, pages 159-164. ACL, 2017).
To extract information from tagged sentences, one or more of several strategies can be applied. One strategy is to use complete or partial string match to identify terms from ontologies. This is effective for concepts which do not vary much in representation, such as medications. Another strategy is extractive rules using regular expressions, which are well suited to predictable elements such as medication dosages, or certain temporal expressions (e.g., dates and durations). Other unsupervised or knowledge-based strategies, such as Lesk-style approaches in which semantic overlap with dictionary definitions of terms is used to normalize semantically equivalent phrases, as has been done successfully for medical concepts. These approaches are suitable for concepts that can vary widely in expression, such as descriptions of symptoms. Fully supervised machine learning approaches can be employed for difficult or highly specialized tasks—e.g., identifying facts not easily tied to an ontology entry, such as symptoms generally worsening.
The knowledge extraction (KE) stage also relies on extractive summary techniques where necessary, in which entire sentences may be copied directly if they refer to information that is tagged as relevant but is difficult to represent in our structured type system—for example, a description of how a patient sustained a workplace injury. At a later stage, extracted text is processed to fit seamlessly into the final report (e.g., changing pronouns).
Reasoning from extracted knowledge. Following the information extraction stage is a reasoning module (340), which performs several functions to validate the structured knowledge and prepare it for natural language generation. Through a series of logical checks, the reasoning module corrects for any gaps or inconsistencies in the extracted knowledge. These may occur when there is critical information that is not explicitly mentioned during the encounter, or if there are errors in diarization, ASR, or KE.
This stage also has access to the templates used when generating the final note. In the event that certain templates can only be partially filled, the reasoning module will attempt to intuit the missing information from existing structured data in the patient's history, if available. Wherever possible, data is also encoded in structures compatible with the HL7 FHIR v3 standard to facilitate interoperability with other systems. For example, if the physician states an intent to prescribe a medication, the extracted information is used to fill a FHIR MedicationRequest resource.
The natural language generation (NLG) module (350) produces and formats the final report. Medical reports follow a loosely standardized format, with sections appearing in a generally predictable order and with well-defined content within each section. Our strategy is a data-driven templatic approach supported by a finite-state “grammar” of report structure.
The template bank consists of sentence templates annotated for the structured data types necessary to complete them. This bank is filled by clustering sentences from a large corpus of medical reports according to semantic and syntactic similarity. The results of this stage are manually curated to ensure that strange or imprecise sentences cannot be generated by the system, and to ensure parsimony in the resulting type system.
Using the same reports, grammar is induced using a probabilistic finite-state graph, where each node is a sentence and a single path through the graph represents one actual or possible report. Decoding optimizes the maximal use of structured data and the likelihood of the path chosen. The grammar helps to improve upon one common criticism of templatic NLG approaches, which is the lack of variation in sentences, in a way that does not require any “inflation” of the template bank with synonyms or paraphrases: during decoding, different semantically equivalent templates may be selected based on context and the set of available facts, thus replicating the flow of natural language in existing notes.
Format does vary between note type—for example, outpatient notes are quite different from hospital discharge summaries—and even between providers. Separate NLG models are built to handle each type of output.
Finally, all notes pass through a processor that handles reference and anaphora (e.g., replacing references to the patient with the appropriate gender pronoun), truecasing, formatting, etc. generate a full template bank from data.
Exemplary Embodiments of an Automated System for, Communicating Information
Other aspects of the inventive subject matter include methods of, and/or an automated systems for, communicating information based in part on an oral communication between a plurality of persons. In these aspects, it is contemplated that the plurality of persons can be in various relationships. For example, two or more persons can be in a medical provider-patient, attorney-client, or other professional-client relationship, which often requires confidential communications. Other contemplated relationships include non-professional relationships, such as parent-child, or salesperson-potential customer relationships.
Contemplated communications can occur using any medium. For example, such communications can be in-person (e.g., face-to-face), over the phone, or over the internet (e.g., via Skype®, etc), and can be conducted entirely through voice, entirely through written or other visual symbols, or through a combination of voice and visual symbols. Other modalities are also contemplated, e.g. video or motion capture. Contemplated communications can be completed in a single-session (e.g., without being intervened for more than an hour, more than a day, etc), or during multiple-sessions communications. The latter, for example, might occur over a multi-month period of hospitalization.
In some embodiments of the inventive subject matter, an automated system converts oral communications between at least first and second persons into a written script (e.g., digitally, etc). Conversion can be in real-time, near-real time (within a minute), or at some subsequent time using a recording of the communication. Conversion can use local and/or remote data storage units.
From the script of an oral communication, the automated system can infer context(s) of the oral communication. As used herein, the term “context” refers to any environment in which the communication takes place. Examples of context include time, place, identity of the speakers (e.g., gender, name, occupation, etc) and relationships between speakers. Further, context can include a speaker's emotion, level of understanding, competence, and intent of the speaker in the communication, etc. Context can be inferred from the content of the voice, or from non-voice aspects of communication. For example, inferences from voice can be made using types of questions, types of answers, use of vocabulary, volume or tone of the speakers' voice, and/or other sounds the speaker makes during the conversation (e.g., laughter, crying, etc). Inferences from non-voice communication include body language (e.g., shrugging, cursing, pushing to show refusal etc) and facial expression (e.g., angry face, sad face, happy face, etc).
Inferences from non-voice communication need not even come from the speaker's voice or body. For example, inferences from oral communication between a real estate agent and a potential buyer could be derived from location, age and appearance of other family members present during the conversation. In a doctor-patient example, context inferences could be derived from the nature of the facility (e.g., an emergency room versus an arthritis clinic).
It is contemplated that both computer-derived content and computer-derived context of a communication can be used to infer diagnostic information. Such information can include the name and status of the disease (or symptom), any physical/mental symptoms related to the disease, potential and/or popular treatment methods and period, any potential side effects of the treatment methods, code(s) of the disease, procedure, diagnoses (e.g., SNOMED, ICD-10, etc), and so on. In some embodiments, the automated system can generate a list of questions related to the oral communication and the inferred diagnostic information to complete the diagnostic information. In these embodiments, the automated system can send (e.g., in real-time, etc) the list of questions or one question in each time to the speaker (e.g., doctor, medical provider), to ensure the inferred diagnostic information is correct or any further information can be collected to further diagnose the symptoms of the patients.
A simple example can be used to help understand some of these concepts. In this example, a patient and a doctor are talking in a hospital. Based on the following exchange, the automated system infers that the patient is an out-patient, and that the conversation is taking place in the doctor's office.
Based on the content and context of the conversation described above, the automated system could infer that the patient has gouty arthritis, and might suggest that diagnosis to the doctor. The automated system might also send one or more questions to the doctor to ensure the right diagnostic information, including “Which wrist, left or right?” or “Is the wrist red and swollen?”
Inferences contemplated herein can be made with inference engines, using known techniques of forward and backward chaining, applied using rules sets acting upon a knowledge base. Examples of suitable inference engines that can be used to execute aspects of the inventive subject matter are referenced elsewhere herein.
Continuing with the previous example, it is contemplated that after a patient's visit to the doctor's office, the doctors or other entity would send a written communication to one or more persons depending on the diagnosis, request, or needs of billing. An important issue here is that different recipients might well need different types of information (e.g., diagnosis result, treatment, financial considerations, etc), and they likely would respond differently to the same type of information. Thus, in order to effectively deliver suitable message to different recipients, individuals, it is important to characterize recipients according to (a) the information they should be given (surface text), (b) any information they should be given as subtext, and (c) appropriate tone.
As used herein, the term “surface text” refers a literal or general meaning (e.g., dictionary definition, etc) of the phrase. The term “subtext” refers to any hidden or implicit meaning of the phrase that would be understood by the listener based on the context of the written communication as a whole or based on the prior communications, etc. The appropriate tone is determined contemplating the listener's potential emotional status (e.g., sad, happy, disappointed, etc), expected response to the information (e.g., resistive, admitting, etc), cultural diversity, and so on.
In the prior art, suitable messages are often prepared using forms. For example, many medical providers have computer systems that complete insurance forms. Although there may be different subparts depending on the diagnosis, treatment provided, and so forth, and although there may be different forms for different insurance companies, the bottom line is that someone in the medical office basically just fills out a form using available information.
Also in the prior art, when it comes time to instruct the patient with respect to treatments, many medical offices provide printed forms for the various different treatments. In the example above, for example, the office might well hand the patient a printed sheet with instructions on how to take medications that block uric acid production. Drugs called xanthine oxidase inhibitors, including Allopurinol or Febuxostat reduce uric acid, and Naproxen helps with pain and inflammation.
If instructions are to be provided to a caregiver (parent of a child, child of an elder parent, spouse, friend, etc) using prior art systems and methods, a medical office would again generally use a form, which might very well be the same instructional form that would be given to an independent patient.
All of this might work very well for simple and routine conditions and treatments. However, there is a trend towards providing more personalized service to patients, caregivers and others. And the need for personalization can increase with conditions and treatments that are less simple or less routine. For example, an elder patient might come into a doctor's office with his/her adult child or other caretaker. If the patient is deemed to have terminal cancer, it might be helpful to provide the patient and caretaker with generic brochures regarding cancer treatment options, but also to provide follow up letters. Such letters might have many different purposes, including providing more personalized information about the location and stage of this patient's condition, a well as specific information designed to protect the doctor and office against malpractice claims.
One way of accomplishing the goal of creating suitable messages is for the doctor, nurse, or other medical professional to speak into the system, perhaps during a conversation with the patient or caregiver, information about the desired letters, using keywords to identify the recipient (a) surface text, (b) the subtext, and (c) the appropriate tone. For example, the doctor could say the following to Mrs. Jones, and to Nancy, her caregiver adult child.
In a contemplated embodiment of the inventive subject matter, the system could key on several words in the doctor's speech to assist in drafting the letter. For example, from the doctor's speech the system could infer that the surface text should include the type of cancer, the treatment options discussed, and what can be expected with each of the different options. The subtext is that the patient could very well make a full recovery, and the tone is upbeat.
The doctor might then speak separately to Nancy, out of earshot of Mrs. Jones.
For this speech the system could infer that the surface text should include the prognosis, and the subtext should be that Nancy and her mother should seriously consider refusing all treatment. The system could infer that the tone of the letter to Nancy should be sad, and possibly apologetic that medical science doesn't have much to offer.
In another aspect of the inventive subject matter, a contemplated system could infer what would be appropriate surface text, subtext and tone from someone other than the doctor or other medical professional. For example, if a patient appears to be confused by terms the medical professional is using, the surface text could be a very simplified, dumbed-down version of the diagnosis, treatment and prognosis, the subtext could be that the patient should avoid reliance on any self-diagnosis, and the tone could be paternal. As another example, it may be that the patient is experiencing considerable denial regarding his condition, and consequently has a serious argument in the doctor's office with a caregiver spouse or friend. From that argument the system could infer that a summary to the patient should provide only superficial information, with subtext that the patient needs to listen to direction provided by the caregiver, and that the tone should be non-confrontational.
Rather than inferring the text, subtext and tone from the doctor's speech, those things could be derived from a database that relies somewhat or even entirely on correlations among relevant factors previously entered into a medical records system, including: diagnosis, treatment, prognosis, patient age, general physical condition, habits (smoking, exercise, etc), whether the recipient is the patient, adult or child caregiver, insurance company, employer, etc.
It is greatly preferred that the contemplated systems would not use “stock phrases” to generate written output. Rather, modern, machine-learning based techniques such as phrase-based or DNN-based machine translation techniques would be employed.
On the other hand, if stock phrases are used, they should be selected to conform to a desired impact respect to surface text, subtext and tone, and that impact derived through inference or otherwise, the automated system can access to a database of a plurality of stock phrases. The stock phrases can include any sentences (complete, incomplete, or partial sentences, etc), phrases, or group of words that can be used to generate a written communication. In some embodiments, at least some of the stock phrases can be tagged with one or more key words that represent the appropriate tone, the surface text, or the subtext such that stock phrase can be sorted/grouped/pulled based on the tagged languages. The types of keywords can vary, including the name of the disease, listener's age or status, level of comfort (high to low), level of explicitness (explicit to implicit), etc. In these embodiments, it is also preferred that the stock phrases are pre-paired with one or more keywords indicating the diagnostic information, conditions of the patients, environment of the patients (e.g., family environment, social status, etc).
Based on the desired impact, inferred diagnostic information, the automated system can select one or more stock phrases, place them in an appropriate order, and generate a written communication to an individual listener. For example, for delivering messages to a patient who is 50 years old and got diagnosed with the third stage of lung cancer, the written communication can comprise a group of stock phrases with encouraging tones so that the patient understands that he might have a serious disease but could overcome by diligently receiving treatments. For other example, for delivering messages to a patient's family, where the patient is 90 years old and got diagnosed with the terminal stage of brain cancer, the written communication can comprise a group of stock phrases that accurately delivers the diagnostic information, expected progress of the cancer in next several months, and what the family members can do for the patient during those period of time.
In some embodiments, the automated system can determine when the written communication should be delivered and designate the future time points for delivery. For example, the automated system can generate multiple written communications by estimating progress/situations of the status or treatment progress of the patient, and send one or more written communications at a different time point than others (e.g., one each per month, one each after each treatment period (e.g., each stage of chemotherapy, etc)). In these embodiments, it is also contemplated that the written communications that are supposed to be sent in a later time point can be automatically updated based on the changed status of patients, progress of the treatment, or response to the earlier written communications (e.g., to the family members, to the patients, etc).
It should be apparent to those skilled in the art that many more modifications besides those already described are possible without departing from the disclosed concepts herein. The disclosed subject matter, therefore, is not to be restricted except in the spirit of the appended claims. Moreover, in interpreting both the specification and the claims, all terms should be interpreted in the broadest possible manner consistent with the context. In particular, the terms “comprises” and “comprising” should be interpreted as referring to elements, components, or steps in a non-exclusive manner, indicating that the referenced elements, components, or steps can be present, or utilized, or combined with other elements, components, or steps that are not expressly referenced. Where the specification claims refers to at least one of something selected from the group consisting of A, B, C . . . and N, the text should be interpreted as requiring only one element from the group, not A plus N, or B plus N, etc.
In some embodiments, keywords extracted from a doctor-patient conversation can be correlated with more potential diagnoses, using the Conversation Keyword Table (Table 1). For example, keywords for communication between doctor and patient suffering from breast cancer can conclude that a patient might have BC3. Some embodiments may have different tables for each different diagnosis, or a single table with an extra column to designate diagnosis. Moreover, different signs and symptoms can be weighted differently, and negative answers can be weighed differently from positive answers.
In some embodiments, potential diagnoses could then be correlated with potential treatments using the Potential Treatments Table (Table 2). Potential treatments could then be correlated with potential prognoses using the Potential Prognoses Table (Table 3). For example, based on the consultation, the system could infer that the patient might have BC3 type or stage of cancer, and needs further tests to confirm. For a BC3 patient, treatment options TX2, TX3 are currently available and recommended. Based on the result of treatment options TX2, TX3, the prognosis of the patient is Prgo3 and Prgo4.
In some embodiments, the surface text of the summary could include potential diagnoses, potential treatments, and potential prognoses. The specific phrases chosen could be taken from the Diagnosis to Phrase (Table 4), the Treatment to Phrase (Table 5), and the Prognosis to Phrase (Table 6) tables. Different phrase tables may have different columns that provide phrasing specific to different types of recipients, according to the recipient relationship to the patient. To the extent that the doctor (or other medical professional) expressly stated a diagnosis, treatment or prognoses, then the system could jump directly to the phrase tables.
In some embodiments, from the other keywords spoken in the conversation, as well as biographical information, diagnostic and treatment options, the system can use the Subtext Table (Table 7) to determine Appropriate Emotional Context (Tone). From the Appropriate Emotional Context (Tone), the system could use the Tone Table (Table 8) to find phrases that could be included in a summary, according to the type of recipient. A summary can be generated using appropriate surface text, subtext and tone. In other embodiments, there are surface text phrases, tone phrases, but no specific subtext phrases, since subtext phrases are already incorporated into the tone phrases.
This application claims priority under 35 U.S.C. 119(e) from U.S. Provisional Patent Application Ser. No. 62/553,071, titled “Artificial Intelligence Scribe”, filed on Aug. 31, 2017.
Number | Date | Country | |
---|---|---|---|
62553071 | Aug 2017 | US |