Embodiments of the invention relate generally to a speech-driven patient care system as well as a mobile or portable communication units used therein.
Within a patient facility, such as a hospital, nursing home, etc., care providers, such as nurses, physicians, and assistants are valuable assets whose time must be allocated in an efficient manner for proper patient care and patient satisfaction. In the facility, a staff of nurses and other care providers are usually assigned a series of patients in individual rooms to whom they provide care. To that end, nurse call systems may be used provide a line of communication between the patient in a room and a nurse or other care provider.
One common nurse call system has a call button that is located in the patient's room and is accessible by the patient. When the button is pressed, a light outside the patient's room and/or an indicator light at a staff desk is typically turned on to visually indicate the request for help from the patient. This “patient call” is attended to if and when a nurse notices either the light outside the patient's room and/or at the staff desk. As may be appreciated, a significant amount of time might elapse from when the patient initiates the call and when the nurse actually responds.
In other nurse call systems, the patient presses a button to initiate two-way full duplex voice communications with a nurse station. The button is typically a call button, and the nurse then calls back to the patient to talk to them in an attempt to understand the problem before attending to the call. While such a system may provide more immediate attention to the patient's call, it is often disruptive and does not allow the nurse to prioritize or organize the call with respect to its urgency or the time of attention needed.
In still other nurse call systems, devices such as pagers, phones, and/or other telecommunications devices are integrated into the system. As such, the system sends out an alphanumeric message to the nurse that is wearing the pager, phone and/or telecommunications device or carrying the phone in response to receiving a patient call. However, as with call systems that utilize full duplex voice communications, until the nurse makes a call to the room, they do not know what the request or call was for and who needs to attend to it.
Thus, it is typical that nurse call systems require some immediate two-way voice communications to provide the nurse with an indication of what is needed by the patient. This is disruptive to the nurse or other care provider, as they may be in the middle of assisting another patient, performing a task, or otherwise unavailable. Thus, such systems are generally inefficient. Therefore, there is still a need in the art to improve upon the communication between a patient and a nurse or care provider that tends to them.
Embodiments of the invention provide a communication system for care providers and a method of managing patient care utilizing same. The system comprises a patient communication unit configured to be positioned proximate to a patient and operable to capture a patient call that includes speech input of the patient, a central console communicably coupled with the patient communication unit for receiving the call from the patient communication unit, and a portable communication unit configured to be carried by a care provider and communicably coupled with the central console, the portable communication unit operable to receive the call from the central console, to capture speech input of the care provider, and to play at least a portion of the patient speech input, the portable communication unit further operable to convert at least a portion of the care provider speech input into at least one command associated with the call.
These and other advantages will be apparent in light of the following figures and detailed description.
a is an illustration of a patient communication unit of
b is an illustration of a portable communication unit of
It should be understood that the appended drawings are not necessarily to scale, presenting a somewhat simplified representation of various preferred features illustrative of the basic principles of embodiments of the invention. The specific design features of the system and/or sequence of operations as disclosed herein, including, for example, specific dimensions, orientations, locations, and shapes of various illustrated components, will be determined in part by the particular intended application and use environment. Certain features of the illustrated embodiments may have been enlarged, distorted or otherwise rendered differently relative to others to facilitate visualization and clear understanding.
Generally, the central console 12 includes at least one processing unit (not shown) coupled to a memory (not shown). Each processing unit is typically implemented in hardware using circuit logic disposed in one or more physical integrated circuit devices, or chips. Each processing unit may be one or more microprocessors, micro-controllers, field programmable gate arrays, or ASICs, while the memory may include random access memory (RAM), dynamic random access memory (DRAM), static random access memory (SRAM), flash memory, and/or another digital storage medium, and also typically implemented using circuit logic disposed on one or more physical integrated circuit devices, or chips. The console 12 may further include a monitor 16 and one or more input devices, such as a keyboard 18 and/or mouse 20, to interface with a user (e.g., a care provider). Moreover, the console 12 may include transceiver circuitry 22 for communicably coupling to remote devices, such as a portable communication unit in the form of a badge 24 or headset 26 utilized to communicate with care providers, as well as a patient communication unit 28 (or patient communication “terminal” 28) configured in one or more patient rooms. In particular, the console 12 is communicably coupled with a unit 28 through a wired link as at 30 or a wireless link as at 32. Central console 12 is communicably coupled with the badge 24 and/or headset 26 of a care provider 34 through a wireless link as at 36. Hereinafter, the care provider 34 will be generically referred to as a user 34. Additionally, the person proximate the unit 28 or who is tended to by one or more users 34 will be generically referred to as a patient 38.
The unit 28 may be operated by a patient 38 and include circuitry to communication with the console 12 (e.g., such as circuitry to communicate through connection 30 and/or transceiver circuitry to communicate wirelessly through connection 32, neither shown), as well as a microphone 40 activated by a button 41 and a speaker 41. In some embodiments, the patient 38 communicates with the console 12 through the unit 28. In particular, the patient 38 may provide speech input that is captured by the microphone 40 of the unit 28. The call from the patient 38, and in particular the speech input of the patient 38, may then be provided, along with additional information (e.g., such as the location of the patient 38 and/or an identity of the patient 38), through console 12 and/or transceiver circuitry 22, to the user 34. In some embodiments, the user 34 may respond through their own speech input, which may include a command and/or message for the patient 38. In particular, the user 34 may respond through console 12, to unit 28. Alternatively, the user 34 may respond to the unit 28 directly through a wireless link as at 44.
For example, and as illustrated in a graphic representation 54 associated with the central console 12 and/or system 10, embodiments of the invention may evaluate the availability of various users 34, indicated as Nurse 1, Nurse 2, etc. As shown in
Once the user 34 associated with a call is determined, the call may be transmitted via link 36 to that user 34. Speech input of the patient 38 included with the call may then be played with an indication of the room and/or patient of origin (block 56). User 34 may be carrying a headset 26 and/or a badge 24 that provides the hardware for communicating with the central console 12 and/or the patient communication unit 28 consistent with embodiments of the invention. For example, the headset 26 and/or badge 24 may include a speaker for playing the patient speech input 56, and a microphone for capturing speech input of the user 34 (e.g., a “response”) (block 58). For example, and as illustrated in
In the example illustrated in
The device 24 and/or 26 may further include radio transceiver circuitry 80 to communicably couple with the transceiver circuitry 22 coupled to the central console 12 or with transceiver circuitry coupled to the unit 28. A power supply 82, such as at least one battery or other energy storage device, may provide the necessary power for the device 24 and/or 26. In some embodiments, each device 24 and/or 26 includes speech recognition/synthesis circuitry 84 (hereinafter, “speech circuitry” 84). The speech circuitry 84 may allow a user 34 to interface with the device 24 and/or 26 and voice a command to handle a call. In particular, the command may allow the user 34 to control the communication between themselves and a patient 38 consistent embodiments of the invention. For example, the speech circuitry 84 may process speech input captured from the microphone 70 and convert at least a portion of that speech input into at least one command for controlling the CPU 76 to handle a call from a patient 28. In some embodiments, the console 12 may also include speech circuitry 84. In alternative embodiments, the device 24 and/or 26 and/or the console 12 may not include speech circuitry 84. As such, the device 24 and/or 26 and/or the console 12 may be configured with program code stored in their respective memory and configured to be executed by a CPU or processing unit thereof to implement speech recognition functionality.
As illustrated in
In addition to prioritizing and/or re-prioritizing a call 90, the user 34 may utilize a “TRANSFER” command 94 to transfer a call 90 to another user 34 pursuant to a defined line of responsibility or hierarchy. Alternatively, the user 34 may specify the other user 34 that the call 90 should be transferred to. As indicated in
Moreover, in accordance with another feature of the invention, the user 34 may receive the call 90 and respond with an appropriate voice message that is played back through the patient communication unit 28 associated with the patient 38 who originated the call 90. In particular, the user 34 may utilize a “MESSAGE” command 96 to capture and supply a reply message to the patient 38 that originated the call 90. In that way, the user 34 may let the patient 38 know that they are responding to their call 90. This may put a patient 38 at ease, as the patient 38 may not be in a particular hurry for a non-urgent request, but would like to know that their call 90 has been received and is being handled.
Through the voice commands 92, 94, 96 and the speech circuitry 84 of the device 24 and/or 26, the user 34 may be able to address how a call is received and handled. If calls are prioritized, they may be attended to in a particular order, as determined by the user 34. Alternatively, the user 34 might pass responsibility for the call to another user 34 through the transfer command 94. Moreover, the user 34 may respond to the call. In some embodiments, this is all done in a hands-free manner while the user 34 attends to other tasks, such as charting or patient care. As noted above, the central console 12 may be aware of the status for each user 34 through information associated with event schedules, real-time activity, and room assignment, so that a line of responsibility might be established for automatic call transfers and call escalations, which may be controlled at the central console 12, as illustrated in
As illustrated in
In accordance with another feature consistent with embodiments of the invention, a full duplex voice dialog may be provided between a user 34 and a patient 38. In particular, the full duplex voice dialog may be provided between a device 24 and/or 26 and a patent communication unit 28. As such, the user 34 may utilize a “CONNECT” command (not shown) and specify the patent room 50 and/or patient 38 with whom to communicate with. For example, the nurse may speak, “CONNECT with Patient Room 320” in order for the device 24 and/or 26 to set up a communication channel for a full duplex voice dialog with the patient 38 in room 320. The full duplex voice dialog may be configured to communicate directly to the patient communication unit 28 from the device 24 and/or 26 carried by a user 34 (as at link 44), or to communicate from the device 24 and/or 26 to the patient communication unit 28 through the central console 12 (as at links 36, 30, and/or 32).
In accordance with another aspect consistent with embodiments of the invention, and as illustrated in
Once a user 34 has completed in answering a call, the user 34 may indicate as such to close documentation and/or tracking thereof. Referring to
In accordance with another feature consistent with embodiments of the invention, the device 24 and/or 26, in combination with the central console 12, monitor and record information associated with various stages that are in turn associated with a call. For example, the time a call is generated may be tracked and recorded, along with the time for a response from a user 34 and/or the identity of the user 34 that responds to the call. Additionally, any users 34 that are transferred a call, to from whom a call is transferred, may also be tracked and recorded. Furthermore, the time of completion of a call may also be tracked and recorded, including the time from which the user initially responds to the call to the time at which the call was completed. Other information associated with calls may also be tracked, including the commands spoken by users 34 and information associated with those commands, the room of origin, and/or the patient 38 of origin. For example, the various users 34 who respond to a call, who prioritize or transfer a call, and/or who actually receive the call and must respond to it, and also the responding user 34, may be monitored and recorded. Due to the ability of embodiments of the invention to not only prioritize calls but also transfer them to other users 34, or have them automatically transferred according to a hierarchal priority, a plurality of users 34 may be involved in the timeline from the beginning to completion of a call. All such information associated with the timing of the call and its completion, along with any user 34 involved, may be monitored and recorded for further review of the healthcare providers and management and to provide concrete evidence from a legal perspective and/or billing perspective, as well as to provide information to share with the patient 38 and/or family members, as illustrated in
In accordance with another feature consistent with embodiments of the invention, speech circuitry 84 of the device 24 and/or 26, speech circuitry 84 that may be alternatively disposed in the central console 12, and/or speech recognition functionality in the form of program code that can be executed by the processing units of the devices 24 and/or 26 and/or control unit 12 to convert speech input of the users 34 and/or patients 38, can provide a particular indication of the urgency of a call as well as the nature of such a call. For example, utilizing speech recognition functionality, the tone, timber, intensity, and/or voice pitch of speech input of a user 34 and/or patient 38 may be determined. From that information, it might be determined that the patient 38 is screaming, yelling, or otherwise making a noise consistent with distress, which may increase the priority and add urgency to the call. Furthermore, the speech recognition functionality may pick out certain words that provide a context to the call, and thus, provide an indication of its urgency. Based on the urgency of such a call, the central console 12 may transfer the call to an appropriate user or department, while simultaneously notifying a primary nurse with the call (e.g., when the central console 12 is configured with such speech recognition functionality). Moreover, a distinct audio tone might be utilized in the call preceding the playback of the speech input of the patient 38 to indicate the acuteness level of that particular call and/or voice message.
Turning to
Embodiments of the invention provide a user 34 with the ability to set vocal reminders and memos to remind them of daily events that may not necessarily be associated with the particular care or care plan in turn associated with a patient 38. For example, as illustrated in
In accordance with another feature consistent with embodiments of the invention, the central console 12 may run a program that provides pre-set patient self-care reminders and informational voice messages to the patient 38 in their room. The reminders and voice messages may be played through unit 28 in the patient room 50, as illustrated in
a is a pictorial representation of one possible patient communication unit 28 with a microphone 40 that may be used to capture speech input of a patient 38 and a speaker 42 to play voice messages to the patient 38.
Embodiments of the invention allow a nurse, user, or other care provider to receive patient calls that include patient speech input captured by in-room terminals, and allows the care provider to respond to the appropriate voice messages, wherein the responses may be played back through the same in-room terminal to the patient. The care provider has the ability to prioritize calls, transfer calls to other staff, or escalate calls as appropriate through voice commands in a hands-free manner. Embodiments of the invention also allows the care provider to establish full duplex voice dialog if they see the need to have such communications with the patient. The present invention further allows the care provider to set care plan reminders for a patient, to set and play memos for daily activities such as meetings, etc., and to do all this through voice activation and playback utilizing a user-worn device with speech recognition capabilities. In that way, patients can quickly communicate their requests for help, and the care provider can respond or forward the call appropriately. Embodiments of the invention allow care providers to use their discretion to prioritize more urgent calls without interrupting their work flow and their interactions with other patients.
The central console of the invention is aware of the status of each care provider through information from event schedules, real-time activity, room assignments, and thus embodiments of the invention allow a line of responsibility prioritization for automatic call transfer and call escalations. The central console may be further configured to monitor and record the timing for each event including call generation, the time of response, the time of completion of the call, and the users (e.g., staff) involved so that records may be kept with respect to each call and event, such as for the purpose of legal evidence, billing purposes, as well as patient and/or family information. Furthermore, embodiments of the invention provide pre-set patient self-care reminders (such as medication reminders), and also provides voice notifications to be played to the patient in their room. The in-room terminal provides the necessary speaker for such communications. While a healthcare provider might request such reminders and notifications, the patient can also request any number of care notifications to be set by the healthcare provider. The central console may also have the ability to decipher the urgency and nature of a call request based on the tone of speech and word or speech recognition, and to transfer a call to the appropriate care giver(s) or department, while simultaneously notifying a primary care giver with distinct tones to indicate the acuteness level of the patient voice message. In that way, as illustrated in
While the present invention has been illustrated by the description of the embodiments thereof, and while the embodiments have been described in considerable detail, it is not the intention of the applicant to restrict or in any way limit the scope of the appended claims to such detail. Additional advantages and modifications will readily appear to those skilled in the art. Therefore, the invention in its broader aspects is not limited to the specific details representative apparatus and method, and illustrative examples shown and described. Accordingly, departures may be made from such details without departure from the spirit or scope of the applicants' general inventive concept.
Other modifications will be apparent to one of ordinary skill in the art. Therefore, the invention lies in the claims hereinafter appended.
This application is related to and claims the benefit of U.S. Provisional Application No. 61/092,641 to Prakash Somasundaram et al., entitled “SPEECH-DRIVEN PATIENT CARE SYSTEM WITH WEARABLE DEVICES”, filed Aug. 28, 2008, which application is incorporated by reference herein.
Number | Name | Date | Kind |
---|---|---|---|
4213253 | Gudelis | Jul 1980 | A |
5077666 | Brimm et al. | Dec 1991 | A |
5465082 | Chaco | Nov 1995 | A |
5536084 | Curtis et al. | Jul 1996 | A |
5754111 | Garcia | May 1998 | A |
5822544 | Chaco et al. | Oct 1998 | A |
5838223 | Gallant et al. | Nov 1998 | A |
5877675 | Rebstock et al. | Mar 1999 | A |
5986568 | Suzuki | Nov 1999 | A |
D420674 | Powell | Feb 2000 | S |
6057758 | Dempsey et al. | May 2000 | A |
6292783 | Rohler et al. | Sep 2001 | B1 |
6442432 | Lee | Aug 2002 | B2 |
6591242 | Karp et al. | Jul 2003 | B1 |
6714913 | Brandt | Mar 2004 | B2 |
6720864 | Wong et al. | Apr 2004 | B1 |
6747556 | Medema et al. | Jun 2004 | B2 |
6772454 | Barry | Aug 2004 | B1 |
6849045 | Iliff | Feb 2005 | B2 |
6872080 | Pastrick | Mar 2005 | B2 |
6890273 | Perez | May 2005 | B1 |
7056289 | Kasper et al. | Jun 2006 | B2 |
7065381 | Jenkins | Jun 2006 | B2 |
7088233 | Menard | Aug 2006 | B2 |
7092376 | Schuman | Aug 2006 | B2 |
7136684 | Matsuura et al. | Nov 2006 | B2 |
7154397 | Zerhusen et al. | Dec 2006 | B2 |
7228429 | Monroe | Jun 2007 | B2 |
7263669 | Denholm | Aug 2007 | B2 |
7283845 | De Bast | Oct 2007 | B2 |
7287031 | Karpf | Oct 2007 | B1 |
D568881 | Hsiau | May 2008 | S |
D569358 | Devenish, III et al. | May 2008 | S |
D569876 | Griffin | May 2008 | S |
D573577 | Huang | Jul 2008 | S |
D583827 | Wahl | Dec 2008 | S |
7487101 | Vasko et al. | Feb 2009 | B1 |
7574370 | Mayaud | Aug 2009 | B2 |
7598852 | Chriss | Oct 2009 | B2 |
7598853 | Becker et al. | Oct 2009 | B2 |
D609246 | Wahl | Feb 2010 | S |
7664657 | Letzt et al. | Feb 2010 | B1 |
7706878 | Freeman | Apr 2010 | B2 |
7868740 | McNeely et al. | Jan 2011 | B2 |
7912733 | Clements et al. | Mar 2011 | B2 |
7978062 | LaLonde et al. | Jul 2011 | B2 |
8054177 | Graves et al. | Nov 2011 | B2 |
20020004729 | Zak et al. | Jan 2002 | A1 |
20020044043 | Chaco et al. | Apr 2002 | A1 |
20020101349 | Rojas, Jr. | Aug 2002 | A1 |
20020146096 | Agarwal et al. | Oct 2002 | A1 |
20020183598 | Teraura et al. | Dec 2002 | A1 |
20030050539 | Naghavi et al. | Mar 2003 | A1 |
20030130016 | Matsuura et al. | Jul 2003 | A1 |
20030160682 | Yamada et al. | Aug 2003 | A1 |
20040034286 | Kasper et al. | Feb 2004 | A1 |
20040122486 | Stahmann et al. | Jun 2004 | A1 |
20040122488 | Mazar et al. | Jun 2004 | A1 |
20040220686 | Cass | Nov 2004 | A1 |
20050033127 | Ciurczak et al. | Feb 2005 | A1 |
20050148890 | Hastings | Jul 2005 | A1 |
20050192845 | Brinsfield et al. | Sep 2005 | A1 |
20050228245 | Quy | Oct 2005 | A1 |
20050251213 | Freeman | Nov 2005 | A1 |
20060026035 | Younkes et al. | Feb 2006 | A1 |
20060253281 | Letzt et al. | Nov 2006 | A1 |
20060279427 | Becker et al. | Dec 2006 | A1 |
20060286960 | Goehler | Dec 2006 | A1 |
20060287014 | Matsuura et al. | Dec 2006 | A1 |
20070080801 | Weismiller et al. | Apr 2007 | A1 |
20070221138 | Mainini et al. | Sep 2007 | A1 |
20070247310 | Ulrich et al. | Oct 2007 | A1 |
20070276270 | Tran | Nov 2007 | A1 |
20070299473 | Matos | Dec 2007 | A1 |
20080018480 | Sham | Jan 2008 | A1 |
20080055071 | Chriss | Mar 2008 | A1 |
20080072153 | Yang | Mar 2008 | A1 |
20080072847 | Liao | Mar 2008 | A1 |
20080082363 | Habashi | Apr 2008 | A1 |
20080106374 | Sharbaugh | May 2008 | A1 |
20080164998 | Scherpbier et al. | Jul 2008 | A1 |
20080211665 | Mazar et al. | Sep 2008 | A1 |
20080224861 | McNeely et al. | Sep 2008 | A1 |
20080262867 | Bayne et al. | Oct 2008 | A1 |
20080262873 | Bayne et al. | Oct 2008 | A1 |
20080306740 | Schuck et al. | Dec 2008 | A1 |
20080316036 | Nixon | Dec 2008 | A1 |
20090063183 | McNeely et al. | Mar 2009 | A1 |
20090063193 | Barton et al. | Mar 2009 | A1 |
20090216534 | Somasundaram | Aug 2009 | A1 |
20090222065 | Dlugos et al. | Sep 2009 | A1 |
20090322514 | Monroe | Dec 2009 | A1 |
20100001838 | Miodownik et al. | Jan 2010 | A1 |
20100036667 | Byford et al. | Feb 2010 | A1 |
20100121157 | Espina et al. | May 2010 | A1 |
20100198608 | Kaboff et al. | Aug 2010 | A1 |
Number | Date | Country |
---|---|---|
1791053 | May 2007 | EP |
WO9525326 | Sep 1995 | WO |
WO02096126 | Nov 2002 | WO |
WO2005043303 | May 2005 | WO |
Entry |
---|
Forty-Seven-page “The Digital Consumer Technology Handbook”, A Comprehensive Guide to Devices, Standards, Future Directions, and Programmable Logic Solutions; by Amit Dhir, Xilinx, Inc. dated Feb. 27, 2004. |
Number | Date | Country | |
---|---|---|---|
20100052871 A1 | Mar 2010 | US |
Number | Date | Country | |
---|---|---|---|
61092641 | Aug 2008 | US |