The following relates generally to the imaging arts, remote imaging assistance arts, remote imaging examination monitoring arts, and related arts.
Medical imaging, such as computed tomography (CT) imaging, magnetic resonance imaging (MRI), positron emission tomography (PET) imaging, fluoroscopy imaging, and so forth, is a critical component of providing medical care, and is used in a wide range of medical fields, such as cardiology, oncology, neurology, orthopedics, to name a few. The operator of the medical imaging device used to acquire the medical images is typically a trained technologists, while interpretation of the medical images is often handled by a medical specialist such as a radiologist. Interpretation of radiology reports or findings by the radiologist can be handled by the patient's general practitioner (GP) physician or a medical specialist such as a cardiologist, oncologist, orthopedic surgeon, or so forth.
Currently, diagnostic imaging is in high demand. As the world population ages, the demand for quick, safe, high quality imaging will only continue to grow, putting further pressure on imaging centers and their staff. Under such conditions, errors are unavoidable, but can be often costly. One approach for imaging centers to boost efficiency and grow operations at no extra labor costs is through a radiology operations command center (ROCC) system. Radiology operations command centers enable teams to work across the entire network of imaging sites, providing their expertise as needed and remotely assisting less experienced technologists in carrying out high quality scans. Remote technologists or experts can monitor the local operators of scanning procedures through cameras installed in the scanning areas (or from other sources, such as sensors (including radar sensors), console video feeds, microphones connected to Internet of Things (IoT) devices, and so forth. In addition, these sources can be supplemented by other data sources like Health-Level 7 (HL7), Digital Imaging and Communications in Medicine (DICOM), Electronic Health Record (EHR) databases, and so forth.
The remote technologist (i.e. “super-tech”) is expected to be concurrently assigned to assist a number of different imaging bays at different sites that may be spread out across different cities or different states. In practice, however, the super-tech can only be paying attention to a single imaging bay at any given time. The super-tech will typically be assisting local technologists who actively call for super-tech support. However, situations may arise in which the super-tech's assistance would be beneficial, but the local technologist is unaware of the need for super-tech assistance, or chooses not to call for such assistance.
Current methods of monitoring imaging workflow via human oversight are also prone to errors, such as simple and subtle cues about workflow events might be easy to miss if the remote expert user is monitoring multiple imaging rooms at the same time. For example, events specific to the patient, such as a need for specialized patient transport, imaging setting adjustments for overweight or underweight patients, patient distress situations, difficulties with administration of a contrast agent to the patient, and so forth may be missed at least initially, leading to delays and/or sub-optimal examination results.
The following discloses certain improvements to overcome these problems and others.
In one aspect, a non-transitory computer readable medium stores instructions executable by at least one electronic processor to perform an assistance method during a medical imaging examination performed using a medical imaging device. The method includes acquiring video and/or audio feeds of the imaging examination; tracking a workflow of the imaging examination based at least on the acquired video and/or audio; detecting an event related to the workflow of the imaging examination at least based on analysis of the video and/or audio using an artificial intelligence (AI) component; determining a modification of the workflow of the imaging examination based on the detected event; and automatically executing the modification.
In another aspect, a non-transitory computer readable medium stores instructions executable by at least one electronic processor to perform an assistance method during a medical imaging examination performed using a medical imaging device. The method includes: tracking a workflow of the imaging examination based at least on acquired video and/or audio of the imaging examination; detecting an event related to the workflow of the imaging examination at least based on analysis of the video and/or audio; determining a modification of the workflow of the imaging examination based on the detected event; displaying an indication of the modification on an electronic processing device operable by a local operator performing the medical imaging examination and automatically executing the modification.
In another aspect, an apparatus is disclosed for providing assistance from a remote expert to a local operator during a medical imaging examination performed using a medical imaging device disposed in a medical imaging bay. The apparatus includes an electronic processing device operable by the remote expert, at least one of a camera and a microphone disposed in the medical imaging bay, and at least one electronic processor. The at least one electronic processor is programmed to: acquire video and/or audio of the imaging examination using the camera and/or the microphone; track a workflow of the imaging examination based at least on the acquired video and/or audio; detect an event related to the workflow of the imaging examination at least based on analysis of the video and/or audio using an artificial intelligence (AI) component; determine a modification of the workflow of the imaging examination based on the detected event; and automatically execute the modification.
One advantage resides in providing a remote expert or radiologist assisting a technologist in conducting a medical imaging examination with positional awareness of local imaging examination(s) which facilitates providing effective assistance to one or more local operators at different facilities.
Another advantage resides in improving efficiency of assistance from a remote expert to one or more local operators by providing the remote expert with automated alerts calling attention to events occurring in imaging workflows being performed by the local operators that may suggest remote expert assistance would be beneficial.
Another advantage resides in providing alerts to a remote expert of events occurring during a procedure operable by a local operator.
Another advantage resides in providing a remote reviewer, such as a remote operator providing imaging examination support from a remote service center, or a remotely located radiologist providing image quality review during the imaging examination, with awareness of the medical imaging examinations which facilitates providing assistance to a one or more local operators at different facilities, and may enable the remote reviewer to provide preemptive corrective advice to the local imaging technologist.
Another advantage resides in avoiding delaying of imaging procedures based on incorrect examination settings.
Another advantage resides in increasing patient safety and comfort during a medical imaging examination.
Another advantage resides in detecting events occurring in an imaging bay during a medical imaging examination that impact examination workflow.
Another advantage resides in automatically adjusting an imaging examination workflow based on one or more such detected events during the examination.
A given embodiment may provide none, one, two, more, or all of the foregoing advantages, and/or may provide other advantages as will become apparent to one of ordinary skill in the art upon reading and understanding the present disclosure.
The disclosure may take form in various components and arrangements of components, and in various steps and arrangements of steps. The drawings are only for purposes of illustrating the preferred embodiments and are not to be construed as limiting the disclosure.
The following discloses leveraging video feeds of a medical imaging device bay, and possibly other input such as an audio feed provided by a microphone, to automatically detect situations in which super-tech assistance may be beneficial and to provide alerts to the super-tech (and possibly also to the local technologist in the medical imaging device bay) in such cases.
The alerting approach utilizes video feeds provided from the imaging bays to which the super-tech is assigned. Microphones are also contemplated to be provided in the imaging bays to provide audio feeds. Video and/or audio feeds may also be provided from other areas, such as the waiting area.
The various video and audio feeds are automatically processed to detect situations in which super-tech assistance may be beneficial, based on detected features such as locations of individuals in the video feeds and their clothing (from which the patient, imaging technologist, nurse, or other roles are identified), detection of relevant hardware (e.g. imaging coils) being used (or not being used) in the imaging examination, patient movements of various types, loud voices in the audio feed, verbal calls for assistance in the audio feed, and so forth.
Additionally, the alerting system preferably tracks the workflow. This may be done based on analysis of the video feeds (e.g., detecting whether the patient is in the bore and therefore in the imaging stage) and timestamps or sequences of such detected features (e.g., if the patient is currently outside the bore and was previously inside the bore then the workflow stage is likely to be the patient unloading stage; whereas, if the patient is currently outside the bore and has not previously been inside the bore then the workflow stage is likely to be the patient loading stage. The alerting system also preferably has information about the specific imaging examination being performed, for example obtained from information contained in an electronic schedule of imaging examinations assigned to the super-tech.
The alerts are then identified based on the detected video and audio features and the workflow tracking. Alerts are typically sent to the remote medical expert or super-tech, and may also be sent to the local technologist. Optionally, alerts may be rated based on severity and the manner of the alerts chosen based on the severity. For example, an alert rated as critical may be issued as an audio alarm in combination with a flashing textual alert; whereas an alert rated as low priority may be issued as a pop-up window or other less obtrusive notification shown on the super-tech's workstation.
In other embodiments, this concept may be extended to other types of intervention based on analysis of audio and/or video, such as: detecting a wheelchair or gurney and modifying the workflow by issuing a call for a nurse or other assistant when the patient is finishing up the examination; detecting obesity of the patient and altering the workflow by suggesting different imaging parameters, scan times, or so forth; detecting an indication of patient distress that is not effectively resolved by the local imaging technician and in response putting a remote expert into direct audio communication with the patient; monitoring an intravascular contrast agent injector to detect an issue with the contrast agent delivery and generating an appropriate workflow modification; detecting an issue recognized during a pre-exam physical assessment of the patient and generating an appropriate workflow modification; and/or so forth.
To handle these complex interventions, a You Only Look Once (YOLO) or another object detection algorithm is used to implement stock and/or specially trained object detectors for relevant objects such as the patient, technician, imaging device, and so forth, optionally along with a combination of speech-to-text conversion and semantic analysis with a package such as Bidirectional Encoder Representations from Transformers (BERT). An object-to-workflow event mapping associates detected object(s) with workflow modifications, and a workflow streamliner/event trigger module then interfaces with information technology (IT) equipment to implement the workflow modification(s).
In a further aspect, a user feedback module may be provided, which provides user feedback as to the effectiveness of the workflow modifications introduced by the system. This is envisioned in some embodiments to be fully automated, e.g. the machine learning (ML) components can be adaptively trained during use based on the user feedback.
ROCC implementations typically include a bay camera (e.g., viewing through the window between the control room and the scanner room), an optional second camera capturing the contrast agent injector display (in examinations that employ intravascular contrast agent administration), and possibly also a bore camera to image the patient in the bore, along with a microphone in the scanner room. Hence, these are the illustrative inputs to the system.
With reference to
The image acquisition device 2 can be a Magnetic Resonance (MR) image acquisition device, a Computed Tomography (CT) image acquisition device; a positron emission tomography (PET) image acquisition device; a single photon emission computed tomography (SPECT) image acquisition device; an X-ray image acquisition device; an ultrasound (US) image acquisition device; or a medical imaging device of another modality. The imaging device 2 may also be a hybrid imaging device such as a PET/CT or SPECT/CT imaging system. While a single image acquisition device 2 is shown by way of illustration in
To provide for contrast-enhanced imaging, a contrast injector 11 is configured to inject the patient with a contrast agent. The contrast injector 11 is a configurable automated contrast injector having a display 13. The user (usually the imaging technologist) loads a vial or syringe of contrast agent (or two, or more, vials of different contrast agent components) into the contrast injector 11, and configures the contrast injector 11 by entering contrast injector settings such as flow rates, volumes, time delays, injection time durations, and/or so forth via a user interface (UI) of the contrast injector 11. The UI may be a touch-sensitive overlay of the display 13, and/or physical buttons, keypad, and/or so forth. In a variant embodiment, the contrast injector 11 is integrated with the imaging device controller 10 (e.g., via a wired or wireless data connection), and the contrast injector 11 is controlled via the imaging device controller 10, including displaying the contrast injector settings in a (optionally selectable) window on the display of the imaging device controller 10. In such an embodiment, the dedicated physical injector display 13 of the contrast injector may optionally be omitted (or, alternatively, the dedicated physical injector display 13 may be retained and the contrast settings displayed at both the dedicated physical injector display 13 and at the imaging device controller 10). In general, the automated contrast injector 11 can employ any suitable mechanical configuration for delivery of the contrast agent (or agents), such as being a syringe injector, a dual-syringe injector, pump-driven injector, or so forth, and may include hardware for performing advanced functions such as saline dilution of the contrast agent, priming and/or flushing of the contrast injection line with saline, and/or so forth.
As used herein, the term “medical imaging device bay” (and variants thereof) refer to a room containing the medical imaging device 2 and also any adjacent control room containing the medical imaging device controller 10 for controlling the medical imaging device. For example, in reference to an MM device, the medical imaging device bay 3 can include the radiofrequency (RF) shielded room containing the MM device 2, as well as an adjacent control room housing the medical imaging device controller 10, as understood in the art of MRI devices and procedures. On the other hand, for other imaging modalities such as CT, the imaging device controller 10 may be located in the same room as the imaging device 2, so that there is no adjacent control room and the medical bay 3 is only the room containing the medical imaging device 2. In addition, while
As diagrammatically shown in
The communication link 14 also provides a natural language communication pathway 19 for verbal and/or textual communication between the local operator and the remote operator. For example, the natural language communication link 19 may be a Voice-Over-Internet-Protocol (VOIP) telephonic connection, an online video chat link, a computerized instant messaging service, or so forth. Alternatively, the natural language communication pathway 19 may be provided by a dedicated communication link that is separate from the communication link 14 providing the data communications 17, 18, e.g. the natural language communication pathway 19 may be provided via a landline telephone. In some embodiments, the natural language communication link 19 allows a local operator LO to call a selected remote expert RE. The call, as used herein, can refer to an audio call (e.g., a telephone call), a video call (e.g., a Skype or Facetime or other screen-sharing program), or an audio-video call. In another example, the natural language communication pathway 19 may be provided via an ROCC device 8, such as a mobile device (e.g., a tablet computer or a smartphone), or can be a wearable device worn by the local operator LO, such as an augmented reality (AR) display device (e.g., AR goggles), a projector device, a heads-up display (HUD) device, etc., each of which having a display device 36. For example, an “app” can run on the ROCC device 8 (operable by the local operator LO) and the remote workstation 12 (operable by the remote expert RE) to allow communication (e.g., audio chats, video chats, and so forth) between the local operator and the remote expert.
The medical imaging device controller 10 in the medical imaging device bay 3 also includes similar components as the remote workstation 12 disposed in the remote service center 4. Except as otherwise indicated herein, features of the medical imaging device controller 10, which includes a local workstation 12′, disposed in the medical imaging device bay 3 similar to those of the remote workstation 12 disposed in the remote service center 4 have a common reference number followed by a “prime” symbol, and the description of the components of the medical imaging device controller 10 will not be repeated. In particular, the medical imaging device controller 10 is configured to display a GUI 28′ on a display device or controller display 24′ that presents information pertaining to the control of the medical imaging device 2, such as configuration displays for adjusting configuration settings an alert 30 perceptible at the remote location when the status information on the medical imaging examination satisfies an alert criterion of the imaging device 2, imaging acquisition monitoring information, presentation of acquired medical images, and so forth. It will be appreciated that the screen mirroring data stream 18 carries the content presented on the display device 24′ of the medical imaging device controller 10. The communication link 14 allows for screen sharing between the display device 24 in the remote service center 4 and the display device 24′ in the medical imaging device bay 3. The GUI 28′ includes one or more dialog screens, including, for example, an examination/scan selection dialog screen, a scan settings dialog screen, an acquisition monitoring dialog screen, among others. The GUI 28′ can be included in the video feed 17 and displayed on the remote workstation display 24 at the remote location 4.
Furthermore, as disclosed herein, the server 14s performs a method or process 100 for providing assistance during a medical imaging examination performed using a medical imaging device 2 (i.e., by assisting local operators LO of respective medical imaging devices 2 during medical imaging examinations by a remote expert RE). The instructions to perform the method 100 are stored in the non-transitory computer readable medium 26 of the remote workstation 12.
With reference to
At an operation 102, as the feed(s) 17, 18 is/are received at the ROCC center on the remote workstation 12, the GUI 28 is provided on the display device 24 of the remote workstation 12 to display the feed(s) 17, 18. In some examples, the feed(s) 17, 18 can include a plurality of feeds at a corresponding number of imaging bays 3. At an optional operation 103, a workflow of the medical imaging examination is tracked. At an operation 104, an event occurring the medical imaging bay 3 is detected based on analysis of the feed(s) 17, 18 (and based on analysis of the tracked workflow where applicable). At an operation 106, an alert 30 is output on the remote workstation 12 indicating the detected event. In some examples, the alert 30 can also be output to the local operator LO.
In some embodiments, operation 104 includes detecting an event based on the analysis of the feed(s) 17, 18 and the analysis of the workflow provided by the tracking operation 103. In one example, the workflow tracking 103 includes analyzing the feed(s) 17, 18 to track the events in the workflow, and the detected event can be a workflow event of the workflow (e.g., the contrast injector 11 activating is an event of the workflow). In another example, the detecting of the event includes analyzing timestamps of the tracked workflow to detect the event.
In other embodiments, the feed 17, 18 includes at the least the video feed 17 acquired by the camera 16 of the imaging bay 3. In one example, the detected event includes a location or position of an individual (e.g., the local operator LO, technologist aides, the patient, and so forth) in the video feed 17 by analysis of the video feed, and the corresponding alert 30 indicates that the location or position of the individual is incorrect. In some examples, to protect patient privacy, instead of conveying a live video feed 17 of the imaging bay 3, the video feed can be processed, and individuals can be replaced by abstract figures in the video feed, with their appropriate roles marked (e.g. color-coded) based on their appearance and actions.
In one example, the video feed 17 is analyzed to confirm a correct positioning of the patient within the medical imaging device 2. To do so, the patient's position can be monitored in real time, and contrasted with the requirements dictated by the imaging workflow (e.g., headfirst vs. feet first, etc.). If the patient is positioned incorrectly, a corresponding alert 30 can be output.
In another example, the video feed 17 is analyzed to detect whether the patient is wearing, for example, protective headphones (not shown), and has an alert ball (not shown) in hand. (An alert ball is a handheld alert button, squeezable ball, or the like which a patient loaded into an MM bore or other constricted imaging examination space can press, squeeze, or otherwise easily activate to generate an alert calling for assistance). A corresponding alert 30 can be output if the medical imaging examination is about to start and the patient is not wearing the protective headphones or does not have the alert ball in hand.
In another example, an alert 30 can be output if the patient is not wearing appropriate clothing compatible with the medical imaging device 2 (e.g., leisure wear vs. a hospital gown; or, a watch or other metal-containing wearable may be detected in the video, which is not permitted in an MM examination).
In yet another example, the video feed 17 of an MM examination is analyzed to detect whether the patient's hands are clasped. This can be a problem because time-varying magnetic fields generate electrical currents, and when the patient's hands are clasped, this can form a closed electrical conduction loop, which can lead to excessive peripheral nerve stimulation. Hence, a corresponding alert 30 can be output if patient's hands are detected to be clasped based on video feed analysis.
In other examples, the detected event includes movement of a patient during an imaging examination detected by analysis of the video feed 17, and the corresponding alert 30 indicates patient movement. In one example, the detected event includes whether patient is attempting to get up (or otherwise move) from during the medical imaging examination. A corresponding alert 30 can be output to indicated movement, and also may include an indication to review the images and potentially restart the examination if the images are unusable due to the detected patient motion.
In further embodiments, the detected event includes an issue with either the medical imaging device 2 itself, or with a workflow step of the medical imaging device. In one example, the detected event comprises a misplacement of a hardware component of the imaging device 2 detected by analysis of the video feed 17, and the corresponding alert 30 indicates the hardware component is misplaced. In other examples, the detected event can include malfunction hardware components, difficulty with positioning or attaching hardware components, or with workflow steps (e.g., how long it takes to position the patient on the medical imaging device 2, how long it takes to clean the room following the examination, a time stamp for when the contrast is being injected, etc.). A corresponding alert 30 can then be output.
In other embodiments, the feed 17, 18 includes at the least the audio feed 18 acquired by the microphone 15 of the medical imaging device controller 10. In one example, the detected event comprises a malfunction of a hardware component of the imaging device 2 detected by analysis of the audio feed 18 (e.g., detected as excessive vibrational noise or the like), and the corresponding alert 30 indicates the hardware component is malfunctioning. In another example, the audio feed 18 is analyzed to troubleshoot the medical imaging device 2, and a corresponding alert 30 is output. In a further example, the audio feed 18 is analyzed to detect whether the patient has squeezed an alarm ball, and a corresponding alert 30 can be output.
In further embodiments, the detected event includes a verbal noise, such as a call for help or assistance from the local operator LO, such as for assistance with a workflow step or to address a combative situation. A corresponding alert 30 can then be output. This could be done based on total audio energy in the 20-20,000 Hz range (or some subset thereof), and/or by applying speech recognition to the audio to extract semantic textual content that may trigger an alert, such as detecting the word or phrase “help” or “need assistance”. In another example, the audio feed 18 can be analyzed to determine one or more baseline audio characteristics. The detected event comprises unusual audio in the audio feed 18 detected by comparing the audio feed with the baseline audio characteristics. A corresponding alert can then be output to indicate that unusual audio (e.g., audio intensity above the baseline intensity, or audio frequency higher than baseline possibly indicating a person speaking in an elevated pitch) has been detected.
In other examples, the remote expert RE can mute the loudspeaker 29 of the remote workstation 12 (e.g., with a mute input) so that the audio feed 18 is not output, which may avoid an unnecessary distraction. As a result, once the audio feed 18 is analyzed, the alert 30 can be output to include an alert on the display device 24 of the remote workstation 12 to have the remote expert RE unmute the loudspeaker 29 to receive the audio feed 18 (or the remote workstation can be programmed to automatically unmute the loudspeaker when such an alert is output, or a flashing light or beep can be output to have the remote expert unmute the loudspeaker). This can be performed by providing a button on the GUI 28 for the patient to select with the at least one user input device 22 to unmute the loudspeaker 29.
In some embodiments, the alert output operation 106 can include determining a significance of the alert 30. For example, if multiple alerts 30 are output (e.g., one or more alerts output on the display device 24 of the remote workstation 12, and/or one or more alerts output via the loudspeaker 29), then the remote workstation 12 can analyze these alerts to determine a significance of the alerts. The alerts 30 can be ranked relative to each other, or the alerts can be ranked according to a predetermined significance threshold. For example, alerts that pertain to patient safety may be ranked higher than alerts that pertain to image quality. Alerts 30 having a higher significance can be output first to the remote expert RE.
In other embodiments, the alert output operation 106 can include correlating a response time to the alerts 30. The remote workstation 12 can analyze the alerts 30 to determine whether any of the alerts 30 need to be resolved more quickly than other output alerts. For example, during initial patient preparation, an alert 30 pertaining to the patient being positioned incorrectly may initially produce only an alert in the form of a text message. However, if the workflow is transitioning to the patient loading phase, then this alert 30 may be escalated, e.g. output as an audible warning that the patient is not correctly positioned. An alert 30 due to detected patient motion during an imaging sequence may not be initially presented at all, but only presented at the end of the imaging sequence as the technician is about to review the acquired images. This avoids distracting the local operator LO during image acquisition. As yet another example, detection that the patient's hands are clasped during MRI examination setup may produce a textual alert 30; whereas detection that the patient's hands are clasped during MRI image acquisition when the patient is being subjected to strong electromagnetic fields (thus creating an immediate patient safety concern) may result in an immediate audible alert to unclasp the hands. To do such analyses, a response time can be correlated to each output alert 30. In some examples, an alert 30 with a response time approaching expiration (e.g., mispositioned patient about to be loaded into the imaging device; or a patient with hands clasped as the imaging acquisition is about to be started; et cetera) can be escalated to the remote expert RE (e.g., by displaying a message on the display device 24, repeating the alert via the loudspeaker 29, and so forth). In other examples, if the remote expert RE who received the alert 30 cannot resolve the alert before expiration of the response time, the remote workstation 12 can transfer the alert 30 to an available remote expert for resolving.
Furthermore, as disclosed herein, the server 14s additionally or alternatively performs a method or process 200 for adjusting the workflow during a medical imaging examination performed using a medical imaging device 2 (i.e., by assisting local operators LO of respective medical imaging devices 2 during medical imaging examinations by a remote expert RE) based on events detected using the one or more cameras 16 and/or microphone 15. The instructions to perform the method 100 are stored in the non-transitory computer readable medium 26s of the server computer 14s.
With reference to
At an operation 202, the video feed 17 (acquired by the one or more cameras 16 and/or the audio feed 18 (acquired by the one or more microphones 15) are acquired, and are transferred to the server computer 14s. At an operation 204, a workflow of the imaging examination is tracked based on the acquired video feed 17 and/or audio feed 18.
At an operation 206, an event related to the workflow of the imaging examination occurring the medical imaging bay 3 is detected based on analysis of the feed(s) 17, 18 (and based on analysis of the tracked workflow where applicable). In some embodiments, the event detecting operation 206 can be performed using at least one artificial intelligence (AI) component 42 implemented in the server computer 14s. For example, the AI component(s) 42 can comprise a You Only Look Once (YOLO) model (for analyzing the video feed 17) and/or a Bidirectional Encoder Representations from Transformers (BERT) model (for analyzing the audio feed 18). In some examples, the event detecting operation 206 includes detecting an object 44 using the AI component(s) 42. As shown in
At an operation 208, a modification of the workflow of the imaging examination is determined based on the detected event. At an operation 210, the modification is automatically executed. These operations 208, 210 can be performed in a variety of manners.
In one example, the detected object 44 is a wheelchair, and the determined modification from the operation 208 includes a patient transport assistance request. The execution operation 210 then includes electronically transmitting the patient transport assistance request (e.g., via an audio intercom, or via an electronic scheduler, or outputting an audio or display alert suggesting ordering a wheelchair, etc.).
In another example, the detected object 44 includes a subject of the imaging examination, and the detected event includes a patient characteristic (e.g., obesity). The modification from the operation 208 includes changing at least one setting of the medical imaging device (2) based on the patient characteristic. The execution operation 210 then includes one of (i) displaying, on the display device 36 of the ROCC device 8, a recommendation of changing the at least one setting of the medical imaging device 2, or (ii) automatically changing the at least one setting of the medical imaging device 2 at the medical imaging device controller 10.
In another example, the detected event includes patient distress detected using the AI component 42 applied to the video and/or audio feed 17, 18. The modification from the operation 208 includes a patient assistance request. The execution operation 210 then includes electronically transmitting the patient assistance request to at least one of the local operator LO and/or the RE remote expert.
In another example, the acquired video and/or audio 17, 18 feed includes a video feed of the contrast injector 11. The detected event includes an event related to operation of the contrast injector 11 detected using the AI component 42 applied to the video feed 17 of the contrast injector 11. The modification from the operation 208 includes changing at least one setting of the contrast injector 11. The execution operation 210 then includes one of (i) displaying, on the display device 36 of the ROCC device 8, a recommendation of changing the at least one setting of the contrast injector 11, or (ii) automatically changing the at least one setting of the contrast injector 11.
These are merely illustrative examples, and should not be construed as limiting.
To perform the modification detection operation 208, a dictionary 48 (e.g., an object (or cue)-to-event dictionary) is stored in the server computer 14s. The server computer 14s is programmed to compare objects detected in the video feed 17 (or cues detected in the audio feed 18), and map these objects or cues to terms included in the dictionary 48.
In some embodiments, an effectiveness of the automatic execution of the modification in remediating the detected event can be automatically assessed, and the AI component(s) 42 can be updated based on the assessed effectiveness.
In other embodiments, an indication 46 of the modification can be displayed on the display device 36 of the ROCC device 8.
As previously discussed, the method 200 can be usefully employed in the context of detection of events relating to patient transport. For example, detection of wheelchair in an imaging room indicates the need for assistance in helping patient on/off the table. The camera(s) 16 may further suitably monitor the examination and alert relevant parties such as nurses or technologists when assistance is about to be required. In another example, if the patient entering for an imaging examination is detected to be obese, this might indicate that extra-long scans should be acquired and therefore the exam may take longer. The camera(s) 16 can detect this and alert the expert user to take remedial action such as transferring the upcoming exam to a different room since this one might be occupied longer than expected.
As another example, the microphone 15 can detect audio cues such as patient requesting help from the technologist verbally. This audio cue can be transmitted to the remote expert RE without requiring local tech to request help manually. Hardware such as audio splitters can allow the signal coming from the patient to the local tech to also go directly to the remote workstation 12 and thus to the remote expert RE. This communication can optionally be a two-way channel (allowing the remote expert RE to talk directly to patient) by connecting the microphone in the control room to the remote workstation 12 as well.
In another example, extravasation of contrast agent into tissue of the patient can be detected as an event using a bore camera of the camera(s) 16. This event can trigger an alert to the relevant parties or at least alert the remote expert user who then can inform staff as appropriate.
As yet another example, an incidental finding observed during the initial physical assessment of the patient in the imaging room before the patient is placed in the imaging device 2 might be detected as an event triggering a workflow modification in which a radiologist or other medical expert is called to review and possibly adjust the scheduled imaging protocol based on the incidental finding. The camera(s) 16 suitably detect cues indicating the incidental finding and automatically trigger an alert to the radiologist or other appropriate medical expert. In a variant embodiment, the detected cues trigger automatic generation of a radiology department assistance ticket (or other extant ticketing system providing efficient communication between technologists and radiologists within ROCC environment).
The disclosure has been described with reference to the preferred embodiments. Modifications and alterations may occur to others upon reading and understanding the preceding detailed description. It is intended that the exemplary embodiment be construed as including all such modifications and alterations insofar as they come within the scope of the appended claims or the equivalents thereof.
This application is a continuation-in-part of U.S. application Ser. No. 17/392,692 filed Aug. 3, 2021, which claims the benefit of U.S. Provisional Application No. 63/061,829 filed Aug. 6, 2020. These applications are hereby incorporated by reference herein.
Number | Date | Country | |
---|---|---|---|
63061829 | Aug 2020 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 17392692 | Aug 2021 | US |
Child | 17737076 | US |