This disclosure relates generally to ultrasound imaging.
The COVID-19 pandemic has emerged as a serious global health crisis, with the primary morbidity and mortality linked to pulmonary involvement. A prompt and accurate diagnostic assessment is thus crucial for understanding and controlling the spread of the disease, with point of care ultrasound scanning (“POCUS”) becoming one of the primary determinative methods for its diagnosis and staging. Although safer and more efficient than other imaging modalities, POCUS requires close contact of radiologists and ultrasound technicians with patients, subsequently increasing the risk for infections.
Tele-operated solutions allow medical experts to remotely control the positioning of an ultrasound probe attached to a robotic system, thus reducing the distance between medical personnel and patients to a safer margin. Several tele-operated systems have been successfully tested amidst the pandemic for various purposes. While they are a better alternative to traditional in-person POCUS, existing tele-operated systems nonetheless involve the presence of at least one healthcare worker in close vicinity of the patient to initialize the setup and assist the remote sonographer. Further, tele-operated systems require a skilled sonographer to remotely control the ultrasound probe, and such experienced technicians are in high demand during the pandemic.
According to various embodiments, a method of autonomously robotically acquiring an ultrasound image of an organ within a bony obstruction of a patient is presented. The method includes: acquiring an electronic three-dimensional patient-specific representation of the bony obstruction of the patient; obtaining an electronic representation of a target location in or on the organ within the bony obstruction of the patient; determining, automatically, position and orientation of an ultrasound probe to acquire an image of the target location; and directing, autonomously, and by a robot, the ultrasound probe on the patient to acquire the image of the target location based on the position and orientation.
Various optional features of the above method include the following. The method may include outputting the image of the target location. The bony obstruction may include a ribcage, and the organ may include at least one of: a lung, a heart, a spleen, a liver, a pancreas, or a kidney. The acquiring may include acquiring a three-dimensional radiological scan of the patient. The acquiring may include acquiring a machine learning representation of the bony obstruction of the patient based on a topographical image of the patient. The obtaining may include obtaining a human specified location in the electronic three-dimensional representation of the bony obstruction of the patient. The method may include: measuring a force on the ultrasound probe; and determining a position of the ultrasound probe, based on the force, relative to the bony obstruction of the patient. The determining may include determining position and orientation based on a weighted function of a plurality of material densities, and the plurality of material densities may include a bone density. The organ may include a lung, and the plurality of material densities may further include a density of air. An image of the target location may be acquired without requiring proximity of a technician to the patient.
According to various embodiments, a system for autonomously robotically acquiring an ultrasound image of an organ within a bony obstruction of a patient is presented. The system includes: an electronic processor that executes instructions to perform operations including: acquiring an electronic three-dimensional patient-specific representation of the bony obstruction of the patient, obtaining an electronic representation of a target location in or on the organ within the bony obstruction of the patient, and determining, automatically, position and orientation of an ultrasound probe to acquire an image of the target location; and a robot communicatively coupled to the electronic processor, the robot comprising an effector couplable to an ultrasound probe, the robot configured to direct the ultrasound probe on the patient to acquire the image of the target location based on the position and orientation.
Various optional features of the above system include the following. The operations may further include outputting the image of the target location. The bony obstruction may include a ribcage, and the organ may include at least one of: a lung, a heart, a spleen, a liver, a pancreas, or a kidney. The acquiring may include acquiring a three-dimensional radiological scan of the patient. The acquiring may include acquiring a machine learning representation of the bony obstruction of the patient based on a topographical image of the patient. The obtaining may include obtaining a human specified location in the electronic three-dimensional representation of the bony obstruction of the patient. The operations may further include: measuring a force on the ultrasound probe; and determining a position of the ultrasound probe, based on the force, relative to the bony obstruction of the patient. The determining may include determining position and orientation based on a weighted function of a plurality of material densities, and wherein the plurality of material densities may include a bone density. The organ may include a lung, and the plurality of material densities may further include a density of air. The robot may be configured to acquire an image of the target location without requiring proximity of a technician to the patient.
The above and/or other aspects and advantages will become more apparent and more readily appreciated from the following detailed description of examples, taken in conjunction with the accompanying drawings, in which:
Embodiments as described herein are described in sufficient detail to enable those skilled in the art to practice the invention and it is to be understood that other embodiments may be utilized and that changes may be made without departing from the scope of the invention. The present description is, therefore, merely exemplary.
The COVID-19 pandemic has emerged as a serious global health crisis, with the predominant morbidity and mortality linked to pulmonary involvement. Point of care ultrasound (POCUS) scanning, becoming one of the primary determinative methods for its diagnosis and staging, requires, however, close contact of healthcare workers with patients, therefore increasing risk of infection.
While tele-operated solutions, which allow medical experts to remotely control the positioning of an ultrasound probe attached to a robotic system, reduce the distance between medical personnel and patients to a safer margin, existing tele-operated ultrasound systems nonetheless require both a skilled remote sonographer and the presence of at least one healthcare worker in close vicinity of the patient to initialize the setup and assist the sonographer.
An autonomous robotic ultrasound system would better limit physical interaction between healthcare workers and infected patients, while offering more accuracy and repeatability to enhance imaging results, and hence patient outcomes. Further, an autonomous system would be a valuable tool for assisting less experienced health care workers, especially amidst the COVID-19 pandemic where trained medical personnel is such a scarce resource. However, existing autonomous ultrasound systems are insufficient for COVID-19 POCUS applications, which involve ultrasound imaging of the patient's lungs.
In general, robotic POCUS of lungs faces several difficulties, including: (a) the large volume of the organ, which cannot be inspected in a single ultrasound scan, implying that during each session, multiple scans from different locations may be sequentially collected for monitoring the disease's progression, (b) the scattering of ultrasound rays through lung air, meaning that an autonomous solution may be patient-specific to account for different lung shapes and sizes to minimize this effect, and (c) the potential obstruction of the lungs by the ribcage, which would result in an uninterpretable scan due to the impenetrability of bone material by ultrasound waves.
Embodiments disclosed herein solve one or more of the above problems. Some embodiments provide autonomous robotic POCUS ultrasound scanning of lungs, e.g., for COVID-19 patient diagnosis, monitoring, and staging.
Some embodiments robotically and autonomously position an ultrasound probe based on a patient's prior CT scan to reach predefined lung infiltrates. Some embodiments provide lung ultrasound scans using force feedback on the robotically controlled ultrasound probe based on a patient's prior CT scan.
Some embodiments predict anatomical features of a patient's ribcage using a surface torso model. Some embodiments utilize a deep learning technique for predicting 3D landmark positions of a human ribcage given a torso surface model. According to such embodiments, the landmarks, combined with the surface model, may be used for estimating ultrasound probe position on the patient for robotically and autonomously imaging infiltrates.
An experimental embodiments, described throughout this disclosure, acquired ultrasound scans with an average accuracy of 20.6±14.7 mm based prior CT scans, and 19.8±16.9 mm based on only ribcage landmark estimation using a surface model. A study of the experimental embodiment used on a full torso ultrasound phantom showed that the autonomously acquired ultrasound images were 100% interpretable when using force feedback with a prior CT and 87.5% with landmark estimation, compared to 75% and 58.3% without force feedback, respectively. This demonstrates the potential for embodiments to acquire accurate POCUS scans while mitigating the spread of COVID-19 in vulnerable environments.
As shown in
If a prior patient chest CT scan is available, then at 202 an expert radiologist or other medical technician marks on the CT regions of interest (typically but not necessarily containing infiltrates), which may then be observed over the course of coming days to evaluate the progression of the disease. The technician may be situated anywhere in the world relative to the patient and need not be present with the patient, or even in the same room or building, at the time of the ultrasound scan. At 204, an algorithm computes the spatial centroid of each region, as shown and described in detail in reference to
If a CT scan is not available, landmarks of the ribcage are estimated using the patient's 3D mesh model, which is generated at 208 from a topographical image of the patient obtained using a depth camera, e.g., depth camera 104. At 210, the location of the patient's ribcage is estimated by mapping landmarks. In the experimental embodiment, a neural network deep learning algorithm was used to estimate the ribcage landmarks from the 3D mesh model. Scanning points are then manually selected at 212 on the model following the 8-point POCUS protocol, e.g., by an expert radiologist or other medical technician. As with the CT scan technique, the technician may be situated anywhere in the world relative to the patient and need not be present with the patient, or even in the same room or building, at the time of the ultrasound scan. Goal positions and orientations are then determined at 214.
Note that in some embodiments, the method may utilize both a CT scan and a surface model as described above to generate probe positions and orientations at 214.
Whether a CT scan is used, a mesh model is used, or both, the probe positions and orientations of 214 are converted to robotic control signals for the robot frame at 216. Further description of such conversion is shown and disclosed herein in reference to
Due to possible kinematic and registration errors, the positioning of the ultrasound probe may suffer from unknown displacements, which can compromise the quality and thus interpretability of the ultrasound scans. To mitigate this, some embodiments employ a force-feedback mechanism through the ultrasound probe, e.g., using force/torque sensor 106, to avoid skeletal structures that will lead to shadowing. In more detail, a force-displacement profile is collected at 218, which is used to correct the end effector's position at 220 to avoid imaging of bones. The specifics of this protocol are further shown and described in detail herein, e.g., in reference to
Finally, the robot collects the ultrasound images at 222. For example, the robot executes the robotic control signals to achieve the ultrasound positions and orientations of an ultrasound probe relative to the patient to acquire images of the target location.
In the experimental embodiment, the ultrasound scanning position and orientation algorithm of 214, as well as the ribcage landmarks estimation of 210, were implemented in Python, whereas the robot control, which includes planning and data processing algorithms, were integrated via Robot Operating System as disclosed in Quigley, M., Conley, K., Gerkey, B., Faust, J., Foote, T., Leibs, J., et al. (2009), ROS: An open-source robot operating system, ICRA workshop on open source software (Kobe, Japan), vol. 3.2, 5. Kinematics and Dynamics Library (KDL) in Open Robot Control Systems (OROCOS) was used to transform the task-space trajectories of 214 of the robot to the joint space trajectories of 216, which are output by the high-level autonomous control system. The drivers developed by Universal Robot allow to apply the low-level controllers to the robot to follow the desired joint-space trajectories.
Details of techniques for determining ultrasound probe scanning position and orientation based on a CT scan are presented in this section.
The objective is to image the pre-determined target points within the lungs, while maximizing the quality of the ultrasound scan, which is influenced by three major factors; (a) proximity of the target to the ultrasound probe, (b) medium through which the ultrasound beam travels, and (c) number of layers with different acoustic properties the beam travels across. The quality of an ultrasound scan is enhanced as the target is closer to the ultrasound probe. In the particular case of lung scanning, directing beams through air should be avoided due to the scattering phenomenon, which significantly reduces the interpretability of the resulting scan. Skeletal structures reflect ultrasound almost entirely, prompting the user to avoid them. Lastly, layers of medium with different attenuation coefficients induce additional refraction and reflection of the signal, negatively impacting the imaging outcomes.
The problem can be formulated as a discrete optimization solved by linear search, whereby the objective is to minimize the sum of weights assigned to various structures in the human body through which the ultrasound beam travels, along with interaction terms modeling refraction, reflection and attenuation of the signal. Let {right arrow over (p)}i ∈2 represent the 2D coordinates of a pixel i inside the ultrasound beam cone (see 312 and 314 for cone reference). Let {right arrow over (p)}f
2 represent the pixel corresponding to the focal point of the ultrasound probe, that is, the point of origin of the ultrasound signal. Note that this pixel may be a virtual pixel that is not imaged. The attenuation of the ultrasound signal may be evaluated through the following equation:
In Equation (1), wi,c represents the weight of the first pixel pertaining to the same class c, α represents the attenuation coefficient of the medium, and Δ represents the spatial resolution of the CT scan. To model the intensity reflection of the ultrasound beam at the interface of two different mediums, first the intensity reflection coefficient γ is evaluated, and subsequently applied to the weight of the first pixel following the interface boundary:
In Equations (2) and (3), ρ represents tissue density, and ν represents the speed of ultrasound. The term ρν effectively represents the impedance of the medium, with medium 1 preceding medium 2. The algorithm thus evaluates the weight of every pixel in between the first point of contact of the ultrasound probe and the target point, with higher weights assigned to more attenuated pixels, as they would drastically reduce the image quality. At 314, the ultrasound path that results in the lowest weight, from among multiple paths generated at 312 whose respective cones include the target location, is selected, e.g., as the optimal one.
To this end, bones were assigned the highest weight of 109, since ultrasound rays cannot travel past them. The second highest weight was assigned to lung air at 5, followed by soft tissues at 1. These example weights are non-limiting. The assumed attenuation coefficients for skeletal tissue, lung air, and soft tissues are 1.1 dB/(mm×MHz), 1.2 dB/(mm×MHz) and 0.12 dB/(mm×MHz) respectively. The assumed densities of each class are 2000 kg/m3, 1.225 kg/m3 and 1000 kg/m3, whereas the speed of sound is 3720 m/s, 330 m/s, and 1575 m/s, respectively.
At 312, for a single image of the images of various orientations generated at 304, multiple possible scanning windows are generated, each containing the target point. The weights are computed within an ultrasound cone that can only be instantiated from the surface of the patient's body across all generated images. The algorithm first determines the scanning position and orientation, e.g., an optimal scanning position and orientation, of the probe for each individual image at 314, and finally selects the image with the overall lowest returned weight after 318. For each image, the position of the ultrasound cone, as well as the orientation of the image, define the selected, e.g., optimal, ultrasound scanning position and orientation in the CT coordinate frame, which is stored at 316. Finally, at 318, the process is repeated for each of the multiple images of various orientations obtained at 304. In the experimental embodiment, the solution was deployed on the Alienware laptop used for the robot control. Pseudocode for the ultrasound probe scanning position and orientation algorithm is presented below.
3, angle resolution Δβ = 5, Δ in mm
(k+1)×L
3 corresponding to wmin evaluated from plane
Uncertainties in patient registration and robot kinematics can result in a partial or complete occlusion of the region of interest due to the misplacement of the ultrasound probe. To mitigate this problem, some embodiments utilize a force-feedback mechanism. According to some embodiments, for a constant force application of, by way of non-limiting example, 20 N, which is the recommended value for abdominal ultrasound imaging, the probe's displacement is higher in-between the ribs as opposed to being on the ribs. Force feedback can thus be used to generate a displacement profile across the ribcage of a patient to detect regions obstructed by ribs. The displacement generally follows a sinusoidal profile, with peaks (i.e., largest displacements) corresponding to a region in-between the ribs, and troughs (i.e., smallest displacement) corresponding to a region on the ribs.
In particular, the force feedback mechanism was validated based on computer-generated solid models 402 of n=3 virtual patient torsos using anonymized CT scans, which were used to simulate displacements using finite element analysis (FEA) in ANSYS (ANSYS, Canonsburg, Philadelphia, USA). Two of the patients were female. The third patient was male, and was used as the model for the phantom's creation. All patients used for the solid models had varying BMI. The different organs of the patients were extracted from the CT scans using Materialize Mimics (Materialize NV, Southport, QLD 4222, Australia) software as STL files, and subsequently converted through SOLIDWORKS (SolidWorks Corp., Dassault Systemes, Velizy-Villacoublay, France) into IGS format, thus transforming the mesh surfaces into solid models that can undergo material assignment and FEA simulations. The tissues” mechanical properties for the female patients were obtained from the literature, whereas those of the phantom were measured experimentally. In the FEA simulations, the force was transmitted onto the bodies through a CAD model of the ultrasound probe. In the robotic implementation, the ultrasound probe did not slip on a patient's body when contact was established because of the setup's rigidity. A lateral displacement of the probe in simulation would be erroneously translated into soft tissue displacement since the probe's tip displacement was used to represent the soft tissue displacement. Thus, to ensure that the motion of the probe is confined to a fixed vector, the virtual probe's motion was locked in all directions except in the z-axis. The virtual force was directly applied to the virtual probe through a force load that gradually increases from 0 to 20 N over a period of five seconds. The virtual probe was initially positioned at a very close proximity from the torso, hence its total displacement was considered to be a measure of the tissue's displacement. The simulations were deployed on a Dell Precision workstation 3620 with an 7 processor and 16 GB of RAM. Each displacement data point required on average 2.5 hours to converge. The location of the collected data points, as well as returned displacement profile for all three virtual patients are shown at 408 of
To verify the outcome of the simulations, corresponding displacement profiles were collected from the physical phantom, which are shown at 406 of
Thus, some embodiments include a force-feedback mechanism in the robot's control process, whereby the system collects several displacement data points around the goal position at 20 N, to ensure that the ultrasound image is obtained in-between the ribs.
Some embodiments use 3D landmarks defined on the ribcage to estimate the appropriate, e.g., optimal, probe position. Some embodiments use a deep convolutional neural network trained to estimate the 3D position of 60 landmarks on the ribcage from the skin surface data. The trained 3D deep convolutional network directly estimated the landmark coordinates in 3D from the 3D volumetric mask representing the skin surface of a patient's torso.
In the experimental embodiment, the landmarks were defined using the segmentation masks of ribs obtained from the CT data. From the segmentation masks, the 3D medial axis for each rib was computed using a skeletonization algorithm. The extremities and center of each rib for the first 10 rib pairs (T1 to T10) were used as landmark locations. The three landmarks thus represent the rib-spine intersection, the center of the rib and the rib-cartilage intersection.
For training the deep neural network of the experimental embodiment, given the skin mask, a 3D bounding box covering the thorax region was estimated, using jugular notch on top and pelvis on the bottom. This region was then cropped and resized to 128×128×128 volume, which was used as input to a deep network. The network output a 3×60 matrix, which represents the 3D coordinates of the 60 rib landmarks. The experimental embodiment used the DenseNet architecture with batch normalization, and LeakyReLU activations with a slope of 0.01 following the 3D convolutional layers with kernel size of 5×5. The network parameters were optimized using AdaDelta.
The autonomous positioning of the ultrasound probe in contact with the patient's body utilizes motion and control planning, a non-limiting example of which is described presently in reference to
For notation, let TBA denote the homogeneous transformation matrix from frames A to B, composed of a rotation matrix RBA ∈SO(3), and a translation vector {right arrow over (p)}BA 3. In the experimental embodiment, the global reference frame for the robotic implementation was chosen as the base frame of the robot, denoted by frame R. Let C and P represent the frames attached to the camera and tip of the ultrasound probe. Since in the experimental embodiment both camera and probe are rigidly affixed to the robot's end effector, TRC and TRP are constant. TRC is estimated by performing an eye-in-hand calibration, whereas TRP is evaluated from the CAD model of the probe and its holder. Note that these transformations are composed of two transformations, namely:
In Equations (4) and (5), EE corresponds to the robot's end effector frame. The holder in the experimental embodiment was designed such that the frame of the ultrasound probe would be translated by a fixed distance along the z-direction of the manipulator's end effector frame. Thus in the physical workspace, the relationship used to map out the point cloud data (frame PC) to the robot's base frame can be expressed as, by way of non-limiting example:
The anonymous CT patient scans (frame CT) were used to generate the mesh model (frame M) of the torsos, and hence the transformation between the two is known, set as RCTM=I, and {right arrow over (p)}CTM[0, 0, 0]T. The transformation TPCM between the point cloud data and mesh model was estimated through the pointmatcher library using the iterative closest point approach. Initial target points were either defined in the CT scans, or on the mesh models, both of which correspond to the M frame. Since the ultrasound probe target position and orientation are defined in the ultrasound probe frame (P), the following transformation may be used:
By way of non-limiting example, the overall control algorithm can be described through three major motion strategies: (a) positioning of the ultrasound probe near the target point, (b) tapping motion along the ribcage in the vicinity of the target point to collect displacement data at 20 N force, and (c) definition of an updated scanning position to avoid ribs, followed by a pre-defined sweeping motion along the probe's z-axis.
The trajectory generation of the manipulator in the experimental embodiment was performed by solving for the joint angles θi, i∈[0,5] through inverse kinematics computation facilitated by Open Robotics Control Software (OROCOS), and the built-in arm controller in the robot driver. Since obstacle avoidance has not been explicitly integrated into the robot's motion generator, the experimental embodiment defined a manipulator home configuration, from which the system can reach various target locations without colliding with the patient and table. The home configuration was centered at the patient's torso at an elevation of ˜0.35 m from the body, with the +z-axis of the end effector corresponding to the −z-axis of the robot's base frame. The robot was driven to the home configuration before each target scan.
For the experimental embodiment, the force-displacement collection task began with the robot maintaining the probe's orientation fixed (as defined by the goal), and moving parallel to the torso at regular intervals of 3 mm, starting at 15 mm away from the goal point, and ending 15 mm past the goal point, resulting in a total of eleven readings. The robot thus moved along the end effector's+z-axis, registering the probe's position when a force reading is first recorded, and when a 20 N force is reached. The L2 norm of the difference of these positions was stored as a displacement data point. The two data points that represent the smallest displacements were assumed to be rib landmarks, representing the center of the corresponding rib. The ideal direction of the applied force would be normal to the centerline of the curved section of the probe, however it may not always be the case as some regions in the lungs might only be reachable with the resultant force pushing the probe on the side. In the case where the measured lateral forces contributed to the overall force by over 20%, the overall force was then considered in the computations. The center of mass of the probe holder is not in line with the assumed center, however, it is stiff enough to prevent bending.
Since the goal point is located between two ribs, it can hence be localized with respect to the center of the two adjacent ribs. The goal point was thus projected onto the shortest straight segment separating the center of the ribs, that is also closest to the goal point itself. Let the distance of the goal point from Rib 1 be dl. Because the ribs are fairly close to each other, a straight line connecting the two was assumed to avoid modeling the curvature of a torso. Once two points with the smallest displacement were identified from the force collection procedure, a line connecting the two was defined in the end effector's coordinate frame, and distance d1 was computed along that line from Rib 1 to define the position of the updated target point. Maintaining the same orientation, the robot was thus driven to the updated goal point, the end effector then moved along the probe's+z-axis until a 20 N force is reached, followed by a sweeping motion of ±30° around the probe's line of contact with the patient.
To evaluate the effectiveness of the scanning point detection algorithm of the experimental embodiment, its results were compared to an expert radiologist's proposed scanning points on the surface of n=3 patients using CT data in 3D Slicer.
The medical expert selected ten different targets within the lungs of each patient at various locations (amounting to a total of 30 data points) and proposed corresponding probe position and orientation on the CT scans that would allow them to image the selected targets through ultrasound. The medical expert only reviewed the CT slices along the main planes (sagittal, transverse and coronal).
The following metrics were used to compare the expert's selection to the algorithm's output: (a) bone obstruction, which is a qualitative metric that indicates whether the path of the ultrasound center beam to the goal point is obstructed by any skeletal structure, and (b) the quality of the ultrasound image, which was estimated using the overall weight structure described herein, whereby a smaller scan weight signals a better scan, with less air travels, scattering, reflection and refraction. For a fair initial comparison, the image search was restricted in the detection algorithm to the plane considered by the medical expert, i.e., the scanning point was evaluated across a single image that passes through the target point.
In this setting, the algorithm did not return solutions that were obstructed by bones, whereas five out of 30 of the medical expert's suggested scanning locations resulted in obstructed paths.
The quality of the scans have been compared on the remaining unobstructed 25 data points, and it was found that the algorithm returned paths with an overall 6.2% improvement in ultrasound image quality as compared to the expert's selection based on the returned sum of weights. However, when the algorithm was reset to search for optimal scanning locations across several tilted 2D images, the returned paths demonstrated a 14.3% improvement across the 25 data points, indicating that it can provide estimates superior to an expert's suggestion that was based on exclusively visual cues. The remaining five points have also been tested on the algorithm, and optimal scanning locations were successfully returned.
The average runtime for the detection of a single scanning position and orientation is 10.5±2.1 min, evaluated from the aforementioned 30 target points. The two most time consuming tasks are the generation of oblique planes from the CT scans, and the Gustafson-Kessel clustering used to delineate lung air. Since this is a pre-processing step, the rather large time consumption is not a concern.
A total of 570 volumes were prepared from thorax CT scans, 550 of which were used for training, and 20 for testing. Each of the 570 volumes came from a different patient. In the training set, the minimum ribcage height was 195 mm, and maximum height 477 mm, whereas in the testing set, the minimum ribcage height was 276 mm, and maximum height 518 mm. The percentile distribution of the training and testing ribcage heights is detailed in Table 1.
The network was trained for 150 epochs optimizing the L1 distance between the predicted coordinates and the Ground truth coordinates using the Adam optimizer. The training took place on an NVIDIA Titan Xp GPU using the PyTorch framework, and converged in 75 min. A mean Euclidean error of 14.8±7 mm was observed on the unseen testing set, with a 95th percentile of 28 mm. The overall inference time was on average 0.15 s.
A total of four experiment sets were devised to evaluate the experimental embodiment: (a) with prior CT scans without force feedback, (b) with prior CT scans with force feedback, (c) with ribcage landmark estimation without force feedback, and (d) with ribcage landmark estimation with force feedback. The overall performance of the robotic system was assessed in comparison to clinical requirements, which encompass three major elements: (a) prevention of acoustic shadowing effects whereby the infiltrates are blocked by the ribcage, (b) minimization of distance traveled by the ultrasound beam to reach targets, particularly through air, and (c) maintaining a contact force below 20 N between the patient's body and ultrasound probe.
Due to the technical limitations imposed by the spread of COVID-19 itself, the real-life implementation of the experimental embodiment was limited to n=1 phantom. Additional results are thus reported using Gazebo simulations. The same three patients described herein were used for the simulation. Since Gazebo is not integrated with an advanced physics engine for modeling tissue deformation on a human torso, the force sensing mechanism was replaced with a ROS node that compensated for the process of applying a force of 20 N and measuring the displacement of the probe through a tabular data lookup, obtained from the FEA simulations. In other words, when the ultrasound probe in the simulation approached the torso, instead of pushing through and measuring the displacement for a 20 N force (which is not implementable in Gazebo for such complex models), the end effector was fixed in place, and returned a displacement value which was obtained from prior FEA simulations on the corresponding torso model.
To replicate a realistic situation with uncertainties and inaccuracies, the torso models were placed in the simulated world at a pre-defined location, corrupted with noise in the x, y, and z directions, as well as roll, pitch, yaw angles. Errors were estimated based on reported camera accuracy, robot's rated precision, and variations between original torsos design and final model. The noises were sampled from Gaussian distributions with the pre-computed means, using a standard deviation of 1% of the mean. The numerical estimates on the errors are reported in Table 2. The exact location of the torsos was thus unknown to the robot. For each torso model, a total of eight points were defined for the robot to image, four on each side. Each lung was divided into four quadrants, and the eight target points correspond to the centroid of each quadrant (see
Two main evaluation metrics were considered: (a) the positional accuracy of the final ultrasound probe placement, which is the L2 norm of the difference between the target ultrasound position, and the actual final ultrasound position, and (b) a binary metric for imaging the goal infiltrate or region, whereby it is either visualized or obstructed. Each experiment set was repeated ten times with different sampled errors at every iteration. The results of all four experiment sets are reported in Table 3.
In Table 3, Case 1 represents using CT data without force feedback; Case 2 represents using CT data with force feedback; Case 3 represents using landmark prediction only without force feedback; and Case 4 represents using landmark prediction only with force feedback. All values in Table 3 are in mm.
It is noticeable that the force feedback component has decreased the overall error on the probe's placement by 49.3% using prior CT scans, and 52.2% using predicted ribcage landmarks. The major error decrease was observed along the z-axis, since the force feedback ensured that the probe was in contact with the patient. It also provided additional information on the ribs' placement near the target points, which were used to define the same target points positions relative to the ribs' location as well. For the ultrasound probe placement using only force-displacement feedback, the average error across all three models was 20.6±14.7 mm. Using the final probe's placement and orientation on the torso, this data was converted into CT coordinates to verify that the point of interest initially specified was imaged. In all of the cases, the sweeping motion allowed the robot to successfully visualize all points of interest. When using ribcage landmark estimation, the displacement error with force feedback for all three patients averaged at 19.8±16.9 mm. Similarly, the data was transformed into CT coordinates, showing that all target points were also successfully swept by the probe. The average time required for completing an eight-point POCUS scan on a single patient was found to be 3.3±0.3 min and 18.6±11.2 min using prior CT scans with and without force feedback, respectively. The average time for completing the same scans was found to be 3.8±0.2 min and 20.3±13.5 min using predicted ribcage landmarks with and without force feedback, respectively. The reported durations do not include the timing to perform camera registration to the robot base, as it is assumed be known a priori.
The same eight points derived from the centroids of the lung quadrants were used as target points for the physical phantom. Since the manufactured phantom does not contain lungs or visual landmarks, the methodology was evaluated qualitatively, categorizing images into three major groups: (a) completely obstructed by bones, whereby 50-100% of the field of view is uninterpretable, with the goal point shadowed, (b) partially obstructed by bones, whereby <$50% of the field of view is uninterpretable, with the goal point not shadowed, and (c) unobstructed, whereby <10% of the field of view is uninterpretable, and the goal point not shadowed. Since the scanning point algorithm focuses on imaging a target point in the center of the ultrasound window, this metric is also reported for completeness. The phantom was assessed by an expert radiologist, confirming that the polycarbonate from which the ribcage is made is clearly discernible from the rest of the gelatin tissues. It does, however, allow for the ultrasound beam to traverse it, meaning that the “shadow” resulting from the phantom's rib obstruction will not be as opaque as that generated by human bones. The robot manipulator was first driven to the specified goal point, and displacement profiles were collected in the vicinity of the target. The ribs' location was estimated from the force-displacement profile, and the final goal point was recomputed as a distance percentage offset from one of the ribs. Each experiment set was repeated three times, the results of which are reported in Table 4. The evaluation of the ultrasound images was performed by an expert radiologist.
In Table 4, Case 1 represents using CT data without force feedback; Case 2 represents using CT data with force feedback; Case 3 represents using landmark prediction only without force feedback; and Case 4 represents using landmark prediction only with force feedback.
Results show that the force feedback indeed assisted with the avoidance of bone structures for imaging purposes, whereby 100% of the ultrasound scans using prior CT data were interpretable i.e., with a visible center, and 87.5% of the scans were interpretable using landmark estimation. Results without using force feedback show that 75% of scans have a visible center region using prior CT scans, and 58.3% using predicted landmarks. The landmark estimation approach demonstrated worse results due to errors associated with the prediction. Select images for all four experiments are shown in
Thus, this disclosure presents an autonomous—e.g., operational without human involvement—robotic ultrasound solution, e.g., for diagnosing and monitoring COVID-19 patients' lungs by imaging specified lung infiltrates. In the prior art, a sonographer initially palpates the patient's torso to assess the location of the ribs for correctly positioning the ultrasound probe. By contrast, some embodiments remove the need for proximity of a human technician and instead act autonomously. For example, some embodiments autonomously locate a position of a patient's ribcage relative to an ultrasound probe. Some embodiments autonomously produce an electronic three-dimensional patient-specific representation of a ribcage of a patient. Some embodiments autonomously provide an electronic representation of a target location, e.g., in or on the organ within the ribcage of the patient. According to some embodiments, a target location, e.g., in or on the organ within the ribcage of the patient, may be specified by a human, which may be the only human involvement in obtaining an ultrasound image of the target location. Some embodiments autonomously determine position and orientation of an ultrasound probe to acquire an image of a target location, e.g., in or on an organ within a ribcage of a patient. Some embodiments autonomously direct, by a robot, an ultrasound probe on a patient to acquire an image of a target location, e.g., in or on an organ within a ribcage of a patient. In some embodiments, no human technician needs to be proximate to the patient, e.g., in the same room as the patient or inside a bioprotective container with the patient, when the ultrasound scan is performed.
Both an autonomous robotic system that targets the monitoring of COVID-19 induced pulmonary diseases in patients and testing thereof is disclosed herein. An algorithm that can estimate the appropriate, e.g., optimal, position and orientation of an ultrasound probe on a patient's body to image target points in lungs identified using prior patient CT scans may be used. The algorithm makes use of the CT scan to assess the location of ribs, which should be avoided in ultrasound scans. In the case where CT data is not available, a deep learning algorithm that can predict 3D landmark positions of a ribcage given a torso surface model that can be obtained using a depth camera may be used. These landmarks are subsequently used to define target points on the patient's body. The target points, whether from prior CT scans or from deep learning applied to a torso surface model, are relayed to a robotic system. An optional force-displacement profile collection methodology allows the system to subsequently correct the ultrasound probe positioning on the phantom to avoid rib obstruction. An experimental embodiment was successfully tested in a simulated environment, as well as on a custom-made patient-specific phantom. Results have suggested that the force feedback enabled the robot to avoid skeletal obstruction, thus improving imaging outcomes, and that landmark estimation of the ribcage is a viable alternative to prior CT data.
Though described herein primarily relative to a patient's lung, where the bony obstruction is the patient's ribcage, embodiments are not so limited. For example, embodiments may be used to acquire an ultrasound image of any location in or on any organ entirely or partially present within a patient's ribcage, e.g., a lung, a heart, a spleen, a liver, a pancreas, or a kidney. As another example, embodiments may be used to acquire an ultrasound image of any location in or on any organ entirely or partially present within a patient's pelvic and/or hip bones, e.g., a urinary bladder, a rectum, a sigmoid colon, a urethra, a uterus, a fallopian tube, an ovary, a seminal vesicle, or a prostate gland.
Certain embodiments can be performed using a computer program or set of programs executed by an electronic processor. The computer programs can exist in a variety of forms both active and inactive. For example, the computer programs can exist as software program(s) comprised of program instructions in source code, object code, executable code or other formats; firmware program(s), or hardware description language (HDL) files. Any of the above can be embodied on a transitory or non-transitory computer readable medium, which include storage devices and signals, in compressed or uncompressed form. Exemplary computer readable storage devices include conventional computer system RAM (random access memory), ROM (read-only memory), EPROM (erasable, programmable ROM), EEPROM (electrically erasable, programmable ROM), and magnetic or optical disks or tapes.
While the invention has been described with reference to the exemplary embodiments thereof, those skilled in the art will be able to make various modifications to the described embodiments without departing from the true spirit and scope. The terms and descriptions used herein are set forth by way of illustration only and are not meant as limitations. In particular, although the method has been described by examples, the steps of the method can be performed in a different order than illustrated or simultaneously. Those skilled in the art will recognize that these and other variations are possible within the spirit and scope as defined in the following claims and their equivalents.
This application is the national stage entry of International Patent Application No. PCT/US2023/014008, filed on Feb. 28, 2023, and published as WO 2023/167830 A1 on Sep. 7, 2023, which claims the benefit of U.S. Provisional Patent Application No. 63/315,115, filed on Mar. 1, 2022, which are hereby incorporated by reference herein in their entireties.
Filing Document | Filing Date | Country | Kind |
---|---|---|---|
PCT/US2023/014008 | 2/28/2023 | WO |
Number | Date | Country | |
---|---|---|---|
63315115 | Mar 2022 | US |