Techniques used to treat fractures and/or deformities of anatomical structures, such as bones, can include the use of external fixators, such as hexapods and other fixation frames, which are surgically mounted to anatomical structure segments on opposed sides of a fracture site. A pair of radiographic images is taken of the fixator and anatomical structure segments at the fracture site. Data from the images is then manipulated to construct a three-dimensional representation of the fixator and the anatomical structures segments that can be used in developing a treatment plan, which may for example comprise realigning the anatomical structure segments through adjustments to the fixator.
Existing techniques for controlling fixator manipulation may, however, involve a number of limitations that may introduce inefficiency, complication, and unreliability. For example, some conventional techniques may rely on a surgeon or other user to indicate locations of certain fixator elements, such as hinges, within images that are displayed in a graphical user interface of a computer. However, it may often be difficult for the user to identify and mark positions of the hinges and other fixator elements within the images. In particular, depending upon the location and orientation from which an image is captured, hinges and other fixator elements may not be identified easily, such as because they may wholly or partially overlap one another or may otherwise be obscured within the images. This may make it cumbersome for the user to identify the fixator elements, thereby increasing time required to identify the elements, increasing the probability of errors, and reducing the reliability of the calculations. This may reduce the reliability of the treatment plan, possibly resulting in improper alignment of anatomical structures segments during the healing process, compromised union between the anatomical structure segments, necessitating additional rounds of radiographic imaging to facilitate alignment corrections, or even necessitating additional surgical procedures.
Techniques for hinge detection for orthopedic fixation, for example for correction of a deformity of an anatomical structure, such as a bone, are described herein. In particular, in some examples, a fixation apparatus may be attached to first and second anatomical structure segments. Images, such as x-rays, of the fixation apparatus and the attached anatomical structure segments may then be captured from different orientations with respect to the fixation apparatus.
In some examples, various manipulations to the fixation apparatus for correction of the anatomical structure deformity may be determined based on positions and orientations of the anatomical structure segments in three-dimensional space. Also, in some examples, the positions and orientations of the anatomical structure segments in three-dimensional space may be determined based on the images. In particular, in some cases, the positions and orientations of the anatomical structure segments in three-dimensional space may be determined by having a surgeon or other user indicate locations of various fixator elements and anatomical structures within the images. However, as described above, it may often be difficult for the user to identify and mark positions of certain fixator elements, such as hinges, within the images. In particular, depending upon the location and orientation from which an image is captured, hinges and other fixator elements may be not be identified easily, such as because they may wholly or partially overlap one another or may otherwise be obscured within the images. This may make it cumbersome for the user to identify the fixator elements, thereby increasing time required to identify the elements, increasing the probability of errors, and reducing the reliability of the calculations.
To alleviate the above and other problems, an automated or semi-automated hinge detection process may be employed. Specifically, in some examples, first and second images may be displayed of the first and the second anatomical structure segments with the fixation device attached thereto. Indications may be received of first image hinge locations associated with the plurality of hinges in the first image. Projected second image hinge locations associated with the plurality of hinges in the second image may then be determined based at least in part on the indications of the first image hinge locations. Hinge candidates may be detected in the second image having shapes associated with the plurality of hinges. The hinge candidates may be detected by computer software using automated software-based image analysis techniques. For example, the hinges may have circular shapes, and the computer software may employ circle detection algorithms, such as a Hough transformation, to identify circular shapes in the second image as hinge candidates. Candidate second image hinge locations of the hinge candidates within the second image may then be identified.
Adjusted second image hinge locations associated with the plurality of hinges within the second image may then be calculated based at least in part on the projected second image hinge locations and the candidate second image hinge locations. In some examples, in order to calculate the adjusted second image hinge locations, the hinge candidates may be grouped into a set of hinge candidate groups, for example based on similarities of their locations and size characteristics (e.g., radius lengths). The set of hinge candidate groups may then be weighted based at least in part on a number of hinge candidates within each of the set of hinge candidate groups. A highest weighted subset of hinge candidate groups from the set of hinge candidate groups may then be selected. A plurality of average group locations may then be calculated, with each of the average group locations being associated with a respective hinge candidate group of the highest weighted subset of hinge candidate groups. A transformation matrix may then be constructed that describes a spatial relationship between the projected second image hinge locations and the plurality of average group locations. The transformation matrix may then be used to adjust the projected second image hinge locations to the adjusted second image hinge locations. The adjusted second image hinge locations may then be used to determine physical locations of the fixation device and the first and second anatomical structure segments in physical three-dimensional space. The physical locations of the fixation device and the first and second anatomical structure segments may then be used to determine manipulations to the fixation device for the correction of the deformity.
The patent or application file contains at least one drawing executed in color. Copies of this patent or patent application publication with color drawing(s) will be provided by the Office upon request and payment of the necessary fee.
The foregoing summary, as well as the following detailed description of the preferred embodiments of the application, will be better understood when read in conjunction with the appended drawings. For the purposes of illustrating the methods and/or techniques of orthopedic fixation with imagery analysis, there are shown in the drawings preferred embodiments. It should be understood, however, that the instant application is not limited to the precise arrangements and/or instrumentalities illustrated in the drawings, in which:
For convenience, the same or equivalent elements in the various embodiments illustrated in the drawings have been identified with the same reference numerals. Certain terminology is used in the following description for convenience only and is not limiting. The words “right”, “left”, “top” and “bottom” designate directions in the drawings to which reference is made. The words “inward”, “inwardly”, “outward”, and “outwardly” refer to directions toward and away from, respectively, the geometric center of the device and designated parts thereof. The terminology intended to be non-limiting includes the above-listed words, derivatives thereof and words of similar import.
Referring initially to
The fixator members can be connected to each other via adjustment members, the adjustment members configured to facilitate the spatial repositioning of the fixator members with respect to each other. For example, in the illustrated embodiment, the orthopedic fixator 100 comprises a pair of fixator members in the form of an upper fixator ring 106 and a lower fixator ring 108. The fixator rings 106, 108 can be constructed the same or differently. For instance, the fixator rings 106, 108 can have diameters that are the same or different. Similarly, the fixator rings 106, 108 can be constructed with varying cross sectional diameters, thicknesses, etc. It should be appreciated that the fixator members of the orthopedic fixator 100 are not limited to the illustrated upper and lower fixator rings 106, 108, and that the orthopedic fixator 100 can be alternatively constructed. For example, additional fixator rings can be provided and interconnected with the fixator ring 106 and/or 108. It should further be appreciated that the geometries of the fixator members are not limited to rings, and that at least one, such as all of the fixator members can be alternatively constructed using any other suitable geometry.
The first and second anatomical structure segments 102, 104 can be rigidly attached to the upper and lower fixator rings 106, 108, respectively, with attachment members that can be mounted to the fixator rings 106, 108. For example, in the illustrated embodiment, attachment members are provided in the form of attachment rods 110 and attachment wires 112.
The rods 110 and the wires 112 extend between proximal ends attached to mounting members 114 that are mounted to the fixator rings 106, 108, and opposed distal ends that are inserted into or otherwise secured to the anatomical structure segments 102, 104. The mounting members 114 can be removably mounted to the fixator rings 106, 108 at predefined points along the peripheries of the fixator rings 106, 108, for example by disposing them into threaded apertures defined by the fixator rings. With respect to each fixator ring 106, 108, the mounting members 114 can be mounted to the upper surface of the ring, the lower surface of the ring, or any combination thereof. It should be appreciated that the attachment members are not limited to the configuration of the illustrated embodiment. For example, any number of attachment members, such as the illustrated rods 110 and wires 112 and any others, can be used to secure the anatomical structure segments to respective fixator members as desired. It should further be appreciated that one or more of the attachment members, for instance the rods 110 and/or wires 112, can be alternatively configured to mount directly to the fixator rings 106, 108, without utilizing mounting members 114.
The upper and lower fixator rings 106, 108 can be connected to each other by at least one, such as a plurality of adjustment members. At least one, such as all, of the adjustment members can be configured to allow the spatial positioning of the fixator rings with respect to each other to be adjusted. For example, in the illustrated embodiment, the upper and lower fixator rings 106, 108 are connected to each other with a plurality of adjustment members provided in the form of adjustable length struts 116. It should be appreciated that the construction of the orthopedic fixator 100 is not limited to the six struts 116 of the illustrated embodiment, and that more or fewer struts can be used as desired.
Each of the adjustable length struts 116 can comprise opposed upper and lower strut arms 118, 120. Each of the upper and lower strut arms 118, 120 have proximal ends disposed in a coupling member, or sleeve 122, and opposed distal ends that are coupled to universal joints 124 mounted to the upper and lower fixator rings 106, 108, respectively. The universal joints of the illustrated embodiment are disposed in pairs spaced evenly around the peripheries of the upper and lower fixator rings 106, 108, but can be alternatively placed in any other locations on the fixator rings as desired.
The proximal ends of the upper and lower strut arms 118, 120 of each strut 116 can have threads defined thereon that are configured to be received by complementary threads defined in the sleeve 122, such that when the proximal ends of the upper and lower strut arms 118, 120 of a strut 116 are received in a respective sleeve 122, rotation of the sleeve 122 causes the upper and lower strut arms 118, 120 to translate within the sleeve 122, thus causing the strut 116 to be elongated or shortened, depending on the direction of rotation. Thus, the length of each strut 116 can be independently adjusted with respect to the remaining struts. It should be appreciated that the adjustment members are not limited to the length adjustable struts 116 of the illustrated embodiment, and that the adjustment members can be alternatively constructed as desired, for example using one or more alternative geometries, alternative length adjustment mechanisms, and the like.
The adjustable length struts 116 and the universal joints 124 by which they are mounted to the upper and lower fixator rings 106, 108, allow the orthopedic fixator 100 to function much like a Stewart platform, and more specifically like a distraction osteogenesis ring system, a hexapod, or a Taylor spatial frame. That is, by making length adjustments to the struts 116, the spatial positioning of the upper and lower fixator rings 106, 108, and thus the anatomical structure segments 102, 104 can be altered. For example, in the illustrated embodiment the first anatomical structure segment 102 is attached to the upper fixator ring 106 and the second anatomical structure segment 104 is attached to the lower fixator ring 108. It should be appreciated that attachment of the first and second anatomical structure segments 102, 104 to the upper and lower fixator rings 106, 108 is not limited to the illustrated embodiment (e.g., where the central longitudinal axes L1, L2 of the first and second anatomical structure segments 102, 104 are substantially perpendicular to the respective planes of the upper and lower fixator rings 106, 108), and that a surgeon has complete flexibility in aligning the first and second anatomical structure segments 102, 104 within the upper and lower fixator rings 106, 108 when configuring the orthopedic fixator 100.
By varying the length of one or more of the struts 116, the upper and lower fixator rings 106, 108, and thus the anatomical structure segments 102 and 104 can be repositioned with respect to each other such that their respective longitudinal axes L1, L2 are substantially aligned with each other, and such that their respective fractured ends 103, 105 abut each other, so as to promote union during the healing process. It should be appreciated that adjustment of the struts 116 is not limited to the length adjustments as described herein, and that the struts 116 can be differently adjusted as desired. It should further be appreciated that adjusting the positions of the fixator members is not limited to adjusting the lengths of the length adjustable struts 116, and that the positioning of the fixator members with respect to each other can be alternatively adjusted, for example in accordance the type and/or number of adjustment members connected to the fixation apparatus.
Repositioning of the fixator members of an orthopedic fixation apparatus, such as orthopedic fixator 100, can be used to correct displacements of angulation, translation, rotation, or any combination thereof, within bodily tissues. A fixation apparatus, such as orthopedic fixator 100, utilized with the techniques described herein, can correct a plurality of such displacement defects individually or simultaneously. However, it should be appreciated that the fixation apparatus is not limited to the illustrated orthopedic fixator 100, and that the fixation apparatus can be alternatively constructed as desired. For example, the fixation apparatus can include additional fixation members, can include fixation members having alternative geometries, can include more or fewer adjustment members, can include alternatively constructed adjustment members, or any combination thereof.
Referring now to
The images can be captured from any position and/or orientation with respect to each other and with respect to the fixator 100 and the anatomical structure segments 102, 104. In other words, there is no requirement that the captured images be orthogonal with respect to each other or aligned with anatomical axes of the patient, thereby providing a surgeon with near complete flexibility in positioning the imagers 130. Preferably, the images 126, 128 are captured from different directions, or orientations, such that the images do not overlap. For example, in the illustrated embodiment, the image planes of the pair of images 126, 128 are not perpendicular with respect to each other. In other words, the angle α between the image planes of the images 126, 128 is not equal to 90 degrees, such that the images 126, 128 are non-orthogonal with respect to each other. Preferably, at least two images are taken, although capturing additional images may increase the accuracy of the method.
The images 126, 128 can be captured using one or more imaging sources, or imagers, for instance the x-ray imagers 130 and/or corresponding image capturing devices 127, 129. The images 126, 128 can be x-ray images captured by a single repositionable x-ray imager 130, or can be captured by separately positioned imagers 130. Preferably, the position of the image capturing devices 127, 129 and/or the imagers 130 with respect to the space origin 135 of the three-dimensional space, described in more detail below, are known. The imagers 130 can be manually positioned and/or oriented under the control of a surgeon, automatically positioned, for instance by a software assisted imager, or any combination thereof. The fixator 100 may also have a respective fixator origin 145.
Referring now to
The remaining operations of the process of
Referring back to
The interface 500 also includes controls for entry of strut information. In particular, interface 500 includes six drop down menus 512 may each be used to indicate a size of a respective strut. Global strut size indicator 511 may also be used to globally select a size for all six struts. Length selectors 513 may be each be used to select a length of a respective strut. Length indicators 514 may be each be used to provide a visual representation of the lengths of the respective struts. It is noted that the length indicators 514 do not necessarily depict the actual exact length of each strut, but rather represent the comparative lengths of the struts with respect to one another.
Save and Update button 516 may be selected to save and update the configuration information values shown in interface 500. In some examples, selection of button 516 may cause interface 500 to display and/or update a graphical representation 520 of the fixation apparatus generated based, at least in part, on the entered configuration information. The graphical representation 520 may be displayed using one or more graphical user interfaces of a computing system. As shown, graphical representation 520 includes six struts that may be color-coded in multiple colors for easy identification. For example, in some cases, each of the struts (or at least two of the struts) may be shown in different colors with respect to one another. The struts in graphical representation 520 may have sizes, lengths, mounting points, and other features corresponding to entered configuration information. Graphical representation 520 also depicts the fixator rings, which may have diameters/lengths, ring types, and other features corresponding to entered configuration information. Graphical representation 520 may, for example, improve efficiency and reliability by providing the user with a visual confirmation of information entered into interface 500, for example to allow fast and easy identification of errors or other problems.
At operation 316, images of the fixation apparatus and the first and second anatomical structure segments attached thereto are displayed, for example using one or more graphical user interfaces of a computing system. The displayed images may include images that were captured at operation 312, such as using x-ray imaging, computer tomography, magnetic resonance imaging, ultrasound, infrared imaging, photography, fluoroscopy, visual spectrum imaging, or any combination thereof. Techniques for acquiring images of the fixation apparatus and the first and second anatomical structure segments are described in detail above and are not repeated here. As set forth above, the acquired and displayed images need not necessarily be orthogonal to one another. Referring now to
At operation 318, first image information is received, for example using one or more graphical user interfaces of a computing system. The first image information may include indications of one or more locations, within the images, of at least part of one or more elements of the fixation apparatus. For example, the first image information may include one or more indications of locations of struts, hinges, rings, and other fixator elements. In some examples, the first image information may also include information about locations, within the images, of marker elements, for example that are mounted to components of the fixation apparatus, such as struts, hinges, and rings. In some cases, the first image information may include points representing locations of hinges and/or lines or vectors representing locations of struts. In some examples, the first image information may be entered into a computing system by selecting or indicating one or more locations within the displayed images, for example using a mouse, keyboard, touchscreen or other user input devices. In particular, using one or more input devices, a user may select points or other locations in the images, draw lines, circles, and generate other graphical indications within the images. For example, in some cases, a user may generate a point or small circle at a particular location in an image to indicate a location (e.g., center point) of a hinge within the image. As another example, in some cases, a user may generate a line and/or vector within an image to indicate a location and/or length of a strut within the image.
For example, as shown in
In some examples, the first image information generated within images 601-A and 601-B may include color-coded graphical representations of the struts, for example to enable the graphical representations to be more clearly associated with their respective struts. For example, in
Referring back to
Referring again to
Referring now to
As shown in
As shown in
Referring again to
The imaging scene parameters can include, but are not limited to image pixel scale factors, image pixel aspect ratio, the image sensor skew factor, the image size, the focal length, the position and orientation of the imaging source, the position of the principle point (defined as the point in the plane of a respective image 126, 128 that is closest to the respective imager 130), positions and orientations of elements of the fixator 100, the position and orientation of a respective image receiver, and the position and orientation of the imaging source's lens.
In a preferred embodiment, at least some, such as all of the imaging scene parameters can be obtained by comparing the locations of representations of particular components, or fixator elements of the fixator 100 within the two-dimensional spaces of the images 126, 128, with the corresponding locations of those same fixator elements in actual, three-dimensional space. The fixator elements comprise components of the orthopedic fixator 100, and preferably are components that are easy to identify in the images 126, 128. Points, lines, conics, or the like, or any combination thereof can be used to describe the respective geometries of the fixator elements. For example, the representations of fixator elements used in the comparison could include center lines of one or more of the adjustable length struts 116, center points of the universal joints 124, center points of the mounting members 114, and the like.
The fixator elements can further include marker elements that are distinct from the above-described components of the fixator 100. The marker elements can be used in the comparison, as a supplement to or in lieu of using components of the fixator 100. The marker elements can be mounted to specific locations of components of the fixator 100 prior to imaging, can be imbedded within components of the fixator 100, or any combination thereof. The marker elements can be configured for enhanced viewability in the images 126, 128 when compared to the viewability of the other components of the fixator 100. For example, the marker elements may be constructed of a different material, such as a radio-opaque material, or may be constructed with geometries that readily distinguish them from other components of the fixator 100 in the images 126, 128. In an example embodiment, the marker elements can have designated geometries that correspond to their respective locations on the fixator 100.
Fixator elements can be identified for use in the comparison. For example, locations, within the images 126, 128 of fixator elements may be indicated using the first image information received at operation 318 and described in detail above. In some examples, the locations of the fixator elements in the two-dimensional space of the images 126, 128 may be determined with respect to local origins 125 defined in the imaging planes of the images 126, 128. The local origins 125 serve as a “zero points” for determining the locations of the fixator elements in the images 126, 128. The locations of the fixator elements can be defined by their respective x and y coordinates with respect to a respective local origin 125. The location of the local origin 125 within the respective image can be arbitrary so long it is in the plane of the image. Typically, the origin is located at the center of the image or at a corner of the image, such as the lower left hand corner. It should be appreciated that the locations of the local origins are not limited to illustrated local origins 125, and that the local origins 125 can be alternatively defined at any other locations.
In some examples, a respective transformation matrix P may then be computed for each of the images 126, 128. The transformation matrices can be utilized to map location coordinates of one or more respective fixator elements in actual three-dimensional space to corresponding location coordinates of the fixator element(s) in the two-dimensional space of the respective image 126, 128. It should be appreciated that the same fixator element(s) need not be used in the comparisons of both images 126, 128. For example, a fixator element used in constructing the transformation matrix associated with image 126 can be the same or different from the fixator element used in constructing the transformation matrix associated with image 128. It should further be appreciated that increasing the number of fixator elements used in computing the transformation matrices can increase the accuracy method. The following equation represents this operation:
The symbols x and y represent location coordinates, with respect to the local origin 125, of a fixator element point in the two-dimensional space of images 126, 128. The symbols X, Y and Z represent corresponding location coordinates, with respect to a space origin 135, of the fixator element point in actual three-dimensional space. In the illustrated embodiment, the point corresponding to the center of the plane defined by the upper surface of the upper fixator ring 106 has been designated as the space origin 135. The illustrated matrix P can be at least four elements wide and three elements tall. In a preferred embodiment, the elements of the matrix P can be computed by solving the following matrix equation:
The vector p can contain eleven elements representing values of the matrix P. The following equations present arrangements of the elements in the vector p and the matrix P:
In the preferred embodiment, the twelfth element p12 of the matrix P can be set to a numerical value of one. The matrices A and B can be assembled using the two-dimensional and three-dimensional information of the fixator elements. For every point representing a respective fixator element, two rows of matrices A and B can be constructed. The following equation presents the values of the two rows added to the matrices A and B for every point of a fixator element (e.g., a center point of a respective universal joint 124):
The symbols X, Y and Z represent location coordinate values of a fixator element point in actual three-dimensional space relative to the space origin 135, and the symbols x and y represent location coordinate values of the corresponding fixator element point in the two-dimensional space of the respective image 126, 128 relative to local origin 125.
For every line representing a respective fixator element, two rows of matrices A and B can be constructed. The following equation presents the values of the two rows added to the matrices A and B for every line of a fixator element (e.g., a center line of a respective adjustable length strut 116):
The symbols X, Y and Z represent location coordinate values of a point belonging to a line of a fixator element in actual three-dimensional space relative to the space origin 135. The symbols dX, dY and dZ represent gradient values of the line in actual three-dimensional space. The symbols a, b and c represent constants defining a line in the two-dimensional space of a respective image 126, 128. For example, a, b, and c can be computed using two points belonging to a line on a respective image 126, 128. In a preferred embodiment, the value of b is assumed to be 1, unless the line is a vertical line, in which case the value of b is zero. A correlation of constants a, b and c with the respective image coordinates x and y is presented in the following equation:
The equation (2) can be over constrained by using six or more fixator elements, for example the adjustable length struts 116. It should be appreciated that it is not necessary for all of the fixator elements to be visible in a single one of the images 126, 128 in order to obtain the matrix P. It should further be appreciated that if one or more of the above-described imaging scene parameters are known, the known parameters can be used to reduce the minimum number of the fixator elements required to constrain equation (2). For instance, such information could be obtained from modern imaging systems in DICOM image headers. Preferably, a singular value decomposition or least squares method can be used to solve equation (2) for values of the vector p.
In some examples, the transformation matrices may then be decomposed into imaging scene parameters. The following equation can be used to relate the matrix P to matrices E and I:
It should be appreciated that additional terms can be introduced when decomposing the matrix P. For example, the method presented by Tsai, described in “A Versatile Camera Calibration Technique for High-Accuracy 3D Machine Vision Metrology Using of-the-shelf TV Cameras and Lenses”, IEEE Journal of Robotics & Automation, RA-3, No. 4, 323-344, August 1987, which is incorporated herein by reference in its entirety, can be used to correct images 126, 128, for radial distortion.
Matrices E and I contain imaging scene parameters. The following equation represents a composition of the matrix I:
The symbols sx and sy represent values of image coordinate scale factors (e.g., pixel scale factors). The symbol f, representing the focal length, corresponds to the value of the shortest distance between a respective imaging source 130 and the plane of a corresponding image 126, 128. The symbols tx and ty represent the coordinates of the principle point relative to the local origin 125 of the respective image 126, 128. The following equation represents the composition of the matrix E:
The symbols ox, oy and oz represent values of the position of the fixator 100 in actual three-dimensional space. The symbols r1 to r9 describe the orientation of the fixator 100. These values can be assembled into a three-dimensional rotational matrix R represented by the following equation:
The methods of Trucco and Verri, as described in “Introductory Techniques of 3-D Computer Vision”, Prentice Hall, 1998, or the method of Hartley, as described in “Euclidian Reconstruction from Uncalibrated Views”, Applications of Invariance in Computer Vision, pages 237-256, Springer Verlag, Berlin Heidelberg, 1994, which are incorporated herein by reference in their entireties, can be used to obtain values of the matrices E and/or I. Utilizing the resulting values of matrices E and I, a complete three-dimensional imaging scene of the fixator 100 and the anatomical structure segments 102, 104 can be reconstructed.
For example,
In some examples, the images 126, 128 and the imaging scene parameters may then be used to obtain the positions and/or orientations of the anatomical structure segments 102, 104 in three-dimensional space. The position and/or orientation data obtained can be used to develop a treatment plan for a patient, for example to change the orientation and/or position of the fractured first and second anatomical structure segments 102, 104 in order to promote union between the anatomical structure segments 102, 104, as described in more detail below. It should be appreciated that the methods and techniques described herein are not limited to applications of repositioning broken anatomical structures, and that orthopedic fixation with imagery analysis can be used in any other type of fixation procedure as desired, for example lengthening of anatomical structures, correction of anatomical defects, and the like.
In some examples, anatomical structure elements comprising representations of particular portions (e.g., anatomical features) of the anatomical structure segments 102, 104, may then be identified and their locations within the images 126, 128 determined. For example, locations, within the images 126, 128 of the first and the second anatomical structure segments may be indicated using the second image information received at operation 320 and described in detail above. In some examples, the locations of the anatomical structure elements may be determined with respect to the respective local origins 125 of images 126, 128.
The anatomical structure elements can be used in the construction of the three-dimensional representation of the position and/or orientation of the anatomical structure segments 102, 104. Preferably, the anatomical structure elements are easy to identify in the images 126, 128. Points, lines, conics, or the like, or any combination thereof can be used to describe the respective geometries of the anatomical structure elements. For example, in the illustrated embodiment, points 134 and 136 representing the fractured ends 103, 105 of the anatomical structure segments 102, 104, respectively, are identified as anatomical structure elements in the images 126, 128.
The anatomical structure elements can further include marker elements that are implanted into the anatomical structure segments 102, 104 prior to imaging. The marker elements can be used as a supplement to or in lieu of the above-described anatomical structure elements identified in the images 126, 128. The marker elements can be configured for enhanced viewability in the images 126, 128 when compared to the viewability of anatomical features of the anatomical structure segments 102, 104. For example, the marker elements may be constructed of a radio-opaque material, or may be constructed with readily distinguishable geometries.
A three-dimensional representation 200 of the anatomical structure segments 102, 104 can be reconstructed. The three-dimensional representation can be constructed with or without a corresponding representation of the fixator 100. In the illustrated embodiment, pairs of ray-lines, such as ray lines 138, 140 and 142, 144 can be constructed for the anatomical structure element points 134, 136, respectively. Each ray line connects an anatomical structure element in one of the images 126, 128 with a respective imager 130. Each pair of ray lines can be analyzed for a common intersection point, such as points 146, 148. The common intersection points 146, 148 represent the respective positions of the anatomical structure element points 134, 136, in the three-dimensional representation of the anatomical structure segments 102, 104. Of course more than a pair of ray lines, such as a plurality, can be constructed, for example if more than two images were captured. If the ray lines of a particular set do not intersect, a point closest to all the ray lines in the set can be used as the common intersection point.
The positions and/or orientations of the anatomical structure segments 102, 104 can be quantified or measured using common intersection points, for instance points 146, 148. For example, lines representing center lines of the anatomical structure segments 102, 104 can be constructed and can be compared to the anatomical axes of the patient. Additionally, the distance between the fractured ends 103, 105 of the anatomical structure segments 102, 104 can be quantified. Using these or similar techniques, the positions and/or orientations of the anatomical structure segments 102, 104 can be determined. It is further noted that, in some examples, in addition to the positions and orientations of the first and second anatomical structure segments, the positions and orientation of rings (and/or other elements of the fixation apparatus) in three-dimensional space may also be determined, for example using any of the techniques described. For example, in some cases, locations of the rings within the images 126, 128 may be determined based on the first image information and/or other provided information. In some examples, these locations may then be used to determine the positions and orientations of the rings in three-dimensional space. Additionally, in some examples, configuration information for the fixation apparatus, such as ring diameters and strut length and mounting information, may also be used to determine positions and orientations of the rings in three-dimensional space.
Referring now to
At operation 326, the deformity parameters calculated at operation 424 are displayed, for example using one or more graphical user interfaces of a computing system. Referring now to
At operation 328, a graphical representation of the position and orientation of the first and the second anatomical structure segments is generated and displayed. The graphical representation of the position and orientation of the first and the second anatomical structure segments may be displayed using one or more graphical user interfaces of a computing system. For example, as shown in
At operation 330, one or more mounting parameters are calculated. The mounting parameters may include parameters relating to mounting of a reference ring of the fixator onto a respective anatomical structure segment. For example, in some cases, the mounting parameters may include an amount of offset (e.g., lateral, medial, anterior, and/or posterior) such as for a center of the reference ring with respect to a reference point, a degree of tilt (e.g., proximal and/or distal), an amount of axial offset, a master tab rotation, and others. In some examples, the mounting parameters may be calculated as part of the process determining the positions and orientations of the first and segment anatomical structure segments described above at operation 322, for example using the techniques described above with reference to operation 322. It is noted that, for the process of
At operation 432, the mounting parameters calculated at operation 430 are displayed, for example using one or more graphical user interfaces of a computing system. Referring now to
At operation 334, a graphical representation of the position and orientation of the reference ring and the respective anatomical structure segment to which it is mounted is generated and displayed. The graphical representation of the position and orientation of the reference ring and the respective anatomical structure segment may be displayed using one or more graphical user interfaces of a computing system. For example, as shown in
At operation 336, one or more treatment plan options are received, for example using one or more graphical user interfaces of a computing system. A treatment plan is a plan for manipulating the fixation apparatus, for example in order to correct the deformity of the first and the second anatomical structure segments. The treatment plan may include, for example, a plan for making gradual adjustments to the positions and orientations of the fixator rings with respect to each other, for example by changing the lengths of the struts of the fixation apparatus. Referring now to
In some examples, the software may allow the treatment plan to be split into multiple treatment phases. This may allow for greater control of the deformity correction, such as by allowing the surgeon to define starting and target poses for each treatment phase, to control the options for each treatment phase, and to control the type of movement in each treatment phase. For example, in some cases, a user may be allowed to create multiple treatment phases. Each of the multiple treatment phases may be defined by an assigned starting pose and an assigned target (i.e., ending) pose. The starting pose for the first (initial) treatment phase may be the initial anatomical structure deformity position from which the treatment begins on the first day of treatment. The target pose for the final treatment phase may be the desired positions of the anatomical structure segments at the conclusion of treatment. The starting pose for each subsequent treatment phase (after the initial treatment phase) may be same as the target pose of each preceding treatment phase. For example, the starting pose of a second treatment phase may be the same as the target pose of the first treatment phase, and so forth. The multiple phases (1 to N) may be combined in a list, and the plans of the individual N phases may be combined into the treatment plan.
The software may provide an interface that allows the user to select the quantity of desired treatment phases, the starting pose and the target pose for each treatment phase, and separate options for each treatment phase. For example, for each treatment phase, the software may allow the user to select a respective duration (e.g., a number of days), a rate of distraction, a quantity of degrees of adjustment per day, of a number of adjustments (e.g., strut movements) per day. The options for each treatment phase may be different from one another. For example, the duration, rate of distraction, quantity of degrees of adjustment per day and/or number of adjustments per day for the first treatment phase may be different from the duration, rate of distraction, quantity of degrees of adjustment per day and/or number of adjustments per day for the second treatment phase. In some examples, the input parameters for each treatment phase may include a starting pose of the distal fragment, a target pose of the distal fragment, and plan options for the phase (duration/distraction rate at given point/degrees per day, etc.). In some examples, the multiple treatment phases may allow for overcorrection of the deformity. For example, the multiple treatment phases may allow for compression, which may be calculated by using an overcorrection with a negative value for the axial distraction. The multiple treatment phases may also provide a simple and intuitive mechanism to allow axial movement to be performed in an initial phase of the treatment, and to allow additional axial lengthening/distraction. In some examples, the user may describe the treatment plan in clinical terms (e.g. residual deformity as overcorrection, number of phases, distraction first, etc.). A software layer may then interpret the clinical language and create corresponding treatment phases according to the phase definitions described above.
At operation 338, manipulations to the fixation apparatus for correction of the anatomical structure deformity (i.e., a treatment plan) are determined. The manipulations to the fixation apparatus may include adjustments to the struts of the fixation apparatus, such as adjustments to the sizes and/or lengths of the struts. In some examples, operation 338 may be performed based, at least in part, on the treatment plan options received at operation 336. For example, operation 338 may be performed based, at least in part, on specified start date, on instructions to perform axial movement first (e.g., in an initial part of the treatment, such as prior to rotational movement), a specified final distance between reference points, instructions to perform additional lengthening by a specified amount, instructions to generate an axial gap to ensure anatomical structure clearance, a specified duration (e.g., a number of days) of treatment, a specified rate of distraction, and/or instructions to perform two perform a specified quantity (e.g., one, two, etc.) of adjustments per day.
In some examples, the treatment plan may also be determined based, at least in part, on a determination of desired changes to the positions and/or orientations of the anatomical structure segments 102, 104, for instance how the anatomical structure segments 102, 104 can be repositioned with respect to each other in order to promote union between the anatomical structure segments 102, 104. For example, in some cases, it may be desirable to change the angulation of the second anatomical structure segment 104 such that the axes L1 and L2 are brought into alignment, and to change the position of the second anatomical structure segment such that the fractured ends 103, 105 of the anatomical structure segments 102, 104 abut each other. Once the desired changes to the positions and/or orientations of the anatomical structure segments 102, 104 have been determined, a treatment plan for effecting the position and/or orientation changes can be determined. In a preferred embodiment, the desired changes to the positions and/or orientations of the anatomical structure segments 102, 104 can be effected gradually, in a series of smaller changes. The positions and/or orientations of the anatomical structure segments 102, 104 can be changed by changing the positions and/or orientations of the upper and lower fixator rings 106, 108 with respect to each other, for instance by lengthening or shortening one or more of the length adjustable struts 116.
The required changes to the geometry of the fixator 100 (i.e., the position and/or orientation of the fixator 100) that can enable the desired changes to the positions and/or orientations of the anatomical structure segments 102, 104 can be computed using the matrix algebra described above. For example, the required repositioning and/or reorientation of the second anatomical structure segment 104 with respect to the first anatomical structure segment 102 can be translated to changes in the position and/or orientation of the lower fixator ring 108 with respect to the upper fixator ring 106.
At operation 340, indications of the determined manipulations to the fixation apparatus are provided to one or more users. For example, in some cases, indications of the determined manipulations to the fixation apparatus may be provided using one or more graphical user interfaces of a computing system, using a printed hard copy, using audio feedback, and/or using other techniques. In particular, referring now to
In the example of
In some examples, the struts of the fixation apparatus attached to the patient may be color-coded, for example using color-coded caps, marker, or other color-coded materials included within and/or attached to the struts. In some examples, the physical color-coding of the struts in the fixation apparatus attached to the patient may match the color-coding of struts used in the software. For example, the physical color-coding of the struts in the fixation apparatus may match the color-coding of struts that may be used to color-code the blocks 1132-A and 1132-B of chart 1130, graphical representation 520, and other color-coded representations of the struts displayed by the software. In some examples, this may make it easier for physicians and/or patients to confirm that, when they physically adjust a strut on the fixation apparatus, they are adjusting the correct strut by the correct amount.
At operation 342, one or more graphical representations of the position and orientation of the first and second anatomical structure segments and the rings of the fixation apparatus is generated and displayed. The graphical representation of the position and orientation of the first and the second anatomical structure segments and the rings of the fixation apparatus may be displayed using one or more graphical user interfaces of a computing system. For example, referring back to
At operation 344, the treatment plan may be implemented, that is the geometry of the fixation apparatus may be changed, for example based on the manipulations determined at operation 338, in order to change positions and orientations of the anatomical structure segments.
Hinge Detection for Orthopedic Fixation
As described above, a frame matching process may be employed to determine positions and orientations of anatomical structure segments in three-dimensional space, such as for generating a treatment plan for correction of an anatomical deformity. As also described above, in some examples, as part of the frame matching process, a surgeon or other user may identify locations of fixator elements (e.g., hinges, struts, etc.) within displayed images (e.g., x-rays) that show the fixator attached to the anatomical structure segments. Some examples of this process are described above with reference to operation 318 of
In some examples, to alleviate the above and other problems, an automated or semi-automated hinge detection technique may be employed. Some examples of these hinge detection technique will now be described with reference to
A first example of the display of the first and the second images at operation 1410 is shown in
It is noted that, in the examples of
At operation 1412, indications are received of first image hinge locations associated with the plurality of hinges in the first image, for example using the one or more graphical user interfaces of the computing system. For example, as described above with respect to
In some examples, after the user indicates locations of the hinges 1541 within the AP View image 1501-A, the software may use the indicated hinge locations to determine locations of the fixator rings 1511 and 1512 within the AP View image 1501-A. The software may then generate ring graphical representations 1531 and 1532, corresponding to the fixator rings 1511 and 1512, respectively, and display the ring graphical representations 1531 and 1532 at the determined locations of the fixator rings 1511 and 1512 within the AP View image 1501-A. Referring now to
At operation 1414, a graphical projection of the fixator is overlaid, for example using the one or more graphical user interfaces of the computing system, on the second image. For example, referring now to
The graphical projection 1600 of the fixator may be rotated relative to fixator elements in the first image, such as based at least in part on an angle (such as at the exact angle or at an approximation of the angle) of image planes of the first and the second images with respect to one another. As shown in
The graphical projection 1600 of the fixator may be rotated based at least in part on the angle between image planes of the images because that rotation may correspond to the expected position of the fixator in the second image. For example, if an image plane of the LAT View image 1501-B is at an angle of ninety degrees to an image plane of the AP View image 1501-A, then it may be expected that the locations of the fixator rings in the LAT View image 1501-B will be rotated ninety degrees relative to the locations of the fixator rings in the AP View image 1501-A. In this way, the overlaying of the graphical projection 1600 on the second image may assist the user in identifying locations of the plurality of fixator elements in the second image. In some examples, a user may provide a numerical value, such as a quantity of degrees (e.g., ninety degrees), that expressly indicates to the software the value of the angle between image planes of the images. In other examples, the value of the angle may be inferred by the software based on descriptions of the images (e.g., anteroposterior, anterior, posterior, lateral, medial, etc.) or using other techniques. In the examples of
Additionally, it is noted that the software may also manipulate other features of the graphical projection 1600 (e.g. size, location, orientations, etc.) such as to correct for other differences (e.g., location, orientation, zoom level, etc.) between the first and the second images. For example, in some cases, if the second image was captured from a closer location to the fixator and/or is more zoomed-in than the first image, then the software may correct for this by enlarging the size of the graphical projection 1600 relative to the size of the fixator elements in the first image. By contrast, in some cases, if the second image was captured from a further location from the fixator and/or is more zoomed-out than the first image, then the software may correct for this by reducing the size of the graphical projection 1600 relative to the size of the fixator elements in the first image.
Thus, in some examples, the graphical projection 1600 of the fixator may be generated based, at least in part, on locations of fixator elements in the first image. Additionally or alternatively, in some examples, the graphical projection 1600 of the fixator may be generated based, at least in part, on configuration information for the fixator that is provided to the software by the user, such as ring types (e.g., full ring, foot plate, etc.), ring sizes, strut lengths, indications of mounting points (e.g., ring holes), and other information. Various types of configuration information and techniques for providing such information to the software are described in detail above, such as with respect to
At operation 1416, the software may allow a user to manipulate (e.g., resize, rotate, move, etc.) the graphical projection and/or the second image. For example, the user may manipulate the graphical projection to make it more precisely align with the positions of the fixator elements in the second image. For example, the software may provide controls that allow resizing (making the graphical projection larger or smaller) or rotating of the graphical projection relative to its initial placement by the software when being overlaid upon the second image at operation 1414. For example, in some cases, it may be necessary to resize and/or rotate the graphical projection to correct for slight differences in the actual angle between the first and the second images relative to the expected angle (e.g., if the images are actually at an angle of ninety-two degrees rather than ninety degrees, etc.), to correct for differences in distance, position or orientation of the first and the second images relative to the objects included in the images, or for other reasons. In some examples, the software may provide various controls, such as buttons, that allow selections of operations such as move, resize and rotate, and the software may be configured to receive input from input devices, such as a mouse or keyboard, to accomplish those manipulations, for example via drag-and-drop, button clicks, keystrokes, etc.
In some examples, in addition or as an alternative to allowing a user to manipulate the graphical projection, the software may allow the user to manipulate the second image (e.g., LAT View image 1501-B) upon which the graphical projection is overlaid. For example, in some cases, the software may allow the user to resize, rotate and/or move the second image and/or elements shown within the second image, such as to assist in aligning the fixator elements shown in the second image with corresponding elements of the graphical projection. Referring now to
At operation 1418, projected second image hinge locations associated with the plurality of hinges in the second image are determined. In some examples, the software may determine the projected second image hinge locations based at least in part on the indications of the first image hinge locations received at operation 1412. The projected second hinge locations are the software's estimated locations of where the software expects the hinges to be located within the second image based on the user's indications of the hinge locations in the first image. For example, because the software knows the spatial relationship (e.g., angle) between the first and second images, the software can use the locations of the hinges in the first image to project/estimate where the locations of the hinges are expected to be in the second image. In some examples, the projected second image hinge locations may be expressed by the software via X and Y coordinate values within the second image.
In some examples, the software may determine the projected second image hinge locations 1841 by rotating the first image hinge locations in the first image, such as based at least in part on an angle (such as at the exact angle or at an approximation of the angle) of image planes of the first and the second images with respect to one another. As shown in
The projected second image hinge locations 1841 may be rotated based at least in part on the angle between image planes of the images because that rotation may correspond to the expected position of the fixator in the second image. For example, if an image plane of the LAT View image 1501-B is at an angle of ninety degrees to an image plane of the AP View image 1501-A, then it may be expected that the locations of the hinges 1541 in the LAT View image 1501-B will be rotated ninety degrees relative to the locations of the hinges 1541 in the AP View image 1501-A. In some examples, a user may provide a numerical value, such as a quantity of degrees (e.g., ninety degrees), that expressly indicates to the software the value of the angle between image planes of the images. In other examples, the value of the angle may be inferred by the software based on descriptions of the images (e.g., anteroposterior, anterior, posterior, lateral, medial, etc.) or using other techniques.
At operation 1420, hinge candidates are detected in the second image. The hinge candidates have shapes associated with the plurality of hinges. A hinge candidate is an area of the second image that has a shape that is associated with (e.g., that resembles) a hinge. For example, a hinge candidate may be an area of the second image that is defined by a same or similar visual feature (e.g., a same or similar shade of white, black or gray or another color) and that has a shape (e.g., a substantially circular shape) that matches or corresponds to a shape of one of the hinges. The hinge candidates may be detected by computer software using automated software-based image analysis techniques that are performed on the second image. For example, the hinges may have circular shapes, and the performing of hinge detection by the computer software may include employing circle detection algorithms, such as a Hough transformation, on the second image to identify circular shapes in the second image as hinge candidates. It is noted that the identification of a circular shape for purposes of hinge detection, as that term is used herein, is meant to encompass identifying of both exactly circular shapes and shapes that are substantially circular, such as a circle that may be partly obfuscated or that has an oval shape. In many cases, the number of hinge candidates that are detected in the second image may be greater than the number of actual fixator hinges. This may be because the, in addition to detecting the actual hinges, the software may detect a number of false positive hinge candidates, such as other circular shapes (e.g., wires, other objects, etc.) in the second image. In addition, in some examples, even a single hinge may sometimes be detected as multiple hinge candidates, such different circles that have similar or adjacent locations but that have different size characteristics (e.g., radius lengths).
In some examples, in order to improve the hinge candidate detection results, the software may use a priori knowledge to detect the hinge candidates. In some cases, the software may determine a range of expected size characteristics (e.g., radius lengths) for the plurality of hinges, and the software may limit the hinge candidates to circular shapes having determined size characteristics that are within the range of expected size characteristics. For example, the software may determine a range of expected radius lengths for a hinge, and the software may limit the hinge candidates to circular shapes having radii whose lengths are within the range of expected radius lengths. For example, a range of expected radius lengths may include a minimum expected radius length and a maximum expected radius length as well as all lengths between the minimum and maximum expected radius lengths. In some examples, the minimum expected radius length may be based on the smallest detected hinge radius in the first image (e.g., AP View image 1501-A). For example, the minimum expected radius length may be equal to the smallest detected hinge radius in the first image (e.g., AP View image 1501-A) minus a selected offset value. Also, in some examples, the maximum expected radius length may be based on the largest detected hinge radius in the first image (e.g., AP View image 1501-A). For example, the maximum expected radius length may be equal to the largest detected hinge radius in the first image (e.g., AP View image 1501-A) plus a selected offset value. In some cases, the radius lengths or other size characteristics of the hinges in the first image may be determined by the software by also performing an automated image analysis (e.g., using a Hough transformation) on the first image to detect the size characteristics of circles at the locations of the first image that were indicated by the user for the hinges in the first image. In some examples, because the second image and the first image may be captured from a same or similar distance to the fixator, the software may reasonably assume that the size characteristics (e.g., radius length) of the hinges should be the same or similar in the first and the second images. Thus, the size characteristics of the hinges in the first image may be used as a priori knowledge to more accurately identify hinge candidates, such as by excluding certain false positives, for example shapes or objects that have a size characteristic (e.g., radius length) that is too big or too small to be an actual hinge. It is noted that, in addition or as an alternative to radius lengths, other size characteristics (e.g., circumference, diameter, etc.) may be used to limit the detected hinge range of hinge candidates in a corresponding fashion as the radius length features described above.
Additionally, in some examples, a priori knowledge used to improve hinge detection results may include hinge orientation. For example, in some cases, the software may expect one or more rings of the fixator to be displayed at a certain angle within the image, such as substantially perpendicular to a bone segment, which may result in the ring being substantially horizontal in the second image. Moreover, the software may also expect hinges corresponding to a particular ring to be aligned with one another in a straight line. For example, the software may expect proximal hinges adjacent to a proximal ring to be aligned with one another in a straight line. The software may also expect distal hinges adjacent to the distal ring to be aligned with one another in a straight line. The software may also expect the line to have the same or similar angle as the respective ring. The software may use this a priori knowledge to more accurately identify hinge candidates, such as by excluding certain false positives. For example, in some cases, if the software identifies an outlying circular shape that is not aligned with any other detected circular shapes, then the software may consider this outlying circular shape to be a false positive and may not include it within the group of detected hinge candidates. Thus, in some examples, the detecting of the hinge candidates may be performed based at least in part on orientations of detected shapes within the second image.
At operation 1422, candidate second image hinge locations are identified. The candidate second image hinge locations are the locations of the hinge candidates in the second image. In some examples, the software may identify the candidate second image hinge locations by determining coordinate values (e.g., X and Y coordinate values) for each of the detected hinge candidates within the second image. Referring now to
At operation 1424, adjusted second image hinge locations associated with the plurality of hinges in the second image are calculated. The adjusted second image hinge locations may be calculated based, at least in part, on the projected second image hinge locations (determined at operation 1418) and the candidate second image hinge locations (determined at operation 1422). For example,
Referring now to
At sub-operation 1424B, the set of hinge candidate groups are weighted. In some examples, the set of hinge candidate groups may be weighted based at least in part on a number of hinge candidates within each of the set of hinge candidate groups. In some examples, hinge candidate groups with more included hinge candidates may be assigned a higher priority weight, while hinge candidate groups with fewer included hinge candidates may be assigned a lower priority weight. In some examples, the number of hinge candidates in each group may correspond to the exact weight assigned to the group. As shown in hinge candidate group list 2000, the weight assigned to each hinge candidate group may be the first (i.e. left-most) value shown in each row that indicates the number of hinge candidates included in the respective group. For example, the underlined row of hinge candidate group list 2000 indicates that the respective hinge candidate group includes four hinge candidates, and this group may therefore receive a weight of four. By contrast, the top row of hinge candidate group list 2000 indicates that the respective hinge candidate group includes five hinge candidates, and this group may therefore receive a weight of five.
At sub-operation 1424C, a highest weighted subset of hinge candidate groups may be selected from the set of hinge candidate groups. For example, if a subset of the four highest weighted hinge candidate groups were selected from the hinge candidate group list 2000, then this subset would include hinge candidate groups represented by the first/top four rows on the hinge candidate group list 2000 (e.g., with respective weights of five or four). As another example, if a subset of the seven highest weighted hinge candidate groups were selected from the hinge candidate group list 2000, then this subset would include hinge candidate groups represented by the first/top seven rows on the hinge candidate group list 2000 (e.g., with respective weights of five, four or three).
At sub-operation 1424D, the software calculates a plurality of average group locations for the subset of highest weighted hinge candidate groups. Each of the average group locations may be associated with a respective hinge candidate group of the highest weighted subset of hinge candidate groups. For example, the average group location for a group may include an average of the X coordinate values of all of the hinge candidates in the group (i.e., the second/center value shown in each row of hinge candidate group list 2000) and an average of the Y coordinate values of all of the hinge candidates in the group (i.e., the third/right-most value shown in each row of hinge candidate group list 2000). Referring now to
At sub-operation 1424E, a transformation matrix is constructed that describes a spatial relationship between the projected second image hinge locations (determined at operation 1418) and the plurality of average group locations (determined at sub-operation 1424D). Referring now to
At sub-operation 1424F, the transformation matrix is used to adjust the projected second image hinge locations to the adjusted second image hinge locations. This may include, for example, determining spatial relationships that correlate the average group locations to the projected second image hinge locations and then using those spatial relationships (e.g., by reversing the spatial relationships) to adjust (e.g., transform) the projected second image hinge locations. For example, referring now to
In some examples, the software may use the adjusted second image hinge locations 2400 to determine locations of the fixator rings 1511 and 1512 within the LAT View image 1501-B. The software may then generate ring graphical representations corresponding to the fixator rings 1511 and 1512 and display the ring graphical representations at the determined locations of the fixator rings 1511 and 1512 within the LAT View image 1501-B. Referring now to
Referring back to
It is noted that the above description of the hinge detection techniques includes examples in which a priori knowledge from the first image is used for various purposes with respect to the second image, such as to determine projected second image hinge locations in the second image and to assist in identifying hinge candidates in the second image. It is noted, however, that the techniques described herein do not necessarily require a priori knowledge from the first image in order to perform hinge detection in the second image (or vice versa). For example, in some cases, hinge candidates could be detected in an image, such as by performing automated software-based image analysis techniques. The image analysis techniques may include performing a Hough transformation to detect circular shapes within the image. Hinge locations within the image may then be determined based at least in part on the detected hinge candidates, in some examples without the use of any a priori knowledge from another image. In some examples, various techniques described above, such as the grouping, weighting, location averaging and/or other techniques, may also optionally be employed. For example, in some cases, the detected hinge candidates may be grouped, such as using the grouping techniques described above. In some examples, only a selected subset of the highest weighted hinge groups may be used. In some examples, average group locations may be calculated for the hinge groups, such as using the techniques described above. In some examples, these average group locations may be used as the determined hinge locations, or the determined hinge locations may otherwise be calculated based at least in part on these average group locations. The determined hinge locations may then be used to determine physical locations of the fixation device and the first and the second anatomical structure segments in three-dimensional space. The physical locations of the fixation device and the first and the second anatomical structure segments may then be used to determine manipulations to the fixation device for the correction of the deformity.
Example Computing Device
Referring to
In an example configuration, the computing device 78 includes a processing portion 80, a memory portion 82, an input/output portion 84, and a user interface (UI) portion 86. It is emphasized that the block diagram depiction of the computing device 78 is exemplary and not intended to imply a specific implementation and/or configuration. The processing portion 80, memory portion 82, input/output portion 84, and user interface portion 86 can be coupled together to allow communications therebetween. As should be appreciated, any of the above components may be distributed across one or more separate devices and/or locations.
In various embodiments, the input/output portion 84 includes a receiver of the computing device 78, a transmitter of the computing device 78, or a combination thereof. The input/output portion 84 is capable of receiving and/or providing information pertaining to communicate a network such as, for example, the Internet. As should be appreciated, transmit and receive functionality may also be provided by one or more devices external to the computing device 78.
The processing portion 80 may include one or more processors. Depending upon the exact configuration and type of processor, the memory portion 82 can be volatile (such as some types of RAM), non-volatile (such as ROM, flash memory, etc.), or a combination thereof. The computing device 78 can include additional storage (e.g., removable storage and/or non-removable storage) including, but not limited to, tape, flash memory, smart cards, CD-ROM, digital versatile disks (DVD) or other optical storage, magnetic cassettes, magnetic tape, magnetic disk storage or other magnetic storage devices, universal serial bus (USB) compatible memory, or any other medium which can be used to store information and which can be accessed by the computing device 78.
The computing device 78 also can contain the user interface portion 86 allowing a user to communicate with the computing device 78. The user interface 86 can include inputs that provide the ability to control the computing device 78, via, for example, buttons, soft keys, a mouse, voice actuated controls, a touch screen, movement of the computing device 78, visual cues (e.g., moving a hand in front of a camera on the computing device 78), or the like. The user interface portion 86 can provide outputs, including visual information (e.g., via a display), audio information (e.g., via speaker), mechanically (e.g., via a vibrating mechanism), or a combination thereof. In various configurations, the user interface portion 86 can include a display, one or more graphical user interfaces, a touch screen, a keyboard, a mouse, an accelerometer, a motion detector, a speaker, a microphone, a camera, a tilt sensor, or any combination thereof. Thus, a computing system including, for example, one or more computing devices 78 can include a processor, a display coupled to the processor, and a memory in communication with the processor, one or more graphical user interfaces, and various other components. The memory can have stored therein instructions that, upon execution by the processor, cause the computer system to perform operations, such as the operations described above. As used herein, the term computing system can refer to a system that includes one or more computing devices 78. For instance, the computing system can include one or more server computing devices that communicate with one or more client computing devices.
While example embodiments of devices for executing the disclosed techniques are described herein, the underlying concepts can be applied to any computing device, processor, or system capable of communicating and presenting information as described herein. The various techniques described herein can be implemented in connection with hardware or software or, where appropriate, with a combination of both. Thus, the methods and apparatuses described herein can be implemented, or certain aspects or portions thereof, can take the form of program code (i.e., instructions) embodied in tangible non-transitory storage media, such as floppy diskettes, CD-ROMs, hard drives, or any other machine-readable storage medium (computer-readable storage medium), wherein, when the program code is loaded into and executed by a machine, such as a computer, the machine becomes an apparatus for performing the techniques described herein. In the case of program code execution on programmable computers, the computing device will generally include a processor, a storage medium readable by the processor (including volatile and non-volatile memory and/or storage elements), at least one input device, and at least one output device, for instance a display. The display can be configured to display visual information. The program(s) can be implemented in assembly or machine language, if desired. The language can be a compiled or interpreted language, and combined with hardware implementations.
It should be appreciated that the orthopedic fixation with imagery analysis techniques described herein provide not only for the use of non-orthogonal images, but also allow the use of overlapping images, images captured using different imaging techniques, images captured in different settings, and the like, thereby presenting a surgeon with greater flexibility when compared with existing fixation and imagery techniques.
The techniques described herein also can be practiced via communications embodied in the form of program code that is transmitted over some transmission medium, such as over electrical wiring or cabling, through fiber optics, or via any other form of transmission. When implemented on a general-purpose processor, the program code combines with the processor to provide a unique apparatus that operates to invoke the functionality described herein. Additionally, any storage techniques used in connection with the techniques described herein can invariably be a combination of hardware and software.
While the techniques described herein can be implemented and have been described in connection with the various embodiments of the various figures, it is to be understood that other similar embodiments can be used or modifications and additions can be made to the described embodiments without deviating therefrom. For example, it should be appreciated that the steps disclosed above can be performed in the order set forth above, or in any other order as desired. Further, one skilled in the art will recognize that the techniques described in the present application may apply to any environment, whether wired or wireless, and may be applied to any number of such devices connected via a communications network and interacting across the network. Therefore, the techniques described herein should not be limited to any single embodiment, but rather should be construed in breadth and scope in accordance with the appended claims.
This application is a continuation of U.S. application Ser. No. 16/839,381 filed Apr. 3, 2020, the contents of which are hereby incorporated by reference in their entirety.
Number | Name | Date | Kind |
---|---|---|---|
2055024 | Bittner, Jr. | Sep 1936 | A |
2391537 | Anderson | Dec 1945 | A |
3977397 | Kalnberz et al. | Aug 1976 | A |
4081686 | Nieuweboer | Mar 1978 | A |
4450834 | Fischer | May 1984 | A |
4489111 | Woodrum | Dec 1984 | A |
4615338 | Ilizarov et al. | Oct 1986 | A |
4620533 | Mears | Nov 1986 | A |
4630203 | Szirtes | Dec 1986 | A |
4768524 | Hardy | Sep 1988 | A |
4784125 | Monticelli et al. | Nov 1988 | A |
4875165 | Fencil et al. | Oct 1989 | A |
4889111 | Ben-Dov | Dec 1989 | A |
4890631 | Hardy | Jan 1990 | A |
4930961 | Weis | Jun 1990 | A |
4964320 | Lee, Jr. | Oct 1990 | A |
4973331 | Pursley et al. | Nov 1990 | A |
5062844 | Jamison et al. | Nov 1991 | A |
5074866 | Sherman et al. | Dec 1991 | A |
5087258 | Schewior | Feb 1992 | A |
5095919 | Monticelli et al. | Mar 1992 | A |
5108393 | Ruffa | Apr 1992 | A |
5156605 | Pursley et al. | Oct 1992 | A |
5179525 | Griffis et al. | Jan 1993 | A |
5180380 | Pursley et al. | Jan 1993 | A |
5209750 | Stef | May 1993 | A |
5275598 | Cook | Jan 1994 | A |
5358504 | Paley et al. | Oct 1994 | A |
5437668 | Aronson et al. | Aug 1995 | A |
5443464 | Russell et al. | Aug 1995 | A |
5451225 | Ross et al. | Sep 1995 | A |
5458599 | Adobbati | Oct 1995 | A |
5540686 | Zippel et al. | Jul 1996 | A |
5601551 | Taylor et al. | Feb 1997 | A |
5630814 | Ross et al. | May 1997 | A |
5653707 | Taylor et al. | Aug 1997 | A |
5681309 | Ross et al. | Oct 1997 | A |
5702389 | Taylor et al. | Dec 1997 | A |
5728095 | Taylor et al. | Mar 1998 | A |
5746741 | Kraus et al. | May 1998 | A |
5766173 | Ross et al. | Jun 1998 | A |
5776132 | Blyakher | Jul 1998 | A |
5871018 | Delp et al. | Feb 1999 | A |
5885282 | Szabo | Mar 1999 | A |
5891143 | Taylor et al. | Apr 1999 | A |
5919192 | Shouts | Jul 1999 | A |
5951556 | Faccioli et al. | Sep 1999 | A |
5961515 | Taylor et al. | Oct 1999 | A |
5963612 | Navab | Oct 1999 | A |
5967777 | Klein et al. | Oct 1999 | A |
5968043 | Ross et al. | Oct 1999 | A |
5971984 | Taylor et al. | Oct 1999 | A |
5976142 | Chin | Nov 1999 | A |
6017341 | Windhagen et al. | Jan 2000 | A |
6021579 | Schimmels et al. | Feb 2000 | A |
6030386 | Taylor et al. | Feb 2000 | A |
6047080 | Chen et al. | Apr 2000 | A |
6129727 | Austin et al. | Oct 2000 | A |
6206566 | Schuetz | Mar 2001 | B1 |
6293947 | Buchbinder | Sep 2001 | B1 |
6320928 | Vaillant et al. | Nov 2001 | B1 |
6363169 | Ritter et al. | Mar 2002 | B1 |
6434278 | Hashimoto | Aug 2002 | B1 |
6501848 | Carroll et al. | Dec 2002 | B1 |
6510241 | Vaillant et al. | Jan 2003 | B1 |
6537275 | Venturini et al. | Mar 2003 | B2 |
6701174 | Krause et al. | Mar 2004 | B1 |
6711432 | Krause et al. | Mar 2004 | B1 |
6912293 | Korobkin | Jun 2005 | B1 |
7113623 | Chen et al. | Sep 2006 | B2 |
7187792 | Fu et al. | Mar 2007 | B2 |
7226449 | Venturini et al. | Jun 2007 | B2 |
7280687 | Ban et al. | Oct 2007 | B2 |
7306601 | McGrath et al. | Dec 2007 | B2 |
7388972 | Kitson | Jun 2008 | B2 |
7490085 | Walker et al. | Feb 2009 | B2 |
RE40914 | Taylor et al. | Sep 2009 | E |
7645279 | Haupt | Jan 2010 | B1 |
7657079 | Lake et al. | Feb 2010 | B2 |
7677078 | Sauer et al. | Mar 2010 | B2 |
7758582 | Ferrante et al. | Jul 2010 | B2 |
7828801 | Mirza et al. | Nov 2010 | B2 |
7837621 | Krause et al. | Nov 2010 | B2 |
7887537 | Ferrante et al. | Feb 2011 | B2 |
7955334 | Steiner et al. | Jun 2011 | B2 |
8029505 | Hearn et al. | Oct 2011 | B2 |
8057474 | Knuchel et al. | Nov 2011 | B2 |
8062293 | Steiner et al. | Nov 2011 | B2 |
8147491 | Lavi | Apr 2012 | B2 |
8157800 | Vvedensky et al. | Apr 2012 | B2 |
8202273 | Karidis | Jun 2012 | B2 |
8257353 | Wong | Sep 2012 | B2 |
8282652 | Mackenzi et al. | Oct 2012 | B2 |
8296094 | Harrison et al. | Oct 2012 | B2 |
8323282 | Taylor | Dec 2012 | B2 |
8333766 | Edelhauser et al. | Dec 2012 | B2 |
8377060 | Vasta et al. | Feb 2013 | B2 |
8419732 | Mullaney | Apr 2013 | B2 |
8425512 | Vasta et al. | Apr 2013 | B2 |
8430878 | Vasta et al. | Apr 2013 | B2 |
8439914 | Ross et al. | May 2013 | B2 |
8444644 | Ross et al. | May 2013 | B2 |
8454604 | Wong | Jun 2013 | B2 |
8469958 | Stevens | Jun 2013 | B2 |
8574232 | Ross et al. | Nov 2013 | B1 |
8654150 | Haskell | Feb 2014 | B2 |
8777946 | Lindahl et al. | Jul 2014 | B2 |
8834467 | Singh et al. | Sep 2014 | B2 |
8858555 | Crozet et al. | Oct 2014 | B2 |
8864763 | Murray et al. | Oct 2014 | B2 |
8906021 | Lehmann et al. | Dec 2014 | B1 |
8945128 | Singh et al. | Feb 2015 | B2 |
8951252 | Steiner et al. | Feb 2015 | B2 |
8952986 | Haskell | Feb 2015 | B2 |
9011438 | Steiner et al. | Apr 2015 | B2 |
9017339 | Edelhauser et al. | Apr 2015 | B2 |
9039706 | Murray et al. | May 2015 | B2 |
9044271 | Edelhauser et al. | Jun 2015 | B2 |
9066756 | Wong | Jun 2015 | B2 |
9078700 | Ross et al. | Jul 2015 | B2 |
9101398 | Singh et al. | Aug 2015 | B2 |
9155559 | Ross et al. | Oct 2015 | B2 |
9204937 | Edelhauser et al. | Dec 2015 | B2 |
9220533 | Singh et al. | Dec 2015 | B2 |
9642649 | Nikonovas | May 2017 | B2 |
9895167 | Edelhauser et al. | Feb 2018 | B2 |
9959683 | Mikheev et al. | May 2018 | B2 |
10932857 | Nikonovas | Mar 2021 | B2 |
20010018617 | Copf et al. | Aug 2001 | A1 |
20020010465 | Koo et al. | Jan 2002 | A1 |
20030106230 | Hennessey | Jun 2003 | A1 |
20030191466 | Austin et al. | Oct 2003 | A1 |
20040068187 | Krause et al. | Apr 2004 | A1 |
20040073211 | Austin et al. | Apr 2004 | A1 |
20040073212 | Kim | Apr 2004 | A1 |
20040082849 | Schweikard et al. | Apr 2004 | A1 |
20040111024 | Zheng et al. | Jun 2004 | A1 |
20040133199 | Coati et al. | Jul 2004 | A1 |
20040167518 | Estrada | Aug 2004 | A1 |
20040208279 | Xiao et al. | Oct 2004 | A1 |
20050149018 | Cooper et al. | Jul 2005 | A1 |
20050215997 | Austin et al. | Sep 2005 | A1 |
20050256389 | Koga et al. | Nov 2005 | A1 |
20060276786 | Brinker | Dec 2006 | A1 |
20070043354 | Koo et al. | Feb 2007 | A1 |
20070043429 | Hegel et al. | Feb 2007 | A1 |
20070161983 | Cresina et al. | Jul 2007 | A1 |
20070161984 | Cresina et al. | Jul 2007 | A1 |
20070238069 | Lovald et al. | Oct 2007 | A1 |
20080012850 | Keating, III | Jan 2008 | A1 |
20080114267 | Lloyd et al. | May 2008 | A1 |
20090036890 | Karidis | Feb 2009 | A1 |
20090036892 | Karidis et al. | Feb 2009 | A1 |
20090105621 | Boyd et al. | Apr 2009 | A1 |
20090143788 | Fang et al. | Jun 2009 | A1 |
20090161945 | Morgan-Mar et al. | Jun 2009 | A1 |
20090177198 | Theodoros et al. | Jul 2009 | A1 |
20090226055 | Dankowicz et al. | Sep 2009 | A1 |
20090275944 | Huebner et al. | Nov 2009 | A1 |
20090326532 | Schulze | Dec 2009 | A1 |
20090326560 | Lampropoulos et al. | Dec 2009 | A1 |
20100030219 | Lerner et al. | Feb 2010 | A1 |
20100039421 | Toyomura et al. | Feb 2010 | A1 |
20100087819 | Mullaney | Apr 2010 | A1 |
20100104150 | Saint et al. | Apr 2010 | A1 |
20100172567 | Prokoski | Jul 2010 | A1 |
20100179548 | Marin | Jul 2010 | A1 |
20100191239 | Sakkers et al. | Jul 2010 | A1 |
20100280516 | Taylor | Nov 2010 | A1 |
20100305568 | Ross et al. | Dec 2010 | A1 |
20110004199 | Ross et al. | Jan 2011 | A1 |
20110029093 | Bojarski et al. | Feb 2011 | A1 |
20110103676 | Mullaney | May 2011 | A1 |
20110131418 | Teng et al. | Jun 2011 | A1 |
20110313419 | Mullaney | Dec 2011 | A1 |
20120029280 | Kucklick | Feb 2012 | A1 |
20120078251 | Benenati et al. | Mar 2012 | A1 |
20120136355 | Wolfson | May 2012 | A1 |
20120232554 | Shaevitz et al. | Sep 2012 | A1 |
20120259343 | Clark et al. | Oct 2012 | A1 |
20120330312 | Burgherr et al. | Dec 2012 | A1 |
20130041288 | Taylor et al. | Feb 2013 | A1 |
20130060146 | Yang et al. | Mar 2013 | A1 |
20130131675 | Vasta et al. | May 2013 | A1 |
20130138017 | Jundt et al. | May 2013 | A1 |
20130211521 | Shenoy | Aug 2013 | A1 |
20130245625 | Vasta et al. | Sep 2013 | A1 |
20130296857 | Barnett et al. | Nov 2013 | A1 |
20140135764 | Ross et al. | May 2014 | A1 |
20140236152 | Walberg et al. | Aug 2014 | A1 |
20140257286 | Lindahl et al. | Sep 2014 | A1 |
20140278325 | Burgherr et al. | Sep 2014 | A1 |
20140303670 | Colloca | Oct 2014 | A1 |
20140379038 | Dogramadzi et al. | Dec 2014 | A1 |
20150080892 | Lehmann et al. | Mar 2015 | A1 |
20150088135 | Singh | Mar 2015 | A1 |
20150112339 | Lindahl et al. | Apr 2015 | A1 |
20150223842 | Murray et al. | Aug 2015 | A1 |
20150238227 | Singh et al. | Aug 2015 | A1 |
20150257788 | Jay et al. | Sep 2015 | A1 |
20150265313 | Wong | Sep 2015 | A1 |
20150272624 | Singh | Oct 2015 | A1 |
20150305776 | Ross et al. | Oct 2015 | A1 |
20150305777 | Singh et al. | Oct 2015 | A1 |
20150313641 | Ross et al. | Nov 2015 | A1 |
20160022314 | Bordeaux et al. | Jan 2016 | A1 |
20160045225 | Edelhauser et al. | Feb 2016 | A1 |
20160092651 | Austin et al. | Mar 2016 | A1 |
20160113681 | Singh | Apr 2016 | A1 |
20160125603 | Tanji | May 2016 | A1 |
20160183979 | Del Deo et al. | Jun 2016 | A1 |
20170181800 | Nikonovas | Jun 2017 | A1 |
20170224520 | Karasahin | Aug 2017 | A1 |
20170303966 | Edelhauser et al. | Oct 2017 | A1 |
20170348054 | Kumar et al. | Dec 2017 | A1 |
20170348057 | Kumar et al. | Dec 2017 | A1 |
20170354439 | Mannanal et al. | Dec 2017 | A1 |
20180055569 | Wahl | Mar 2018 | A1 |
20200305977 | Gutmann et al. | Oct 2020 | A1 |
Number | Date | Country |
---|---|---|
1494397 | May 2004 | CN |
101296664 | Oct 2008 | CN |
102883671 | Jan 2013 | CN |
103270513 | Aug 2013 | CN |
105852985 | Aug 2016 | CN |
1100048 | May 2001 | EP |
1690506 | Aug 2006 | EP |
2767252 | Aug 2014 | EP |
2576774 | Aug 1986 | FR |
2756025 | May 1998 | FR |
2001-523985 | Nov 2001 | JP |
2003-144454 | May 2003 | JP |
2003-530177 | Oct 2003 | JP |
2004-254899 | Sep 2004 | JP |
2006-507056 | Mar 2006 | JP |
2006-218298 | Aug 2006 | JP |
2009-505736 | Feb 2009 | JP |
2011-512883 | Apr 2011 | JP |
2013-526377 | Jun 2013 | JP |
20-0443058 | Jan 2009 | KR |
2159091 | Nov 2000 | RU |
2352283 | Apr 2009 | RU |
9812975 | Apr 1998 | WO |
9959100 | Nov 1999 | WO |
0115611 | Mar 2001 | WO |
0178015 | Oct 2001 | WO |
0330759 | Apr 2003 | WO |
2007024904 | Mar 2007 | WO |
2009102904 | Aug 2009 | WO |
2010002587 | Jan 2010 | WO |
2010104567 | Sep 2010 | WO |
2011026475 | Mar 2011 | WO |
2011060264 | May 2011 | WO |
2011060266 | May 2011 | WO |
2011146703 | Nov 2011 | WO |
2012021307 | Feb 2012 | WO |
2014186453 | Nov 2014 | WO |
2019040829 | Feb 2019 | WO |
2020023686 | Jan 2020 | WO |
Entry |
---|
U.S. Appl. No. 16/839,381, filed Apr. 3, 2020. |
Changjiang Yang et al: “Planar conic based camera calibration”, Proceedings / 15th International Conference on Pattern Recognition Barcelona, Spain, Sep. 3-7, [Proceedings of the International Conference on Pattern Recognition. (ICPR)], IEEE Computer Society, Los Alamitos, Calif. [U.A.], vol. 1, Sep. 3, 2000 (Sep. 3, 2000) pp. 555-558. |
Charlton, an Investigation into the Effect of Lateral Hillslope inputs on Floorplain Hydraulic Model Predictions, Diss.University of Bristol, Sep. 1995, 289 pages. |
Circle Hough Transform, Wikipedia, https://en.wikipedia.org/wiki/Circle_Hough_Transform, web-archive capture from Jan. 23, 2020, accessed on Mar. 4, 2021 from web.archive.org/web/20200123161407/https://en.wikipedia.org/wiki/Circle_Hough_Transform, 5 pages. |
Circle Hough Transform, Wikipedia, https://en.wikipedia.org/wiki/Circle_Hough_Transform, web-archive capture from Jul. 31, 2017, accessed on Nov. 24, 2020 from web.archive.org/web/20170731074826/https://en.wikipedia.org/wiki/Circle_Hough_Transform, 4 pages. |
Circle Hough Transform, Wikipedia, https://en.wikipedia.org/wiki/Circle_Hough_Transform, web-archive capture from Jul. 31, 2017, accessed on Jun. 16, 2021 from web.archive.org/web/20170731074826/https://en.wikipedia.org/wiki/Circle_Hough_Transform, 4 pages. |
Circle Hough Transform, Wikipedia, https://en.wikipedia.org/wiki/Circle_Hough_Transform; webpage accessed Apr. 3, 2020, 5 pages. |
Decision to Grant (Translation) dated Mar. 2016 in Russian patent application 2012147835, 6 pages. |
Durali M, Shameli E. Full order neural velocity and acceleration observer for a general 6-6 Stewart platform. InNetworking, Sensing and Control, 2004 IEEE International Conference on Mar. 2, 20041 (vol. 1, pp. 333-338). |
Garreau et al., “A Knowledge-Based Approach for 3-D Reconstruction and Labeling of Vascular Networks from Biplane Angiographic Projections”, IEEE Transactions on Medical Imaging, Jun. 1991, vol. 10, No. 2, 122-131. |
Hartley, Euclidian Reconstruction from Uncalibrated Views, Applications of Invariance in Computer Vision, pp. 237-256, Springer Verlag, Berlin Heidelberg, 1994. |
Iterative Closest Point, Wikipedia, https://en.wikipedia.org/wiki/iterative_closest_point, web-archive capture from Jan. 17, 2019, accessed on Nov. 24, 2020 from web.archive.org/web/20190117001205/https://en.wikipedia.org/wiki/iterative_closest_point, 3 pages. |
Iterative Closest Point, Wikipedia, https://en.wikipedia.org/wiki/iterative_closest_point, web-archive capture from Jan. 17, 2019, accessed on Mar. 4, 2021 from web.archive.org/web/20190117001205/https://en.wikipedia.org/wiki/iterative_closest_point, 3 pages. |
Iterative Closest Point, Wikipedia, https://en.wikipedia.org/wiki/iterative_closest_point, web-archive capture from Oct. 28, 2010, accessed on Oct. 26, 2020 from web.archive.org/web/20101028140305/https://en.wikipedia.org/wiki/iterative_closest_point, 3 pages. |
Iterative Closest Point, Wikipedia, https://en.wikipedia.org/wiki/iterative_closest_point, web-archive capture from Sep. 13, 2006, accessed on Oct. 26, 2020 from web.archive.org/web/20060913000000/http://en.wikipedia.org/wiki/iterative_closest_point, 1 page. |
Iterative Closest Point, Wikipedia, https://en.wikipedia.org/wiki/Iterative_closest_point, webpage accessed Apr. 3, 2020, 3 pages. |
Kelly, “How to calculate 3D coordinates with two cameras, a calibration object, a java program, and a lot of MS Excel macros”, Jun. 10, 2002, 9 pages. |
Larionova, et al: “Roentgen absorptiometry for analyzing bone tissue mineral density in orthopaedic-and-traumatological patients”, Genius Orthopedics, No. 3, 2009, pp. 98-102. |
Maiocchi et al., “Operative Principles of Ilizarov”, Chapter 2, 1991, 26 pages. |
Maxframe Software User's Manual, Data Feb. 8, 2017, Section 9.1 Edit Strut Mounting Points, pp. 88-89. |
Maxframe Software User's Manual, Dated Feb. 8, 2017, 138 pages. |
Maxframe Software Users Manual, Dated Feb. 8, 2017, Section 7.4.2.6 “Display Frame Configuration Model” and Section 7.4.3 View Options, pp. 72-73. |
Maxframe Surgical Technique Guide, DePuy Synthes, Jan. 27, 2017, 92 pages. |
Maxframe Surgical Technique Guide, Radiographic Markers, DePuy Synthes, Jan. 27, 2017, pp. 43-44. |
Maxframe User's Manual, Dated Feb. 8, 2017, Section 7.4-7.5, “Perform The Frame Matching”, pp. 63-73. |
Nikonovas, Arkadijus. Taylor Spatial Frame: Kinematics, Mechanical Properties and Automation. Diss. University of Bristol, May 2005, 230 pages. |
Ortho-SUV Frame—Art of Deformity Correction, Ortho-SUV Ltd, captured by https://web.archive.org from http://www.miito.org/download/ortho-suv-frame-eng.pdfon Jun. 13, 2010; 11 pages. |
Ortho-SUV Frame—Art of Deformity Correction, Ortho-SUV Ltd, Jun. 13, 2010, 11 pages. |
Orthofix, TL-HEX Software User's Guide: Software version 1.4, Nov. 2015, 60 pages. |
Paley et al., “Deformity Correction By The Ilizarov Technique”, Operative Orthopaedics, 1993, 883-948. |
Paley, “The principles of deformity correction by the Ilizarov technique: Technical aspects”, Techniques in Orthopaedics, 1989, vol. 4, Issue 1,15-29. |
Parikh PJ, Lam SS. A hybrid strategy to solve the forward kinematics problem in parallel manipulators. IEEE Transactions on Robotics. Feb. 2005; 21(1): 18-25. |
Point set registration, Wikipedia, https://en.wikipedia.org/wiki/Point_set_registration, web-archive capture from Jul. 21, 2017, accessed on Mar. 11, 2021 from web.archive.org/web/20170721011854/https://en.wikipedia.org/wiki/Point_set-registration, 12 pages. |
Point set registration, Wikipedia, https://en.wikipedia.org/wiki/Point_set_registration, web-archive capture from Oct. 16, 2019, accessed on Mar. 4, 2021 from web.archive.org/web/2019016232144/https://en.wikipedia.org/wiki/Point_set-registration. |
Point set registration, Wikipedia, https://en.wikipedia.org/wiki/Point_set_registration, webpage accessed Apr. 3, 2020, 11 pages. |
Ren L, Feng Z, Mills JK. A self-tuning iterative calculation approach for the forward kinematics of a Stewart-Gough platform. In Mechatronics and Automation, Proceedings of the 2006 IEEE International Conference on Jun. 25, 2006, 2018-2023. |
Russakoff et al., “Intensity-Based 2D-3D Spine Image Registration Incorporating a Single Fiducial Marker”, Academic Radiology, Jan. 2005, vol. 12, No. 1, 37-50. |
Simard et al., “The Ilizarov Procedure: Limb Lengthening and Its Implications”, Physical Therapy, Jan. 1992, vol. 72, No. 1, 25-35. |
Solomin et al., Deformity Correction and Fracture Treatment by software-based Ortho-SUV Frame User Manual Draft, Sep. 2011, 90 pages. |
Solomin et al., Deformity Correction and Fracture Treatment by software-based Ortho-SUV Frame, User Manual, For SUV-Software vp 1.0 and vr 1.0, Vreden Russian Research Institute of Traumatology and Orthopedics, (Ortho-SUV) Ltd., Saint Petersburg, 2013, 144 pages. |
Solomin et al., Deformity Correction and Fracture Treatment by software-based Ortho-SUV Frame, User Manual, For SUV-Software vp 2.1, Vreden Russian Research Institute of Traumatology and Orthopedics, (Ortho-SUV) Ltd., Saint Petersburg, 2016, 158 pages. |
Solomin, The Basic Principles of External Fixation Using The Ilizarov Device, 2005, 371 pages. |
Stoughton et al., “A Modified Stewart Platform Manipulator with Improved Dexterity”, IEEE Transactions on Robotics and Automation, Apr. 1993, vol. 9, No. 2, 166-173. |
Stryker, Hoffman LRF, Gradual Correction, Operative technique, 2016, 36 pages. |
Styker, Hoffmann LRF Hexapod, Operative technique, Jul. 2016, 44 pages. |
Trucco et al., “Introductory Techniques of 3-D Computer Vision”, 1998, pp. 178-194. |
Tsai, “A Versatile Camera Calibration Technique for High-Accuracy 3D Machine Vision Metrology Using of-the-shelf TV Cameras and Lenses”, IEEE Journal of Robotics & Automation, RA-3, No. 4, Aug. 1987, 323-344. |
U.S. Appl. No. 15/247,333, filed Aug. 25, 2016, Non-final Rejection dated Sep. 19, 2018, 21 pages. |
U.S. Appl. No. 13/111,180, filed May 19, 2011, entitled Orthopedic Fixation with Imagery Analysis, Final Rejection dated Feb. 14, 2014. |
U.S. Appl. No. 13/111,180, filed May 19, 2011, entitled Orthopedic Fixation with Imagery Analysis, Final Rejection dated Oct. 23, 2015. |
U.S. Appl. 13/111,180, filed May 19, 2011, entitled Orthopedic Fixation with Imagery Analysis, Non-final Rejection dated Apr. 9, 2015. |
U.S. Appl. No. 13/111,180, filed May 19, 2011, entitled Orthopedic Fixation with Imagery Analysis, Non-final Rejection dated Jun. 6, 2013. |
U.S. Appl. No. 13/111,180, filed May 19, 2011, entitled Orthopedic Fixation with Imagery Analysis, Notice of Allowability dated Nov. 17, 2016. |
U.S. Appl. No. 13/111,180, filed May 19, 2011, entitled Orthopedic Fixation with Imagery Analysis, Notice of Allowance dated Aug. 26, 2016. |
U.S. Appl. No. 13/111,180, filed May 19, 2011, entitled Orthopedic Fixation with Imagery Analysis, Notice of Allowance dated Dec. 16, 2016. |
U.S. Appl. No. 13/111,180, filed May 19, 2011, entitled Orthopedic Fixation with Imagery Analysis, Notice of Allowance dated Sep. 17, 2014. |
U.S. Appl. No. 15/461,969, filed Mar. 17, 2017, entitled Orthopedic Fixation with Imagery Analysis, Non-final Rejection dated Feb. 13, 2018. |
U.S. Appl. No. 15/247,333, Restriction Requirement dated Apr. 5, 2018, 6 pages. |
Viceconti et al., “A software simulation of tibial fracture reduction with external fixator”, Computer Methods and Programs in Biomedicine, 1993, 40, 89-94. |
Mur-Artal, et al., ORB-SLAM: A Versatile and Accurate Monocular SLAM System, IEEE Transactions on Robotics, vol. 31, No. 5, Oct. 2015, pp. 1147-1163. |
Zhang, A Flexible New Technique for Camera Calibration, Technical Report, MSR-TR-98-71, Microsoft Research, Microsoft Corporation, Mar. 25, 1999, 22 pages. |
Number | Date | Country | |
---|---|---|---|
20220237797 A1 | Jul 2022 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 16839381 | Apr 2020 | US |
Child | 17720371 | US |