Rotating marker for image guided surgery

Information

  • Patent Grant
  • 12186028
  • Patent Number
    12,186,028
  • Date Filed
    Tuesday, June 15, 2021
    4 years ago
  • Date Issued
    Tuesday, January 7, 2025
    12 months ago
Abstract
A marker (40) for image guided surgery, consisting of a base (94), having a base axis, connecting to a clamp (42), and an alignment target (44). The alignment target includes a target region (120) having an alignment pattern formed thereon, and a socket (124) connected to the target region and configured to fit rotatably to the base, whereby the alignment target is rotatable about the base axis. The alignment target also includes an optical indicator (162) for the socket indicating an angle of orientation of the alignment target about the base axis.
Description
FIELD OF THE INVENTION

The present invention relates generally to surgery, and specifically to surgery performed using augmented reality.


BACKGROUND OF THE INVENTION

In an augmented reality system used by a physician performing surgery, it is typically necessary to register a frame of reference of a patient with a frame of reference of the augmented reality system used by the physician. Methods for registration are known in the art.


U.S. Pat. No. 8,848,977 to Bammer et al., describes a method for optical pose detection. A self-encoded marker where each feature on the pattern is augmented with a 2-D barcode is provided.


U.S. Pat. No. 9,220,573 to Kendrick et al., describes a system for tracking a tracking device for use with a surgical navigation system. The system can include at least one tracking device having a plurality of faces, and the faces can be operable to generate a signal upon activation.


U.S. Pat. No. 9,378,558 to Kajiwara et al., describes a self-position/self-orientation calculation unit calculating self-position and/or self-orientation in a predetermined coordinate system based on a marker in acquired imaged image data when it is determined that the marker exists within a predetermined area.


U.S. Pat. No. 9,495,585 to Bicer et al., describes methods to find one to one mapping between fiducial markers on a tracked object and fiducial marker projections on an image plane captured by a camera in optical object tracking systems.


U.S. Pat. No. 9,943,374 to Merritt et al., describes an image guidance system for tracking a surgical instrument during a surgical procedure. The image guidance system includes a plurality of cameras adapted to be located external to a surgical area for capturing images of optically visible patterns.


U.S. Pat. No. 10,022,104 to Sell et al., describes a marker that includes a first marker component having a first hydrogen proton density and a first mass density; and a second marker component having a second hydrogen proton density different than the first hydrogen proton density.


U.S. Pat. No. 10,080,616 to Wilkinson et al., describes a system which generates a three-dimensional representation of a bone and reference markers, defines a coordinate system for the three-dimensional representation, and determines locations of the reference markers relative to the coordinate system.


U.S. Pat. No. 10,108,833 to Hong et al., describes a marker with a pattern formed thereon, and which includes an optical system. At least a part of the pattern that uniquely appears depending on a direction in which the pattern is viewed from outside of the marker, through the optical system, is visually identified from the outside of the marker.


U.S. Pat. No. 10,251,724 to McLachlin et al., describes a reference tie that may be secured around a portion of a spine during a surgical procedure and that may be tracked by a surgical navigation system.


U.S. Pat. No. 10,296,805 to Yang et al., describes a marker wherein at least one of a position and pose with respect to a capturing unit is estimated.


U.S. Pat. No. 10,420,626 to Tokuda et al., describes methods for automated detection and registration of medical images using fiducial markers and processing algorithms.


U.S. Pat. No. 10,463,434 to Siegler et al., describes tracking marker support structures that include one or more fiducial reference markers, where the tracking marker support structures are configured to be removably and securely attached to a skeletal region of a patient.


U.S. Pat. No. 10,504,231 to Fiala describes fiducial markers that are printed patterns detected by algorithms in imagery from image sensors for applications such as automated processes and augmented reality graphics.


U.S. Pat. No. 10,537,395 to Perez describes a kinematic connector assembly for kinematically coupling two objects. The kinematic connector assembly comprises a receiver defining a cavity and having a plurality of constraint surfaces accessible in the cavity.


U.S. Patent Application 2003/0210812 to Khamene et al., describes an apparatus for pose determination using single camera tracking in a workspace The apparatus includes a computer programmed for making the pose determination and a tracker camera coupled to the computer for providing a tracking image and for which calibration information is stored.


U.S. Patent Application 2011/0098553 to Robbins et al., describes automatic registration of a magnetic resonance (MR) image that is carried out in an image guidance system by placing MR visible markers at known positions relative to markers visible in a camera tracking system.


U.S. Patent Application 2013/0106833 to Fun describes an input device for providing three-dimensional, six-degrees-of-freedom data input to a computer. The device includes a tracker having tracking points. One array of tracking points defines a first axis. Another array defines a second axis or plane orthogonal to the first axis.


U.S. Patent Application 2015/0150641 to Daon et al., describes a three-dimensional position and orientation tracking system that comprises one or more pattern tags, each comprising a plurality of contrasting portions, and a tracker for obtaining image information about the pattern tags.


U.S. Patent Application 2016/0324583 to Kheradpir et al., describes a patient reference device that includes a housing having a back side and a front side, and at least three tracking markers attached to the front side of the housing. The housing extends around the at least three tracking markers and beyond a horizontal plane defined by tops of the at least three tracking markers.


U.S. Patent Application 20170239015 to Sela et al., describes an apparatus that is at least partially visible by both a three dimensional (3D) scanner system of a medical navigation system and a tracking system of the medical navigation system.


Documents incorporated by reference in the present patent application are to be considered an integral part of the application except that, to the extent that any terms are defined in these incorporated documents in a manner that conflicts with definitions made explicitly or implicitly in the present specification, only the definitions in the present specification should be considered.


SUMMARY OF THE INVENTION

An embodiment of the present invention provides a marker for image guided surgery, including:


a base, having a base axis, connecting to a clamp; and


an alignment target, including:

    • a target region having an alignment pattern formed thereon;
    • a socket connected to the target region and configured to fit rotatably to the base, whereby the alignment target is rotatable about the base axis; and
    • an optical indicator for the socket indicating an angle of orientation of the alignment target about the base axis.


In a disclosed embodiment the socket is configured to only fit to the base in a plurality of at least two discrete orientations about the base axis. Typically, the plurality of discrete configurations is distributed symmetrically about the base axis. The plurality may consist of four discrete orientations.


In a further disclosed embodiment the socket consists of a plurality of apertures equal to the plurality of discrete orientations, and the optical indicator is configured to be visible through one of the apertures indicative of one of the discrete orientations.


In a yet further disclosed embodiment the socket consists of a plurality of apertures equal to the plurality of discrete orientations, and the optical indicator is configured to be visible through apertures selected and arranged so as to provide an unambiguous identification of each of the discrete orientations.


In an alternative embodiment the socket is configured to fit to the base in a plurality of non-discrete orientations about the base axis. The socket may include an aperture, and the optical indicator may be congruent with the aperture, and a fraction of the optical indicator visible through the aperture may be indicative of one of the non-discrete orientations. The aperture may consist of a semicircular arc.


In a further alternative embodiment the socket is at a fixed distance from the target region, and the marker further includes:


an augmented reality system operative during surgery on a patient; and


a processor configured to:


track the alignment target during the surgery,


provide a patient tracking vector to the augmented reality system in response to the tracking of the alignment target,


calculate a change in the angle of orientation of the alignment target in response to changes in images of the optical indicator, and


add a change-of-orientation vector, based only on the fixed distance and the change in the angle of orientation, to the patient tracking vector so as to update the patient tracking vector.


An embodiment of the present invention also provides a method for enabling rotation of a marker during surgery without requiring re-registration, including:


connecting a base, having a base axis, to a clamp;


forming an alignment pattern on a target region of an alignment target;


connecting a socket to the target region, the socket being at a fixed distance from the target region and being configured to fit rotatably to the base, whereby the alignment target is rotatable about the base axis;


providing an optical indicator for the socket indicating an angle of orientation of the alignment target about the base axis;


operating an augmented reality system during the surgery on a patient;


tracking the alignment target during the surgery;


providing a patient tracking vector to the augmented reality system in response to the tracking of the alignment target;


calculating a change in the angle of orientation of the alignment target in response to changes in images of the optical indicator; and


adding a change-of-orientation vector, based only on the fixed distance and the change in the angle of orientation, to the patient tracking vector so as to update the patient tracking vector.


In accordance with aspects of the present disclosure, a marker for image guided surgery is disclosed, the marker includes a base, having a base axis, connecting to an anchoring device; and an alignment target, including: a target region having an alignment pattern formed thereon; a socket connected to the target region and configured to fit rotatably to the base, whereby the alignment target is rotatable about the base axis; and an optical indicator for the socket indicating an angle of orientation of the alignment target about the base axis.


In various embodiments of the marker, the socket is configured to only fit to the base in a plurality of at least two discrete orientations about the base axis.


In various embodiments of the marker, the plurality of discrete configurations is distributed symmetrically about the base axis.


In various embodiments of the marker, the plurality comprises four discrete orientations.


In various embodiments of the marker, the socket comprises a plurality of apertures equal to the plurality of discrete orientations, and the optical indicator is configured to be visible through one of the apertures indicative of one of the discrete orientations.


In various embodiments of the marker, the socket comprises a plurality of apertures equal to the plurality of discrete orientations, and the optical indicator is configured to be visible through apertures selected and arranged so as to provide an unambiguous identification of each of the discrete orientations.


In various embodiments of the marker, the socket is configured to fit to the base in a plurality of non-discrete orientations about the base axis.


In various embodiments of the marker, the socket comprises an aperture, and the optical indicator is congruent with the aperture, and a fraction of the optical indicator visible through the aperture is indicative of one of the non-discrete orientations.


In various embodiments of the marker, the aperture comprises a semicircular arc.


In accordance with aspects of the present disclosure, an augmented reality system or a navigation and/or tracking system operative during surgery on a patient is provided, the system includes the marker, wherein the socket is at a fixed distance from the target region; and a processor configured to or a storage medium storing machine-executable instructions configured to execute on a computing system the instructions, when executed, cause the computing system to: track the alignment target during the surgery, provide a patient tracking vector in response to the tracking of the alignment target, calculate a change in the angle of orientation of the alignment target in response to changes in images of the optical indicator, and add a change-of-orientation vector, based only on the fixed distance and the change in the angle of orientation, to the patient tracking vector so as to update the patient tracking vector.


In various embodiments of the marker, the marker further includes the anchoring device.


In various embodiments of the marker, the anchoring device is a clamp or a pin.


In accordance with aspects of the present disclosure, a marker for image guided surgery is provided, the marker includes: a base, having a base axis; an interface configured to be coupled to an anchoring device; and an alignment target including: a target region having an alignment pattern formed thereon; a socket connected to the target region and configured to fit rotatably to the base, whereby the alignment target is rotatable about the base axis; and an optical indicator for the socket indicating an angle of orientation of the alignment target about the base axis.


In various embodiments of the marker, the base comprises the interface.


In various embodiments of the marker, the interface is coupled to the anchoring device in a manner that aligns the base axis with a base axis of the anchoring device.


In accordance with aspects of the present disclosure, a method for enabling rotation of a marker during surgery without requiring re-registration is provided, the method includes: connecting a base, having a base axis, to an anchoring device; forming an alignment pattern on a target region of an alignment target; connecting a socket to the target region, the socket being at a fixed distance from the target region and being configured to fit rotatably to the base, whereby the alignment target is rotatable about the base axis; providing an optical indicator for the socket indicating an angle of orientation of the alignment target about the base axis; operating an augmented reality system during the surgery on a patient; tracking the alignment target during the surgery; providing a patient tracking vector to the augmented reality system in response to the tracking of the alignment target; calculating a change in the angle of orientation of the alignment target in response to changes in images of the optical indicator; and adding a change-of-orientation vector, based only on the fixed distance and the change in the angle of orientation, to the patient tracking vector so as to update the patient tracking vector.


In various embodiments of the method, the socket is configured to only fit to the base in a plurality of at least two discrete orientations about the base axis.


In various embodiments of the method, the method further includes the socket includes a plurality of apertures equal to the plurality of discrete orientations, and wherein the optical indicator is configured to be visible through one of the apertures indicative of one of the discrete orientations.


In various embodiments of the method, the socket includes a plurality of apertures equal to the plurality of discrete orientations, and wherein the optical indicator is configured to be visible through apertures selected and arranged so as to provide an unambiguous identification of each of the discrete orientations.


In various embodiments of the method, the socket is configured to fit to the base in a plurality of non-discrete orientations about the base axis.


In various embodiments of the method, the socket includes an aperture, and wherein the optical indicator is congruent with the aperture, and wherein a fraction of the optical indicator visible through the aperture is indicative of one of the non-discrete orientations.


In accordance with aspects of the present disclosure, a computer-implemented method for enabling rotation of a marker during surgery without requiring re-registration, wherein the marker is anchored to a patient and configured to fit rotatably to a base, and wherein the marker comprises an alignment pattern and an optical indicator, the optical indicator configured to indicate an angle of orientation of the marker about a base axis of the base and the optical indicator is being at a fixed distance from the alignment pattern, the method including: operating a navigation system during the surgery on the patient; tracking the alignment target during the surgery; providing a patient tracking vector to the navigation system in response to the tracking of the alignment target; calculating a change in the angle of orientation of the alignment target in response to changes in images of the optical indicator; and adding a change-of-orientation vector, based only on the fixed distance and the change in the angle of orientation, to the patient tracking vector so as to update the patient tracking vector.


The present disclosure will be more fully understood from the following detailed description of the embodiments thereof, taken together with the drawings, in which:





BRIEF DESCRIPTION OF THE DRAWINGS


FIG. 1, which is a schematic illustration of a medical procedure, according to an embodiment of the present invention;



FIG. 2 is a schematic diagram illustrating an exemplary augmented reality assembly used in the procedure, according to an embodiment of the present invention;



FIG. 3 schematically illustrates the situation after an exemplary patient marker has been attached to an exemplary clamp which is inserted and adjusted in a patient, according to an embodiment of the present invention;



FIG. 4A is a schematic perspective view of the marker, and FIGS. 4B-4E are schematic views of different orientations of the marker, according to an embodiment of the present invention;



FIG. 5 is a flowchart describing the use of the marker in the medical procedure, according to an embodiment of the present invention;



FIGS. 6A-6D are schematic views of different orientations of an alternative patient marker, according to an embodiment of the present invention;



FIGS. 7A-7E are schematic views of different orientations of another alternative patient marker, according to an embodiment of the present invention;



FIG. 8 is a flowchart describing the use of the marker of FIGS. 7A-7E, according to an embodiment of the present invention; and



FIG. 9 is a diagram explaining some of the steps of the flowchart of FIG. 8, according to an embodiment of the present invention.





DETAILED DESCRIPTION OF EMBODIMENTS
Overview

In an augmented reality system that is used during surgery or a medical procedure on a patient it is necessary to track the position or movement (or both) of the patient. The system typically comprises a head-mounted display worn by a professional performing the surgery, and the tracking performed e.g., via a tracking system, is required so as to maintain registration of images of the patient that are presented in the display with the professional's view of the patient.


To track the patient's position or movement relative to the display, an alignment target (may be also referred as “target” herein below) may be fixed to the patient and a processor (e.g., via a tracking system) may be configured to track or determine the relative location of the target. In embodiments of the invention the target is fixed to a base of a patient anchoring device such as a clamp or a pin that is clamped or inserted to a bone of the patient, so that the target, when attached to the patient anchoring device, e.g., a clamp base, acts as a patient marker. According to other embodiments of the invention, the patient marker comprises the alignment target coupled to a marker base, wherein the patient marker may be anchored to the patient via an anchoring device, such as a clamp (e.g., a base of the anchoring device). If the surgery is on the spine of the patient, the bone may comprise one or more spinous processes of the patient's vertebrae.


Typically, the tracked position or location of the target of the patient marker is “registered” with or represents the patient (e.g., utilized as a fiducial marker), and this “registration” or representation is used by the processor during the surgery. Tracking or determining of the target location allows maintaining the registration of images of the patient that are presented in the display with the professional's view of the patient and may further allow the display of other tracked elements on the images of the patient, such as tools used during the surgery. The display of the tools may facilitate their navigation by the professional performing the surgery.


However, during the procedure the alignment target may interfere with the surgery being performed, for example, by obstructing the professional's view and/or by restricting the professional's action. In this case, the alignment target may be re-oriented or re-positioned with respect to the anchoring device, e.g., a clamp, and/or the patient, e.g., by re-positioning the target and/or the clamp, to overcome the interference. The re-orientation or re-positioning, may be performed by detaching the target from the clamp, then re-attaching the target to the clamp, or by re-positioning of the clamp itself. The above may necessitate re-registration of the target with the clamp, re-positioning of the clamp with respect to the bone and/or even the performance of an additional incision.


Embodiments of the present invention allow for re-orientation of the target with respect to the patient anchoring device without the necessity of re-registration by the professional and/or the performance of any additional operations. The alignment target comprises a target region having an alignment pattern formed thereon. A socket comprised in the alignment target is fixedly connected at a known distance to the target region. The alignment target and/or the anchoring device base also comprises an optical indicator for the socket indicating an angle of orientation of the alignment target about the base axis. According to some aspects, the socket is configured to fit rotatably to the base of the anchoring device, e.g., a clamp, so that the alignment target is rotatable about a base axis defined by the clamp base. According to some aspects, the patient marker comprises an alignment target coupled to a marker base. The alignment target comprises the optical indicator and configured to rotate around an axis of the marker base.


During the procedure the processor operating the augmented reality system may track the alignment target so as to provide a patient tracking vector to the system, the vector maintaining the registration referred to above. The processor may then calculate a change in the angle of orientation of the alignment target in response to changes in images of the optical indicator. Based only on the change in the angle of orientation and the known target region-socket distance, the processor may calculate a change-of-orientation vector, and then add this vector to the patient tracking vector so as to update the patient tracking vector.


The updated patient tracking vector acts to automatically update the registration of the tracking of the alignment target, indicating the professional's view of the patient, with the patient image data, so that no re-registration and/or any additional operations are necessary.


The terms “position” and “location” may be hereby interchangeably used. Although the description refers to a surgery, the disclosed systems, devices, and methods may be applicable, mutatis mutandis, to any suitable medical procedure. Although the description refers to a spinal surgery or a spinal medical procedure, the disclosed systems, devices, and methods may be applicable, mutatis mutandis, to medical procedures performed with respect to structures or parts of the body other than the spine, including joint trauma related procedures, hip surgery and hip replacement or cranial procedures. Although the description refers to a clamp, the disclosed systems, devices, and methods may be applicable, mutatis mutandis, to any other patient or bone anchoring device, such as a pin inserted into a patient's bone, e.g., into the ilium or a cranial anchoring frame. Although the description refers to an augmented reality system, the disclosed systems, devices and methods may be applicable, mutatis mutandis, to other navigation and/or tracking and display systems, such as stationary tracking and/or display systems, which are not head-mounted, and the like.


System Description


In the following, all directional references (e.g., upper, lower, upward, downward, left, right, top, bottom, above, below, vertical, and horizontal) are only used for identification purposes to aid the reader's understanding of the present invention, and do not create limitations, particularly as to the position, orientation, or use of embodiments of the invention.


Reference is now made to FIG. 1, which is a schematic illustration of a medical procedure, according to an embodiment of the present invention. During the procedure, performed by a professional 22, the professional uses a surgical navigation system 20, which assists the professional in performance of the procedure. Surgical navigation system 20 comprises a processor 26, which operates elements of the system, and which communicates with an augmented reality assembly 24, worn by professional 22, that is incorporated in the system. While assembly 24 may be incorporated for wearing into a number of different retaining structures on professional 22, in the present description the retaining structure is assumed to be similar to a pair of spectacles. Those having ordinary skill in the augmented reality art will be aware of other possible structures, such as incorporation of the augmented reality assembly into a head-up display that is integrated into a helmet worn by the user of system 20, and all such structures are assumed to be comprised within the scope of the present invention.


In one embodiment processor 26 is assumed to be incorporated within a stand-alone computer, and the processor typically communicates with other elements of the system, including assembly 24, wirelessly, as is illustrated in FIG. 1. Alternatively or additionally, processor 26 may use optical and/or conducting cables for the communication. In further alternative embodiments processor 26 is integrated within assembly 24, or in the mounting of the assembly. Processor 26 is typically able to access a database 38, wherein are stored images, other visual elements and any other type of data, including computer code, used by system 20. Software enabling processor 26 to operate system 20 or assembly 24 (or both) may be downloaded to the processor or to database 38 in electronic form, over a network, for example. Alternatively or additionally, the software may be provided on non-transitory tangible media, such as optical, magnetic, or electronic storage media.


Assembly 24 comprises, inter alia, an image capturing device 72, also termed herein a camera 72, that has a field of view 74. Camera 72 may be configured to capture images in the visible spectrum, non-visible spectrum or both. Assembly 24 and functions of system 20, processor 26, and device 72 are described below. An assembly similar to augmented reality assembly 24, and its operation, are described in U.S. Pat. No. 9,928,629, to Benishti, et al., whose disclosure is incorporated herein by reference.


The medical procedure exemplified here is performed on a patient 30, and during an initial stage of the procedure professional 22 makes an incision 32 into the patient's back. The professional then inserts a spinous process clamp 42, into the incision, so that opposing jaws of the clamp are located on opposite sides of a spinous process. The professional adjusts clamp 42 to grip one or more spinous processes, selected by the professional, of the patient.


The professional attaches an alignment target 44 to a base 94 of the clamp, the target when attached to the base operating as a patient marker 40. Patient marker 40 thus comprises alignment target 44 coupled to base 94. As is described below, patient marker 40 is used by system 20 to determine the position and orientation of patient 30 during the medical procedure.



FIG. 2 is a schematic diagram illustrating assembly 24, according to an embodiment of the present invention. As stated above, assembly 24 is configured, by way of example, as a pair of spectacles 50 mounted on a frame 54.


At least one image capturing device 68 or 72 is attached to frame 54. Typically, devices 68 comprise cameras configured to capture images of scenes viewed by the professional's eyes, including images of marker 40 in the visible spectrum.


As stated above assembly 24 comprises camera 72, which is configured to capture images of elements of a scene, including marker 40, in front of assembly 24. The images are produced from radiation projected by a projector 73 that is in the spectrum detected by camera 72. Projector 73 is located in close proximity to camera 72, so that radiation from the projector, that has been retroreflected, is captured by camera 72. The camera typically has a bandpass filter configured to block other radiation, such as that projected by surgical lighting. Typically, camera 72 and projector 73 operate in a non-visible region of the spectrum, such as in the near infra-red spectrum. As is described below, at least some retroreflected radiation is typically received from marker 40, and processor 26 uses the image of the marker produced by camera 72 from the received radiation to track the marker, and thus the position and orientation of patient 30. By tracking the position and orientation of patient 30, the processor is able to present, to professional 22 in assembly 24, images of the patient (e.g., a Computerized Tomography scan) that are correctly registered with the physician's actual view of the patient.


According to some aspects, camera 72 may be mounted on frame 54 between the two lenses of spectacles 50. According to some aspects, projector 73 may include two or more projectors. According to another exemplary embodiment, assembly 24 comprises only one capturing device, e.g., camera 72, configured to operate in the non-visible region of the spectrum, such as in the near infra-red spectrum. Camera 72 may be then configured to capture images of scenes viewed by the professional's eyes, including images of marker 40 in the non-visible spectrum. According to some aspects, camera 72 and two projectors 73 may be mounted on frame 54 between the two lenses of spectacles 50. Each of projector 73 may be mounted on each side of camera 72.


Capturing devices 68 and/or 72 and projector 73 may be mounted or positioned on other positions or arranged in other configurations with respect to frame 54 than as illustrated and described with respect to FIGS. 1 and 2, and as may be practiced, implemented or envisioned by persons skilled in the art.



FIG. 3 schematically illustrates the situation after clamp 42 has been inserted and adjusted in patient 30, according to an embodiment of the present invention. Target 44 is then attached to base 94 of the clamp, forming marker 40. The figure illustrates that clamp 42 has been attached to grip a bone 21 of patient 30, specifically to grip a spinous process 25 of vertebrae of the spine of the patient. After attachment, alignment target 44 is external to the patient. As shown in FIG. 3, exemplary clamp 42 comprises teeth, protruding internally from jaws of the clamp, the teeth facilitating the clamp fixedly gripping the spinous processes. An exemplary configuration of marker 40 is described in more detail below with reference to FIGS. 4A-4E.



FIG. 4A is a schematic perspective view of marker 40, and FIGS. 4B-4E are schematic views of different orientations of the marker, according to an embodiment of the present invention. As stated above, marker 40 is formed by attaching alignment target 44 to base 94 of clamp 42. The clamp is described below.


Exemplary clamp 42 comprises a pair of jaws 90, 92 in a lower section of the clamp. The jaws are coupled to clamp base 94 in an upper section of the clamp, the base comprising a jaw adjustment mechanism 96. In the embodiment described herein, jaw 92 is fixed to base 94, and jaw 90 moves with respect to jaw 92, by being rotated about a hinge pin 98. Jaw adjustment mechanism 96 comprises an adjustment screw 100, which is coupled by a lever 102 to jaw 90 so that rotation of the screw causes jaw 90 to approach or retreat from jaw 92. Thus professional 22 is able to cause the jaws of clamp 42 to grip or release a bone, such as spinous process 25, by rotating screw 100. Hinge pin 98 defines a hinge axis 106 about which jaw 90 rotates, and each jaw 90, 92 is substantially parallel to the hinge axis.


For clarity, in the description herein, elements of marker 40 are assumed, by way of example, to be referenced to an xyz set of orthogonal axes, with origin at the center of hinge pin 98. The xyz set of axes is illustrated in FIG. 4A, wherein a y-axis is assumed to correspond to hinge axis 106, an x-axis is orthogonal to a plane including jaws 90 and 92, and a z-axis is orthogonal to the x- and y-axes.


Alignment target 44 comprises a target region 120 and a socket 124, the target region and the socket being fixedly connected together by a connecting rod 46. Alignment target 44, together with its components target region 120 and socket 124, are generally planar, herein termed xy planar since they are assumed to be in a plane parallel to the x- and y-axes. Embodiments of the present invention measure an angle of orientation of alignment target 44 to clamp 42, so that a line 126, constructed from a center 130 of the socket to a center 134 of the target region and extended therefrom, is herein assumed to indicate a direction of orientation of the alignment target 44.


Target region 120, by way of example, is approximately rectangular and comprises optical elements 138. Elements 138 are arranged in a three-dimensional (3D) pattern with no rotational axis of symmetry (other than a trivial axis of symmetry for rotating by 360°), and no mirror plane of symmetry, so that an image of the elements enables an unambiguous determination of the location and orientation of the target region. Elements 138 are typically retroreflectors. According to some aspects, elements 138 are arranges in a two-dimensional (2D) pattern. An entity with an arrangement of optical elements similar to the arrangement herein is described in PCT Patent Application WO2019211741A1, which is incorporated herein by reference.


As stated above, socket 124 is generally planar, and is assumed to define an axis 152 through socket center 130 and orthogonal to an xy plane. In cases where socket center 130 lies on the z-axis, axis 152 is coincident with the z-axis, as is illustrated in FIG. 4A. Socket 124 comprises four substantially similar apertures 150 which are distributed symmetrically about axis 152. Socket 124 comprises a central hole 156, and a screw 154 is configured to penetrate the hole and connect the socket to an upper surface 170 of clamp base 94, as is illustrated in the call-out of FIG. 4C. Once connected, a base axis, comprising an axis orthogonal to surface 170 through central hole 156, corresponds to axis 152.


Surface 170 is xy planar, being parallel to an xy plane, and comprises four protrusions 174 distributed symmetrically about the z-axis. There is an aperture 178 in the surface providing access to adjustment screw 100, and the positions of apertures 150 are selected so that regardless of the orientation of target 44, access to screw 100 is available through one of apertures 150. There is also a shaped indentation 182, comprising an arc 186, in surface 170, the indentation being shaped to accept a colored or retroreflective insert 160.


As is also illustrated in FIG. 4C, socket 124 comprises a planar lower surface 190, within which are inset four indents 194, distributed symmetrically about socket central hole 156, and configured to mate with protrusions 174. Surface 190 is surrounded by a circular wall 196. Extending from surface 190 are a plurality of arcs 198, also distributed symmetrically about socket central hole 156, and configured to mate with arc 186 of indentation 182.



FIGS. 4B-4E illustrate the four different discrete orientations that exemplary alignment target 44 is able to make with clamp 42, when the target is connected to the clamp with screw 154 so that socket 124 mates with base 94. Assuming that the positive y-axis of the clamp corresponds to an orientation of 0°, and those orientations are measured as clockwise rotations about the z-axis from the y-axis, FIGS. 4B, 4C, 4D, and 4E correspond respectively to the target having discrete orientations of 0°, 90°, 180°, and 270°. At each orientation arcs 198 mate with arc 186, protrusions 174 mate with indents 194, and wall 196 mates with outer circular edges of base 94, the mating ensuring that socket 124 is centered with respect to the z-axis.


As is illustrated in FIGS. 4B-4E, at each orientation insert 160 is visible through one of apertures 150, and the visible insert acts as an optical indicator 162 of the orientation. During operation of system 20, processor 26 calculates coordinates of a directed line segment between indicator 162 and center 134, the coordinates acting as an orientation metric. For each orientation there is a unique directed line segment, i.e., a unique orientation metric, so that processor 26 is able to use the calculated coordinates as an orientation indicator.


Table I below shows coordinates, the orientation metric, of the directed line segment for each of the exemplary four orientations of target 44. (For clarity, Table I, and Table II below, are drawn to a two-dimensional system, and may be adapted, mutatis mutandis, to a three-dimensional or higher system.) The coordinates are calculated assuming that indicator 162 lies on a circle radius r centered at center 130 of socket 120, and that there is a separation of D between center 130 and center 134.












TABLE I








Orientation Indicator



Orientation
Orientation Metric



















 0°
(−r, 0) − (0, D) = (−r, −D)



 90°
(−r, 0) − (D, 0) = (−r − D, 0)



180°
(−r, 0) − (0, −D) = (−r, D)



270°
(−r, 0) − (−D, 0) = (D − r, 0)










As is described herein, marker 40 is used to track the location of patient 30, typically the patient's bone, with respect to assembly 24, by tracking the location of target region 120. Since the location of the target region is fixed with respect to the patient's bone to which marker 40 is clamped, because the marker is inflexible, tracking of the patient's bone may be accomplished by tracking of the target region, and adding a fixed adjustment vector due to the differing physical positions of the target region and the bone.


Furthermore, since the target region positions have a one-to-one correlation with the orientations, and since the different target region positions are in known geometric relations to each other, these geometric relations may be pre-programmed as change-of-orientation vectors and used to continue tracking the patient when the target region orientation is changed.


For example, if target region 120 is a distance D from socket 124, then the target region positions for the orientations 0°, 90°, 180°, and 270° (illustrated in FIGS. 4B, 4C, 4D, and 4E) may be respectively represented by the two-dimensional ordered pairs (0, D), (D, 0), (0, −D), and (−D, 0). If an initial target region is in the 0° orientation, then the geometric relations, i.e., the change-of-orientation vectors, to the other three orientations are as given in Table II:










TABLE II





New
Change-of-orientation


Orientation
Vector (From 0° Orientation)







 90°
(D, 0) − (0, D) = (D, −D)


180°
(0, −D) − (0, D) = (0, −2D)


270°
(−D, 0) − (0, D) = (−D, −D)









It will be understood that the three change-of-orientation vectors presented in Table II do not vary with movement of marker 40. The vectors depend only on the initial and final orientation of the target region, and so, as stated above, may be pre-programmed. It will also be understood that sets of three change-of-orientation vectors from the other possible initial orientations (90°, 180°, and 270°) may be calculated as for Table II, and may also be pre-programmed.


As is explained further below, in embodiments of the present invention, when the target region orientation changes a processor adds an appropriate change-of-orientation vector to an initial tracking vector of a patient marker. This enables continuous tracking of a patient by the marker without re-registration of the marker.



FIG. 5 is a flowchart describing the use of a marker such as marker 40 in the medical procedure referred to above, according to an embodiment of the present invention.


In an initial step 200, professional 22 attaches clamp 42 or any other patient anchoring device to a bone of patient 30, herein assumed to comprise spinous process 25 of the patient, by rotating screw 100.


In a target attachment step 204, the professional attaches an alignment target to the patient anchoring device in a specific orientation. The alignment target includes an optical indicator indicating the marker current orientation. For example, alignment target 44 may be attached to the clamp, by aligning socket 124 with surface 170 of clamp base 94, and screwing screw 154. It will be understood that the attached orientation is one of the four orientations illustrated in FIGS. 4B-4E, so that after the attachment, insert 160 is visible through one of apertures 150, and acts as optical indicator 162.


The values of r and D, and the sets of orientation metrics for the orientations of the target, described above with reference to Table I, may be input to or accessed by processor 26. In addition, the sets of change-of-orientation vectors coordinates, described above with reference to Table II, may also be input to or accessed by the processor.


In an image acquisition step 208, an image of optical elements and of the optical indicator of the alignment marker is captured via the head mounted assembly. With reference to FIGS. 1 and 2, an image of optical elements 138 of target region 120 and of indicator 162, is captured, e.g., using camera 72 and/or one or more devices 68. Optionally the image may include the bone of patient 30, or of a region close to the bone, such as incision 32 and/or a tool navigated in a region close to the bone or alignment target 44. The tool may be mounted with or include a tool marker. The tool location may be then identified with respect to the patient via the tool marker and the alignment target (the patient marker).


In an analysis step 212, the processor analyzes the acquired image to identify the alignment target and/or to find a location, comprising an orientation and a position, of the alignment target.


For example, the processor analyzes the acquired image to identify and/or find a location, comprising an orientation and a position, of target region 120. The position of target region is herein, by way of example, assumed to correspond to center 134. Once the target region location has been found, the processor may initiate tracking of the target region, for example, by capturing images of the target region, e.g., once in a predefined time interval.


The processor also identifies the optical indicator of the alignment target in the acquired image. For example, processor 26 may find a position of indicator 162 in the image. From coordinates of the two positions the processor calculates coordinates of an orientation metric (as in Table I) joining the positions, and from the metrics stored in step 204, identifies the orientation of step 204.


In an adjustment step 216, the processor utilizes the image, e.g., of the bone, of the region close to the bone, and/or of a tool navigated in the region close to the bone, acquired in step 208, to generate an augmented reality display to professional 22 in augmented reality assembly 24. The professional adjusts coordinates of the presented image to correspond with the actual image visible to the professional through the assembly, and the processor stores the adjustment coordinates. The processor then applies the stored adjustment coordinates as an adjustment vector, together with the tracking of the tracked region (initiated in step 212), to the presented image, so as to register the presented image with patient 30 on a continuing basis.


The processor continues to apply the stored adjustment vector, which acts as a patient tracking vector, as long as the orientation of step 204 is unchanged. It will be understood that the registration using the stored adjustment vector counteracts any relative movement between the patient and assembly 24.


Steps 200-216 correspond to an initial setup of navigation system 20. Steps 220-232, described below, correspond to steps of the flowchart that may be implemented during the medical procedure for which system 20 is used.


In a continuing imaging step 220, an image of the alignment target and optionally of additional elements of interest (e.g., tools used during the procedure or of a region of interest on the patient's body) is continuously acquired (e.g., once in a predefined time interval). For example, an image of target 44, including target region 120 and indicator 162 is continuously acquired via assembly 24 (optionally controlled by processor 26 but not necessarily). From the images of target region 120 and indicator 162, the processor calculates an orientation metric.


In a decision step 224 the processor checks if the metric calculated in step 220 is different from that calculated in step 212, so as to check if the target region has changed orientation. If the decision returns negative, i.e., there is no change in orientation, then in a continuation step 228 the processor continues to use the existing adjustment coordinates, i.e., those of step 216.


If the decision returns positive, i.e., there is a change in orientation, then in an update step 232 the processor calculates updated adjustment coordinates, by adding the appropriate change-of-orientation vector, from step 204, to the existing adjustment coordinates. The processor applies the updated coordinates in presenting the image of the patient, e.g., of the region close to the bone, to professional 22 in augmented reality assembly 24.


It will be understood that a positive return of decision 224 is typically caused by professional 22 changing the orientation of the alignment target. The professional may change the orientation for example by unscrewing screw 154, realigning socket 124 on surface 170 so that target 44 is in a new orientation, then screwing screw 154 to fix the target in its new orientation. In some embodiments the professional may pause the tracking of target 44 while positioning target 44 in its new orientation.


The professional typically repositions target 44 to improve access to, and/or to improve visibility to, part of the patient. It will be understood that there is no need to repeat steps 200-216 after the repositioning, since the new adjustment coordinates can be calculated from the known geometric relations of the two orientations of target 44. (It will also be understood that regardless of any repositioning, adjustment screw 100 is always accessible via one of apertures 150, since the screw aligns with the apertures.) From steps 228 and 232 control returns to decision step 220, so that during the procedure the processor applies steps 220-232 iteratively.



FIGS. 6A-6D are schematic views of different orientations of a patient marker 240, according to an embodiment of the present invention. Marker 240 comprises a clamp base 294 which is coupled to an alignment target 244. Clamp base 294 is the base of a clamp 242. Apart from the differences described below, the operation of marker 240, clamp 242, base 294, and target 244 is generally similar to that of marker 40, clamp 42, base 94, and target 44, and elements indicated by the same reference numerals in both markers are generally similar in construction and in operation. As for marker 40, target 244 has four different discrete orientations with respect to clamp 242. The axes for marker 240 are as for marker 40.


In contrast to upper surface 170 of clamp base 94 of clamp 42, an upper xy surface 270 of clamp base 294 of clamp 242 is surmounted by a circular turret 300. Fixed to turret 300 are four protuberances 304, distributed symmetrically about a center of the turret, and lying in an xy plane, parallel to surface 270. Turret 300 comprises a recess 308, which is configured to accept an insert 260 that typically is colored or retroreflective.


As for socket 124, a socket 224 of target 244 comprises four apertures 250 distributed symmetrically about a center of the socket. However, socket 224 is smaller than socket 124, and regardless of target orientation, an aperture 250 does not align with screw 100. Rather, socket 224 is small enough so that in at least some target orientations, e.g. those of FIGS. 6A, 6C and 6D, screw 100 is accessible from outside the socket.


A connecting rod 246 connects target 120 to socket 224, but typically has a smaller width than rod 46.


Socket 224 has a lower circular wall 316 which has an internal surface which is configured to mate with turret 300. Set within the wall are four clips 312 which are distributed symmetrically about a center of the socket, and which are positioned to mate with protuberances 304. When clips 312 mate with protuberances 304, they hold socket 224 so that wall 316 surrounds and contacts an outer surface of turret 300, and so the socket is fixed to the turret.


Because protuberances 304 and mating clips 312 are distributed symmetrically, it will be understood that target 244 can mate with clamp 242 in one of four orientations illustrated in FIGS. 6A-6D, and that in each orientation a center of socket 224 aligns with a center of turret 300.


Unlike marker 40, where optical indicator 162 is formed from the visibility of insert 160 through one of apertures 150, in marker 240 an optical indicator 262 comprises insert 260 as viewed through three apertures 250, as illustrated in FIGS. 6A-6D.


The description of the flowchart of FIG. 5 applies to marker 240, mutatis mutandis. For example, a position of indicator 262 may be assumed to be the position of the central one of the three apertures 260.


Markers 40 and 240 each have four symmetrically distributed discrete orientations. However, embodiments of the present invention may have other numbers of symmetrically distributed orientations, where the number may be as little as two.


The number of apertures 150 corresponds to the number of discrete orientations. As exemplified by indicator 162 the number of apertures used to generate the optical indicator may be a single aperture. Alternatively, as exemplified by indicator 262, the number of apertures used to generate the optical indicator may comprise any fixed number of apertures that is at least one less than the total number of apertures. In this case the apertures are selected and arranged so that when rotated, they provide an unambiguous identification of each of the discrete orientations.


Thus, for four apertures, corresponding to four discrete orientations, the indicator may be two adjacent apertures, but not two apertures opposite each other, since two apertures opposite each do not provide an unambiguous identification of each orientation.



FIGS. 7A-7E are schematic views of different orientations of a patient marker 440, according to an embodiment of the present invention. Marker 440 comprises a clamp base 494 which is coupled to an alignment target 444. Clamp base 494 is the base of a clamp 242. Apart from the differences described below, the operation of marker 440, clamp 442, base 494, and target 444 is generally similar to that of marker 40, clamp 42, base 94, and target 44, and elements indicated by the same reference numerals in both markers are generally similar in construction and in operation. Unlike marker 40, where target 44 can only make discrete orientations with respect to clamp 42, target 444 in marker 440 can make multiple non-discrete, substantially continuous, orientations varying from 0°-360° with respect to clamp 442.



FIGS. 7A-7E have been drawn on the same set of xyz axes as for marker 40 (although the axes are rotated 180° compared to those of FIGS. 4B-4E), and orientations are measured as clockwise rotations about the z-axis from the y-axis. FIGS. 7A, 7B, 7C, and 7D, correspond respectively to the target having orientations of 0°, 90°, 180°, and 270° relative to the clamp.



FIG. 7E illustrates the target having an orientation of θ relative to the clamp, where 0°≤θ<360°, and coordinates of a center point of the target region have been marked as (D sin θ, D cos θ) where D is the distance of the target region center point from the z-axis.


In contrast to upper surface 170 of clamp base 94 of clamp 42, an upper xy plane surface 470 of an upper plate 476 of clamp base 494 is circular. Surface 470 has a central circular indent 454 symmetrically situated in the surface, and the indent is terminated at its lower end by a female thread. Surface 470 also has an indent 464 which is in the form of a semicircular arc, centered on a center of circular surface 470. An insert 460 that is a semicircular arc and that is typically colored or retroreflective is inserted into indent 464, and the insert is dimensioned so that an upper surface of the insert is level with surface 470.


A socket 424 of target 444 comprises a planar lower surface 490 which is surrounded by a circular wall 496 that is configured to mate with an outer cylindrical surface 474 of plate 476. Extending from surface 490 are a plurality of arcs 498, distributed symmetrically about a socket central hole 456, configured to mate with indent 454. Socket 424 also comprises a semicircular aperture 468, which is congruent to insert 460.


Target 444 is coupled to clamp 442 by fitting socket 424 to plate 476 so that wall 496 mates with surface 474, and so that arcs 498 mate with indent 454. Once so coupled, target 444 may be held fixedly in place in any orientation selected by professional 22, by screwing a screw 472 into the female thread terminating indent 454.


During a procedure processor 26 is able to determine the orientation of the target, as a value between 0° and 360°, by the imaging of insert 460, and the use of the imaged insert as an optical indicator 462 of the orientation. In one embodiment processor 26 determines the orientation by finding the fraction of the insert visible through aperture 468, as well as a location of the visible insert.


In embodiments of the invention, the fraction may comprise a fractional area of the insert, or alternatively or additionally, a fractional linear dimension, such as an arc length, of the insert. In some embodiments the fractional linear dimension may be measured using a Vernier scale.


Thus, FIG. 7A illustrates a maximum of the insert visible through the aperture, corresponding to an orientation of 0°, and FIG. 7C illustrates a minimum of the insert visible through the aperture, corresponding to an orientation of 180°. FIG. 7B, corresponding to an orientation of 90° illustrates half of the insert visible, the visible half being located below the x-axis, and FIG. 7D, corresponding to an orientation of 270°, illustrates half of the insert visible, the visible half being located above the x-axis.


Other methods for determining the orientation of the target from the imaged insert, such as by finding coordinates of the endpoints of the imaged insert as well as coordinates of an intermediate point on the image, will be apparent, and all such methods are assumed to be comprised within the scope of the present invention.


During a procedure, processor 26 determines the orientation of target 444 from imaging optical indicator 462, as described above. The flowchart of FIG. 8 below describes how the processor uses the values of the orientation during the procedure.



FIG. 8 is a flowchart describing the use of a marker such as marker 440 in the medical procedure referred to above, and FIG. 9 is a diagram explaining some of the steps of the flowchart, according to an embodiment of the present invention.


An initial step 600, wherein clamp 442 is attached to the patient's bone, is substantially similar to initial step 200, described above.


In an attachment step 604, professional 22 attaches socket 424 to base 494, in any convenient orientation of target 444 to clamp 442. The professional uses screw 472 to fix the target to the clamp base.


An imaging step 608 is substantially similar to step 208, described above.


In an analysis step 612, the processor analyzes the image of indicator 462, as described above with reference to FIG. 7E, to determine an angle of orientation θ of the target with clamp 42. In addition, the processor calculates coordinates of the location of target region 120 from the image of the region acquired in step 608.


An adjustment or registration step 616 is substantially as described above for step 216. Thus, in step 616, coordinates of a presented image are adjusted or registered to correspond with an actual image visible to the professional through augmented reality assembly 24. Processor 26 stores the adjustment coordinates, and applies the stored coordinates as an adjustment vector so as to register and track the presented image with patient 30.


Steps 600-616 correspond to an initial setup of navigation system 20 for marker 440. Steps 620-632, described below, correspond to steps of the flowchart that may be implemented during the medical procedure for which system 20 is used.


In a continuing imaging step 620, an image of target 444, including target region 120 and indicator 462 is acquired, e.g., once in a predefined time interval. Optionally, an image of patient 30 may be also acquired. From the image of indicator 462, the processor calculates an angle θ of the orientation of target region 120 relative to the clamp.


In a decision step 624 the processor checks if the angle calculated in step 620 is different from that calculated in step 612, so as to check if the target region has changed orientation. If the decision returns negative, i.e., there is no change in orientation, then in a continuation step 628 the processor continues to use the existing adjustment coordinates, i.e., those of step 616, as an adjustment vector.


If decision 624 returns positive, i.e., there is a change of orientation, then in an update step 632 the processor calculates a change-of-orientation vector, to be added to the existing adjustment vector, so as to enable the processor to maintain registration of images of patient 30 with the patient.



FIG. 9 is a schematic diagram illustrating how the processor calculates the change-of-orientation vector, according to an embodiment of the present invention. A line segment PA, having a length D, represents an initial orientation of adjustment target 444, where A is the center of region 120 and θ1 is the orientation of the center, both values being measured in step 612. A has coordinates (D sin θ1, D cos θ1).


A line segment PB, having a length D, represents a subsequent orientation of adjustment target 444, where B is the center of region 120 and θ2 is the orientation of the center, both values being measured in step 620. B has coordinates (D sin θ2, D cos θ2).


Processor 26 calculates a change-of-orientation vector [AB] as the difference between the coordinates of B and the coordinates of A, as in equation (1):

[AB]=(D(sin θ2−sin θ1),D(cos θ2−cos θ1))  (1)


Returning to the flowchart of FIG. 8, in step 632 the processor adds a change-of-orientation vector, calculated as described for equation (1), to the existing adjustment vector.


From steps 628 and 632 control returns to imaging step 620 and following that decision step 624, so that during the procedure the processor applies steps 620-632 iteratively.


A positive return of decision 624 is typically caused by professional 22 changing the orientation of target 444 by loosening then tightening screw 472. In some embodiments the professional may pause the tracking of target 444 while positioning the target in a new orientation. It will be understood that there is no need to repeat steps 600-616 after any re-orientation, since by iterating steps 620-632 the processor continues to correctly register any acquired image of patient 30 with the patient.


According to some aspects, a patient marker comprising an alignment target and a marker base, which may be anchored to the patient via an anchoring device, is herein disclosed. The alignment target is coupled to the marker base. The alignment target is configured to rotate around an axis of the marker base. The alignment target comprises an optical indicator, The optical indicator indicates the angle of orientation of the alignment target about the marker base axis. The patient marker further comprises an anchoring interface. The anchoring interface is coupled to the marker base. According to some aspects, the marker base comprises the anchoring interface. The alignment target may comprise a socket configured to fit rotatably to the target base. The patient marker is configured to be mounted on a patient anchoring device, e.g., on a base of the anchoring device, via the anchoring interface in a fixed manner and such that the marker base axis is aligned with the anchoring device base axis (e.g., a clamp base axis). The optical indicator may be, for example, in the form of and may be configured to operate according to optical indicators 162, 262 or 462, as described in and with respect to FIGS. 4A-4E, 6A-6D and 7A-7E correspondingly. The alignment target may be configured to rotate around an axis of the marker base, for example according to the rotation mechanisms describe in and with respect to FIGS. 4A-4E. 6A-6D and 7A-7E. Persons skilled in the art will understand how to implement different mechanisms for the optical indicator and different rotation mechanisms for the alignment target, including those disclosed herein, in a patient marker comprising a rotatable alignment target coupled to a base of the marker. The description above explains how, for example, an optical indicator such as indicator 162, together with the optical elements 138 of alignment target 44, may be used to find an angle of orientation of the alignment target about an axis of a base to which the alignment target is connected. The base may be a marker base or an anchoring device base, such as a clamp base. It will be understood that optical elements 138 identify the alignment target.


In some embodiments of the present invention, professional 22 uses a surgical tool in the medical procedure referred to above (FIG. 1). The tool has a tool marker connected to the tool or integrated with the tool in a predefined location, and the augmented reality assembly (e.g., augmented reality assembly 24) is configured to enable the tracking (e.g., by processor 26) of the position of the tool tip. In these embodiments, instead of using an optical indicator such as optical indicator 162 to identify the angle of orientation of the alignment target, professional 22 may use the tool to find the angle, as is described below.


The patient anchoring device or the marker base may comprise a fixed tool point (also will be referred as “tool point” herein below). The tool point location is configured to be fixed with respect to the anchoring device and thus with respect to the patient. For example, the fixed tool point may be connected to a base of the anchoring device. In general, the tool having the tool marker mounted thereon is calibrated prior to its use in a medical procedure via the augmented reality assembly. In the calibration process, a spatial ratio is determined between the tool marker and the tip of the tool, such as a vector between the tool marker (e.g., a retroreflective predefined point on the tool marker) and the tool tip. The professional may touch the fixed tool point with the tip of the calibrated tool, while the alignment target of the patient marker is oriented in a specific orientation with respect to the anchoring device. An image of the tool marker and the oriented patient marker is captured by the capturing device of the augmented reality assembly. A location of the fixed tool point with respect to the alignment target is determined by a processor, such as processor 26, based on the image. Following that, the orientation of the alignment target with respect to the anchoring device and thus the patient may be determined. According to some aspects, the professional may be directed by the augmented reality assembly (e.g., by a software executed by processor 26) to touch the fixed tool point with the tip of the tool. According to some aspects, the alignment target has a plurality of predefined orientation and a vector between the alignment target and the fixed tool point in each orientation is calculated in advance. In such a case, a spatial area corresponding to each predefined orientation of the target alignment may be determined and used in determining the orientation of the alignment target, e.g., instead of an accurate calculation of the alignment target location. As described above, the capturing device continuously capture images of the area of interest, including the patient marker and the tool marker. For each image captured via the capturing device, the processor may check if the tool tip is substantially located at a possible tool point location, corresponding to a possible alignment target orientation, to identify the touching of the tool point by the tool tip. Identification of a specific such location in one or more subsequent images may determine the orientation of the alignment marker.


Those having ordinary skill in the art will be able to adapt the description, mutatis mutandis, for various configurations of alignment targets, including targets 44, 244 and 444.


It will also be understood embodiments of the invention may measure discrete changes of orientation e.g., of 0°, 90°, 180°, or 270°, or continuous changes of orientation through substantially any angle between 0° and 360°.


It will be appreciated that the embodiments described above are cited by way of example, and that the present invention is not limited to what has been particularly shown and described hereinabove. Rather, the scope of the present invention includes both combinations and subcombinations of the various features described hereinabove, as well as variations and modifications thereof which would occur to persons skilled in the art upon reading the foregoing description and which are not disclosed in the prior art.

Claims
  • 1. A marker for image guided surgery, comprising: a base, having a base axis, connecting to an anchoring device; andan alignment target, comprising: a target region having an alignment pattern formed thereon;a socket connected to the target region and configured to fit rotatably to the base, whereby the alignment target is rotatable about the base axis; andan optical indicator for the socket indicating an angle of orientation of the alignment target about the base axis.
  • 2. The marker according to claim 1, wherein the socket is configured to only fit to the base in a plurality of at least two discrete orientations about the base axis.
  • 3. The marker according to claim 2, wherein the plurality of discrete configurations is distributed symmetrically about the base axis.
  • 4. The marker according to claim 2, wherein the plurality comprises four discrete orientations.
  • 5. The marker according to claim 2, wherein the socket comprises a plurality of apertures equal to the plurality of discrete orientations, and wherein the optical indicator is configured to be visible through one of the apertures indicative of one of the discrete orientations.
  • 6. The marker according to claim 2, wherein the socket comprises a plurality of apertures equal to the plurality of discrete orientations, and wherein the optical indicator is configured to be visible through apertures selected and arranged so as to provide an unambiguous identification of each of the discrete orientations.
  • 7. The marker according to claim 1, wherein the socket is configured to fit to the base in a plurality of non-discrete orientations about the base axis.
  • 8. The marker according to claim 7, wherein the socket comprises an aperture, and wherein the optical indicator is congruent with the aperture, and wherein a fraction of the optical indicator visible through the aperture is indicative of one of the non-discrete orientations.
  • 9. The marker according to claim 8, wherein the aperture comprises a semicircular arc.
  • 10. An augmented reality system operative during surgery on a patient, comprising: the marker according to claim 1, wherein the socket is at a fixed distance from the target region; anda processor configured to: track the alignment target during the surgery, provide a patient tracking vector in response to the tracking of the alignment target, calculate a change in the angle of orientation of the alignment target in response to changes in images of the optical indicator, and add a change-of-orientation vector, based only on the fixed distance and the change in the angle of orientation, to the patient tracking vector so as to update the patient tracking vector.
  • 11. The marker according to claim 1, the marker further comprising the anchoring device.
  • 12. The marker according to claim 1, wherein the anchoring device is a clamp or a pin.
  • 13. A marker for image guided surgery, comprising: a base, having a base axis;an interface configured to be coupled to an anchoring device; andan alignment target, comprising: a target region having an alignment pattern formed thereon;a socket connected to the target region and configured to fit rotatably to the base, whereby the alignment target is rotatable about the base axis; andan optical indicator for the socket indicating an angle of orientation of the alignment target about the base axis.
  • 14. The marker of claim 13, wherein the base comprises the interface.
  • 15. The marker of claim 13, wherein the interface is coupled to the anchoring device in a manner that aligns the base axis with a base axis of the anchoring device.
  • 16. A method for enabling rotation of a marker during surgery without requiring re registration, comprising: connecting a base, having a base axis, to an anchoring device;forming an alignment pattern on a target region of an alignment target;connecting a socket to the target region, the socket being at a fixed distance from the target region and being configured to fit rotatably to the base, whereby the alignment target is rotatable about the base axis;providing an optical indicator for the socket indicating an angle of orientation of the alignment target about the base axis;operating an augmented reality system during the surgery on a patient; tracking the alignment target during the surgery;providing a patient tracking vector to the augmented reality system in response to the tracking of the alignment target;calculating a change in the angle of orientation of the alignment target in response to changes in images of the optical indicator; andadding a change-of-orientation vector, based only on the fixed distance and the change in the angle of orientation, to the patient tracking vector so as to update the patient tracking vector.
  • 17. The method according to claim 16, wherein the socket is configured to only fit to the base in a plurality of at least two discrete orientations about the base axis.
  • 18. The method according to claim 17, wherein the socket comprises a plurality of apertures equal to the plurality of discrete orientations, and wherein the optical indicator is configured to be visible through one of the apertures indicative of one of the discrete orientations.
  • 19. The method according to claim 17, wherein the socket comprises a plurality of apertures equal to the plurality of discrete orientations, and wherein the optical indicator is configured to be visible through apertures selected and arranged so as to provide an unambiguous identification of each of the discrete orientations.
  • 20. The method according to claim 16, wherein the socket is configured to fit to the base in a plurality of non-discrete orientations about the base axis.
  • 21. The method according to claim 20, wherein the socket comprises an aperture, and wherein the optical indicator is congruent with the aperture, and wherein a fraction of the optical indicator visible through the aperture is indicative of one of the non discrete orientations.
  • 22. A computer-implemented method for enabling rotation of a marker during surgery without requiring re-registration, wherein the marker is anchored to a patient and configured to fit rotatably to a base, and wherein the marker comprises an alignment pattern and an optical indicator, the optical indicator configured to indicate an angle of orientation of the marker about a base axis of the base and the optical indicator is being at a fixed distance from the alignment pattern, the method comprising: operating a navigation system during the surgery on the patient;tracking the alignment pattern during the surgery;providing a patient tracking vector to the navigation system in response to the tracking of the alignment pattern;calculating a change in the angle of orientation of the alignment pattern in response to changes in images of the optical indicator; andadding a change-of-orientation vector, based only on the fixed distance and the change in the angle of orientation, to the patient tracking vector so as to update the patient tracking vector.
PCT Information
Filing Document Filing Date Country Kind
PCT/IB2021/055242 6/15/2021 WO
Publishing Document Publishing Date Country Kind
WO2021/255627 12/23/2021 WO A
US Referenced Citations (1479)
Number Name Date Kind
3690776 Zaporoshan Sep 1972 A
4459358 Berke Jul 1984 A
4711512 Upatnieks Dec 1987 A
4863238 Brewster Sep 1989 A
4944739 Torre Jul 1990 A
5100420 Green et al. Mar 1992 A
5357292 Wiedner Oct 1994 A
5410802 Buckley May 1995 A
5441042 Putman Aug 1995 A
5442146 Bell et al. Aug 1995 A
5510832 Garcia Apr 1996 A
D370309 Stucky May 1996 S
5620188 McCurry et al. Apr 1997 A
5636255 Ellis Jun 1997 A
5665092 Mangiardi et al. Sep 1997 A
5743731 Lares et al. Apr 1998 A
5771121 Hentschke Jun 1998 A
5792046 Dobrovolny Aug 1998 A
5841507 Barnes Nov 1998 A
6006126 Cosman Dec 1999 A
6038467 De Bliek et al. Mar 2000 A
6125164 Murphy et al. Sep 2000 A
6138530 McClure Oct 2000 A
6147805 Fergason Nov 2000 A
6227667 Halldorsson et al. May 2001 B1
6256529 Holupka et al. Jul 2001 B1
6285505 Melville et al. Sep 2001 B1
6314310 Ben-Haim et al. Nov 2001 B1
6349001 Spitzer Feb 2002 B1
6444192 Mattrey Sep 2002 B1
6447503 Wynne et al. Sep 2002 B1
6449090 Omar et al. Sep 2002 B1
6456405 Horikoshi et al. Sep 2002 B2
6456868 Saito et al. Sep 2002 B2
6474159 Foxlin et al. Nov 2002 B1
6518939 Kikuchi Feb 2003 B1
6527777 Justin Mar 2003 B2
6529331 Massof et al. Mar 2003 B2
6549645 Oikawa et al. Apr 2003 B1
6578962 Amir et al. Jun 2003 B1
6609022 Vilsmeier et al. Aug 2003 B2
6610009 Person Aug 2003 B2
D480476 Martinson et al. Oct 2003 S
6659611 Amir et al. Dec 2003 B2
6675040 Cosman Jan 2004 B1
6683584 Ronzani et al. Jan 2004 B2
6690964 Bieger et al. Feb 2004 B2
6714810 Grzeszczuk et al. Mar 2004 B2
6737425 Yamamoto et al. May 2004 B1
6740882 Weinberg May 2004 B2
6757068 Foxlin Jun 2004 B2
6759200 Stanton, Jr. Jul 2004 B1
6847336 Lemelson et al. Jan 2005 B1
6856324 Sauer et al. Feb 2005 B2
6856826 Seeley et al. Feb 2005 B2
6891518 Sauer et al. May 2005 B2
6900777 Hebert et al. May 2005 B1
6919867 Sauer Jul 2005 B2
6921167 Nagata Jul 2005 B2
6966668 Cugini et al. Nov 2005 B2
6980849 Sasso Dec 2005 B2
6993374 Sasso Jan 2006 B2
6997552 Hung Feb 2006 B1
6999239 Martins et al. Feb 2006 B1
7000262 Bielefeld Feb 2006 B2
7035371 Boese et al. Apr 2006 B2
7043961 Pandey et al. May 2006 B2
7072435 Metz et al. Jul 2006 B2
7103233 Stearns Sep 2006 B2
7107091 Jutras et al. Sep 2006 B2
7112656 Desnoyers et al. Sep 2006 B2
7141812 Appleby et al. Nov 2006 B2
7157459 Ohta et al. Jan 2007 B2
7169785 Timmer et al. Jan 2007 B2
7171255 Holupka et al. Jan 2007 B2
7176936 Sauer et al. Feb 2007 B2
7187792 Fu et al. Mar 2007 B2
7190331 Genc et al. Mar 2007 B2
7194295 Vilsmeier Mar 2007 B2
7215322 Genc et al. May 2007 B2
7229078 Lechot Jun 2007 B2
7231076 Fu et al. Jun 2007 B2
7235076 Pacheco Jun 2007 B2
7239330 Sauer et al. Jul 2007 B2
7259266 Carter et al. Aug 2007 B2
7260426 Schweikard et al. Aug 2007 B2
7269192 Hayashi Sep 2007 B2
7281826 Huang Oct 2007 B2
7315636 Kuduvalli Jan 2008 B2
7320556 Vagn-Erik Jan 2008 B2
7330578 Wang et al. Feb 2008 B2
7359535 Salla et al. Apr 2008 B2
7364314 Nilsen et al. Apr 2008 B2
7366934 Narayan et al. Apr 2008 B1
7379077 Bani-Hashemi et al. May 2008 B2
7431453 Hogan Oct 2008 B2
7435219 Kim Oct 2008 B2
7450743 Sundar et al. Nov 2008 B2
7458977 McGinley et al. Dec 2008 B2
7462852 Appleby et al. Dec 2008 B2
7493153 Ahmed et al. Feb 2009 B2
7505617 Fu et al. Mar 2009 B2
7507968 Wollenweber et al. Mar 2009 B2
7518136 Appleby et al. Apr 2009 B2
7525735 Sottilare et al. Apr 2009 B2
D592691 Chang May 2009 S
D592692 Chang May 2009 S
D592693 Chang May 2009 S
7536216 Geiger et al. May 2009 B2
7542791 Mire et al. Jun 2009 B2
7556428 Sukovic et al. Jul 2009 B2
7557824 Holliman Jul 2009 B2
7563228 Ma et al. Jul 2009 B2
7567834 Clayton et al. Jul 2009 B2
7570791 Frank et al. Aug 2009 B2
7586686 Hall Sep 2009 B1
D602620 Cristoforo Oct 2009 S
7605826 Sauer Oct 2009 B2
7606613 Simon et al. Oct 2009 B2
7607775 Hermanson et al. Oct 2009 B2
7620223 Xu et al. Nov 2009 B2
7623902 Pacheco Nov 2009 B2
7627085 Boyden et al. Dec 2009 B2
7630753 Simon et al. Dec 2009 B2
7633501 Wood et al. Dec 2009 B2
7645050 Wilt et al. Jan 2010 B2
7653226 Guhring et al. Jan 2010 B2
7657075 Viswanathan Feb 2010 B2
7689019 Boese et al. Mar 2010 B2
7689042 Brunner et al. Mar 2010 B2
7689320 Prisco et al. Mar 2010 B2
7699486 Beiner Apr 2010 B1
7699793 Goette et al. Apr 2010 B2
7719769 Sugihara et al. May 2010 B2
D617825 Chang Jun 2010 S
7734327 Colquhoun Jun 2010 B2
D619285 Cristoforo Jul 2010 S
7751865 Jascob et al. Jul 2010 B2
7758204 Klipstein et al. Jul 2010 B2
7768702 Hirose et al. Aug 2010 B2
7769236 Fiala Aug 2010 B2
7773074 Arenson et al. Aug 2010 B2
7774044 Sauer et al. Aug 2010 B2
7822483 Stone et al. Oct 2010 B2
D628307 Krause-Bonte Nov 2010 S
7826902 Stone et al. Nov 2010 B2
7831073 Fu et al. Nov 2010 B2
7831096 Williamson, Jr. Nov 2010 B2
7835778 Foley et al. Nov 2010 B2
7835784 Mire et al. Nov 2010 B2
7837987 Shi et al. Nov 2010 B2
7840093 Fu et al. Nov 2010 B2
7840253 Tremblay et al. Nov 2010 B2
7840256 Lakin et al. Nov 2010 B2
7853305 Simon et al. Dec 2010 B2
7854705 Pawluczyk et al. Dec 2010 B2
7857271 Lees Dec 2010 B2
7860282 Boese et al. Dec 2010 B2
D630766 Harbin Jan 2011 S
7865269 Prisco et al. Jan 2011 B2
7874686 Rossner et al. Jan 2011 B2
7881770 Melkent et al. Feb 2011 B2
7893413 Appleby et al. Feb 2011 B1
7894649 Fu et al. Feb 2011 B2
7920162 Masini et al. Apr 2011 B2
7922391 Essenreiter et al. Apr 2011 B2
7938553 Beiner May 2011 B1
7945310 Gattani et al. May 2011 B2
7953471 Clayton et al. May 2011 B2
7969383 Eberl et al. Jun 2011 B2
7974677 Mire et al. Jul 2011 B2
7985756 Barlow et al. Jul 2011 B2
7991557 Liew et al. Aug 2011 B2
7993353 Roner et al. Aug 2011 B2
7996064 Simon et al. Aug 2011 B2
8004524 Deinzer Aug 2011 B2
8021300 Ma et al. Sep 2011 B2
8022984 Cheong et al. Sep 2011 B2
8045266 Nakamura Oct 2011 B2
8060181 Rodriguez et al. Nov 2011 B2
8068581 Boese et al. Nov 2011 B2
8068896 Daghighian et al. Nov 2011 B2
8077943 Williams et al. Dec 2011 B2
8079957 Ma et al. Dec 2011 B2
8081812 Kreiser Dec 2011 B2
8085075 Huffman et al. Dec 2011 B2
8085897 Morton Dec 2011 B2
8090175 Fu et al. Jan 2012 B2
8092400 Warkentine et al. Jan 2012 B2
8108072 Zhao et al. Jan 2012 B2
8112292 Simon Feb 2012 B2
8116847 Gattani et al. Feb 2012 B2
8120847 Chang Feb 2012 B2
8121255 Sugiyama Feb 2012 B2
8155479 Hoffman et al. Apr 2012 B2
8180132 Gorges et al. May 2012 B2
8180429 Sasso May 2012 B2
8208599 Ye et al. Jun 2012 B2
8216211 Mathis et al. Jul 2012 B2
8221402 Francischelli et al. Jul 2012 B2
8239001 Verard et al. Aug 2012 B2
8244012 Liang et al. Aug 2012 B2
8253778 Atsushi Aug 2012 B2
8271069 Jascob et al. Sep 2012 B2
8280491 Kuduvalli et al. Oct 2012 B2
8285021 Boese et al. Oct 2012 B2
8300315 Kobayashi Oct 2012 B2
8305685 Heine et al. Nov 2012 B2
8306305 Porat et al. Nov 2012 B2
8309932 Haselman et al. Nov 2012 B2
8317320 Huang Nov 2012 B2
8328815 Farr et al. Dec 2012 B2
8335553 Rubner et al. Dec 2012 B2
8335557 Maschke Dec 2012 B2
8340379 Razzaque et al. Dec 2012 B2
8369925 Giesel et al. Feb 2013 B2
8386022 Jutras et al. Feb 2013 B2
8394144 Zehavi et al. Mar 2013 B2
8398541 Dimaio et al. Mar 2013 B2
8444266 Waters May 2013 B2
8457719 Moctezuma De La Barrera et al. Jun 2013 B2
8467851 Mire et al. Jun 2013 B2
8469902 Dick et al. Jun 2013 B2
8475470 Von Jako Jul 2013 B2
8494612 Vetter et al. Jul 2013 B2
8509503 Nahum et al. Aug 2013 B2
8511827 Hua et al. Aug 2013 B2
8531394 Maltz Sep 2013 B2
8540364 Waters Sep 2013 B2
8545012 Waters Oct 2013 B2
8548567 Maschke et al. Oct 2013 B2
8556883 Saleh Oct 2013 B2
8559596 Thomson et al. Oct 2013 B2
8567945 Waters Oct 2013 B2
8571353 Watanabe Oct 2013 B2
8585598 Razzaque et al. Nov 2013 B2
8600001 Schweizer Dec 2013 B2
8600477 Beyar et al. Dec 2013 B2
8605199 Imai Dec 2013 B2
8611988 Miyamoto Dec 2013 B2
8612024 Stone et al. Dec 2013 B2
8634897 Simon et al. Jan 2014 B2
8641621 Razzaque et al. Feb 2014 B2
8643950 König Feb 2014 B2
8644907 Hartmann et al. Feb 2014 B2
8674902 Park et al. Mar 2014 B2
8686923 Eberl et al. Apr 2014 B2
8690581 Ruf et al. Apr 2014 B2
8690776 Razzaque et al. Apr 2014 B2
8692845 Fedorovskaya et al. Apr 2014 B2
8693632 Allison Apr 2014 B2
8694075 Groszmann et al. Apr 2014 B2
8699765 Hao et al. Apr 2014 B2
8705829 Frank et al. Apr 2014 B2
8737708 Hartmann et al. May 2014 B2
8746887 Shestak et al. Jun 2014 B2
8764025 Gao Jul 2014 B1
8784450 Moskowitz et al. Jul 2014 B2
8786689 Liu Jul 2014 B1
D710545 Wu Aug 2014 S
D710546 Wu Aug 2014 S
8827934 Chopra et al. Sep 2014 B2
8831706 Fu et al. Sep 2014 B2
8836768 Rafii et al. Sep 2014 B1
8838199 Simon et al. Sep 2014 B2
8848977 Bammer et al. Sep 2014 B2
8855395 Baturin et al. Oct 2014 B2
8878900 Yang et al. Nov 2014 B2
8879815 Miao et al. Nov 2014 B2
8885177 Ben-Yishai et al. Nov 2014 B2
8890772 Woo et al. Nov 2014 B2
8890773 Pederson Nov 2014 B1
8890943 Lee et al. Nov 2014 B2
8897514 Feikas et al. Nov 2014 B2
8900131 Chopra et al. Dec 2014 B2
8903150 Star-Lack et al. Dec 2014 B2
8908952 Isaacs et al. Dec 2014 B2
8911358 Koninckx et al. Dec 2014 B2
8917268 Johnsen et al. Dec 2014 B2
8920776 Gaiger et al. Dec 2014 B2
8922589 Laor Dec 2014 B2
8941559 Bar-Zeev et al. Jan 2015 B2
8942455 Chou et al. Jan 2015 B2
8950877 Northey et al. Feb 2015 B2
8953246 Koenig Feb 2015 B2
8965583 Ortmaier et al. Feb 2015 B2
8969829 Wollenweber et al. Mar 2015 B2
8989349 Thomson et al. Mar 2015 B2
8992580 Bar et al. Mar 2015 B2
8994729 Nakamura Mar 2015 B2
8994795 Oh Mar 2015 B2
9004711 Gerolemou Apr 2015 B2
9005211 Brundobler et al. Apr 2015 B2
9011441 Bertagnoli et al. Apr 2015 B2
9057759 Klingenbeck et al. Jun 2015 B2
9060757 Lawson et al. Jun 2015 B2
9066751 Sasso Jun 2015 B2
9081436 Berme et al. Jul 2015 B1
9084635 Nuckley et al. Jul 2015 B2
9085643 Svanborg et al. Jul 2015 B2
9087471 Miao Jul 2015 B2
9100643 McDowall et al. Aug 2015 B2
9101394 Arata et al. Aug 2015 B2
9104902 Xu et al. Aug 2015 B2
9111175 Strommer et al. Aug 2015 B2
9123155 Cunningham et al. Sep 2015 B2
9125556 Zehavi et al. Sep 2015 B2
9129054 Nawana et al. Sep 2015 B2
9129372 Kriston et al. Sep 2015 B2
9132361 Smithwick Sep 2015 B2
9135706 Zagorchev et al. Sep 2015 B2
9141873 Takemoto Sep 2015 B2
9142020 Deguise et al. Sep 2015 B2
9149317 Arthur et al. Oct 2015 B2
9165203 McCarthy Oct 2015 B2
9165362 Siewerdsen et al. Oct 2015 B2
9179984 Teichman et al. Nov 2015 B2
D746354 Chang Dec 2015 S
9208916 Appleby et al. Dec 2015 B2
9220573 Kendrick et al. Dec 2015 B2
9225895 Kozinski Dec 2015 B2
9232982 Soler et al. Jan 2016 B2
9235934 Mandella et al. Jan 2016 B2
9240046 Carrell et al. Jan 2016 B2
9244278 Sugiyama et al. Jan 2016 B2
9247240 Park et al. Jan 2016 B2
9259192 Ishihara Feb 2016 B2
9265572 Fuchs et al. Feb 2016 B2
9269192 Kobayashi Feb 2016 B2
9283052 Ponce Mar 2016 B2
9286730 Bar-Zeev et al. Mar 2016 B2
9289267 Sauer et al. Mar 2016 B2
9300949 Ahearn Mar 2016 B2
9305354 Burlon et al. Apr 2016 B2
9310591 Hua et al. Apr 2016 B2
9320474 Demri Apr 2016 B2
9323055 Baillot Apr 2016 B2
9330477 Rappel May 2016 B2
9335547 Takano et al. May 2016 B2
9335567 Nakamura May 2016 B2
9341704 Picard May 2016 B2
9344686 Moharir May 2016 B2
9349066 Koo May 2016 B2
9349520 Demetriou May 2016 B2
9364294 Razzaque et al. Jun 2016 B2
9370332 Paladini et al. Jun 2016 B2
9373166 Azar Jun 2016 B2
9375639 Kobayashi et al. Jun 2016 B2
9378558 Kajiwara et al. Jun 2016 B2
9380287 Nistico Jun 2016 B2
9387008 Sarvestani Jul 2016 B2
9392129 Simmons Jul 2016 B2
9395542 Tilleman et al. Jul 2016 B2
9398936 Razzaque et al. Jul 2016 B2
9400384 Griffith Jul 2016 B2
9414041 Ko Aug 2016 B2
9424611 Kanjirathinkal et al. Aug 2016 B2
9424641 Wiemker et al. Aug 2016 B2
9427286 Siewerdsen et al. Aug 2016 B2
9438894 Park Sep 2016 B2
9443488 Borenstein Sep 2016 B2
9453804 Tahtali Sep 2016 B2
9456878 Macfarlane et al. Oct 2016 B2
9465235 Chang Oct 2016 B2
9468373 Larsen Oct 2016 B2
9470908 Frankel Oct 2016 B1
9473766 Douglas Oct 2016 B2
9492222 Singh Nov 2016 B2
9495585 Bicer et al. Nov 2016 B2
9498132 Maier-Hein et al. Nov 2016 B2
9498231 Haider et al. Nov 2016 B2
9499999 Nanqing Nov 2016 B2
9507155 Morimoto Nov 2016 B2
9513495 Waters Dec 2016 B2
9521966 Schwartz Dec 2016 B2
9526443 Berme Dec 2016 B1
9530382 Simmons Dec 2016 B2
9532846 Nakamura Jan 2017 B2
9532849 Anderson et al. Jan 2017 B2
9533407 Ragner Jan 2017 B1
9538962 Hannaford et al. Jan 2017 B1
9545233 Sirpad Jan 2017 B2
9546779 Rementer Jan 2017 B2
9547174 Gao et al. Jan 2017 B2
9547940 Sun et al. Jan 2017 B1
9557566 Fujimaki Jan 2017 B2
9560318 Reina et al. Jan 2017 B2
9561095 Nguyen Feb 2017 B1
9561446 Brecher Feb 2017 B2
9565415 Zhang et al. Feb 2017 B2
9572661 Robin Feb 2017 B2
9576398 Zehner et al. Feb 2017 B1
9576556 Simmons Feb 2017 B2
9581822 Morimoto Feb 2017 B2
9610056 Lavallee et al. Apr 2017 B2
9612657 Bertram et al. Apr 2017 B2
9626936 Bell Apr 2017 B2
9629595 Walker Apr 2017 B2
9633431 Merlet Apr 2017 B2
9645395 Bolas et al. May 2017 B2
9646423 Sun et al. May 2017 B1
9672597 Amiot Jun 2017 B2
9672607 Demri et al. Jun 2017 B2
9672640 Kleiner Jun 2017 B2
9675306 Morton Jun 2017 B2
9675319 Razzaque Jun 2017 B1
9684980 Royalty et al. Jun 2017 B2
9690119 Garofolo et al. Jun 2017 B2
RE46463 Feinbloom Jul 2017 E
9693748 Rai et al. Jul 2017 B2
9710968 Dillavou et al. Jul 2017 B2
9713502 Finkman Jul 2017 B2
9724119 Hissong Aug 2017 B2
9724165 Arata et al. Aug 2017 B2
9726888 Giartisio Aug 2017 B2
9728006 Varga Aug 2017 B2
9729831 Birnkrant Aug 2017 B2
9746739 Alton et al. Aug 2017 B2
9757034 Desjardins Sep 2017 B2
9757087 Simon et al. Sep 2017 B2
9766441 Rappel Sep 2017 B2
9766459 Alton et al. Sep 2017 B2
9767608 Lee et al. Sep 2017 B2
9770203 Berme Sep 2017 B1
9772102 Ferguson Sep 2017 B1
9772495 Tam Sep 2017 B2
9791138 Feinbloom Oct 2017 B1
9800995 Libin Oct 2017 B2
9805504 Zhang Oct 2017 B2
9808148 Miller Nov 2017 B2
9839448 Reckling et al. Dec 2017 B2
9844413 Daon et al. Dec 2017 B2
9851080 Wilt Dec 2017 B2
9858663 Penney et al. Jan 2018 B2
9861446 Lang Jan 2018 B2
9864214 Fass Jan 2018 B2
9872733 Shoham et al. Jan 2018 B2
9875544 Rai et al. Jan 2018 B2
9877642 Duret Jan 2018 B2
9885465 Nguyen Feb 2018 B2
9886552 Dillavou et al. Feb 2018 B2
9886760 Liu et al. Feb 2018 B2
9892564 Cvetko et al. Feb 2018 B1
9898866 Fuchs et al. Feb 2018 B2
9901414 Lively Feb 2018 B2
9911187 Steinle Mar 2018 B2
9911236 Bar et al. Mar 2018 B2
9927611 Rudy Mar 2018 B2
9928629 Benishti et al. Mar 2018 B2
9940750 Dillavou et al. Apr 2018 B2
9943374 Merritt et al. Apr 2018 B2
9947110 Haimerl Apr 2018 B2
9952664 Border et al. Apr 2018 B2
9956054 Aguirre-Valencia May 2018 B2
9958674 Border May 2018 B2
9959620 Merlet May 2018 B2
9959629 Dillavou et al. May 2018 B2
9965681 Border et al. May 2018 B2
9968297 Connor May 2018 B2
9980780 Lang May 2018 B2
9986228 Woods May 2018 B2
D824523 Paoli et al. Jul 2018 S
10010379 Gibby et al. Jul 2018 B1
10013531 Richards Jul 2018 B2
10015243 Kazerani et al. Jul 2018 B2
10016243 Esterberg Jul 2018 B2
10022064 Kim et al. Jul 2018 B2
10022065 Yishai et al. Jul 2018 B2
10022104 Sell et al. Jul 2018 B2
10023615 Bonny Jul 2018 B2
10026015 Cavusoglu Jul 2018 B2
10034713 Yang et al. Jul 2018 B2
10042167 McDowall et al. Aug 2018 B2
10046165 Frewin Aug 2018 B2
10055838 Elenbaas et al. Aug 2018 B2
10066816 Chang Sep 2018 B2
10067359 Ushakov Sep 2018 B1
10073515 Awdeh Sep 2018 B2
10080616 Wilkinson et al. Sep 2018 B2
10082680 Chang Sep 2018 B2
10085709 Lavallee et al. Oct 2018 B2
10105187 Corndorf et al. Oct 2018 B2
10107483 Oren Oct 2018 B2
10108833 Hong et al. Oct 2018 B2
10123840 Dorman Nov 2018 B2
10130378 Bryan Nov 2018 B2
10132483 Feinbloom Nov 2018 B1
10134166 Benishti et al. Nov 2018 B2
10134194 Kepner Nov 2018 B2
10139652 Windham Nov 2018 B2
10139920 Isaacs Nov 2018 B2
10142496 Rao Nov 2018 B1
10151928 Ushakov Dec 2018 B2
10154239 Casas Dec 2018 B2
10159530 Lang Dec 2018 B2
10163207 Merlet Dec 2018 B2
10166079 McLachlin et al. Jan 2019 B2
10175507 Nakamura Jan 2019 B2
10175753 Boesen Jan 2019 B2
10181361 Dillavou et al. Jan 2019 B2
10186055 Takahashi Jan 2019 B2
10188672 Wagner Jan 2019 B2
10194131 Casas Jan 2019 B2
10194990 Amanatullah et al. Feb 2019 B2
10194993 Roger et al. Feb 2019 B2
10195076 Fateh Feb 2019 B2
10197803 Badiali et al. Feb 2019 B2
10197816 Waisman Feb 2019 B2
10207315 Appleby Feb 2019 B2
10212517 Beltran et al. Feb 2019 B1
10230719 Vaugn Mar 2019 B2
10231893 Lei Mar 2019 B2
10235606 Miao Mar 2019 B2
10240769 Braganca Mar 2019 B1
10247965 Ton Apr 2019 B2
10251724 McLachlin et al. Apr 2019 B2
10261324 Chuang et al. Apr 2019 B2
10262424 Ketcha et al. Apr 2019 B2
10274731 Maimone Apr 2019 B2
10278777 Lang May 2019 B1
10292768 Lang May 2019 B2
10296805 Yang et al. May 2019 B2
10319154 Chakravarthula et al. Jun 2019 B1
10326975 Casas Jun 2019 B2
10332267 Rai et al. Jun 2019 B2
10339719 Jagga et al. Jul 2019 B2
10352543 Braganca Jul 2019 B1
10357146 Fiebel Jul 2019 B2
10357574 Hilderbrand Jul 2019 B2
10366489 Boettger et al. Jul 2019 B2
10368947 Lang Aug 2019 B2
10368948 Tripathi Aug 2019 B2
10382748 Benishti et al. Aug 2019 B2
10383654 Yilmaz et al. Aug 2019 B2
10386645 Shousha Aug 2019 B2
10388076 Bar-Zeev et al. Aug 2019 B2
10398514 Ryan et al. Sep 2019 B2
10401657 Jiang et al. Sep 2019 B2
10405825 Rai et al. Sep 2019 B2
10405927 Lang Sep 2019 B1
10413752 Berlinger et al. Sep 2019 B2
10419655 Sivan Sep 2019 B2
10420626 Tokuda et al. Sep 2019 B2
10420813 Newell-Rogers Sep 2019 B2
10424115 Ellerbrock Sep 2019 B2
D862469 Sadot et al. Oct 2019 S
10426554 Siewerdsen et al. Oct 2019 B2
10429675 Greget Oct 2019 B2
10431008 Djajadiningrat Oct 2019 B2
10433814 Razzaque Oct 2019 B2
10434335 Takahashi Oct 2019 B2
10441236 Bar-Tal et al. Oct 2019 B2
10444514 Abou Shousha et al. Oct 2019 B2
10447947 Liu Oct 2019 B2
10448003 Grafenberg Oct 2019 B2
10449040 Lashinski Oct 2019 B2
10453187 Peterson Oct 2019 B2
10463434 Siegler et al. Nov 2019 B2
10465892 Feinbloom Nov 2019 B1
10466487 Blum et al. Nov 2019 B2
10470732 Baumgart Nov 2019 B2
10473314 Braganca Nov 2019 B1
10485989 Jordan Nov 2019 B2
10488663 Choi Nov 2019 B2
D869772 Gand Dec 2019 S
D870977 Berggren et al. Dec 2019 S
10492755 Lin et al. Dec 2019 B2
10499997 Weinstein et al. Dec 2019 B2
10502363 Edwards et al. Dec 2019 B2
10504231 Fiala Dec 2019 B2
10507066 DiMaio Dec 2019 B2
10511822 Casas Dec 2019 B2
10517544 Taguchi Dec 2019 B2
10537395 Perez Jan 2020 B2
10540780 Cousins Jan 2020 B1
10543485 Ismagilov Jan 2020 B2
10546423 Jones et al. Jan 2020 B2
10548557 Lim Feb 2020 B2
10555775 Hoffman Feb 2020 B2
10568535 Roberts et al. Feb 2020 B2
10571696 Urey et al. Feb 2020 B2
10571716 Chapiro Feb 2020 B2
10573086 Bar-Zeev et al. Feb 2020 B2
10573087 Gallop et al. Feb 2020 B2
10602114 Casas Feb 2020 B2
10577630 Zhang Mar 2020 B2
10586400 Douglas Mar 2020 B2
10591737 Yildiz et al. Mar 2020 B2
10592748 Cousins Mar 2020 B1
10595716 Nazareth Mar 2020 B2
10601950 Devam et al. Mar 2020 B2
10603113 Lang Mar 2020 B2
10603133 Wang et al. Mar 2020 B2
10606085 Toyama Mar 2020 B2
10594998 Casas Apr 2020 B1
10610172 Hummel et al. Apr 2020 B2
10610179 Altmann Apr 2020 B2
10613352 Knoll Apr 2020 B2
10617566 Esmonde Apr 2020 B2
10620460 Carabin Apr 2020 B2
10621738 Miao et al. Apr 2020 B2
10625099 Takahashi Apr 2020 B2
10626473 Mariani Apr 2020 B2
10631905 Asfora et al. Apr 2020 B2
10631907 Zucker Apr 2020 B2
10634331 Feinbloom Apr 2020 B1
10634921 Blum et al. Apr 2020 B2
10638080 Ovchinnikov Apr 2020 B2
10646285 Siemionow et al. May 2020 B2
10650513 Penney et al. May 2020 B2
10650594 Jones May 2020 B2
10652525 Woods May 2020 B2
10653495 Gregerson et al. May 2020 B2
10660715 Dozeman May 2020 B2
10663738 Carlvik May 2020 B2
10665033 Bar-Zeev et al. May 2020 B2
10670937 Alton et al. Jun 2020 B2
10672145 Albiol et al. Jun 2020 B2
10682112 Pizaine Jun 2020 B2
10682767 Grafenberg et al. Jun 2020 B2
10687901 Thomas Jun 2020 B2
10691397 Clements Jun 2020 B1
10702713 Mori Jul 2020 B2
10706540 Merlet Jul 2020 B2
10709398 Schweizer Jul 2020 B2
10713801 Jordan Jul 2020 B2
10716643 Justin et al. Jul 2020 B2
10722733 Takahashi Jul 2020 B2
10725535 Yu Jul 2020 B2
10731832 Koo Aug 2020 B2
10732721 Clements Aug 2020 B1
10742949 Casas Aug 2020 B2
10743939 Lang Aug 2020 B1
10743943 Razeto et al. Aug 2020 B2
10747315 Tungare Aug 2020 B2
10748319 Tao et al. Aug 2020 B1
10758315 Johnson et al. Sep 2020 B2
10777094 Rao Sep 2020 B1
10777315 Zehavi Sep 2020 B2
10781482 Gubatayao Sep 2020 B2
10792110 Leung et al. Oct 2020 B2
10799145 Found Oct 2020 B2
10799296 Lang Oct 2020 B2
10799298 Crawford et al. Oct 2020 B2
10799316 Sela et al. Oct 2020 B2
10810799 Tepper et al. Oct 2020 B2
10818019 Piat Oct 2020 B2
10818101 Gallop et al. Oct 2020 B2
10818199 Buras et al. Oct 2020 B2
10825563 Gibby et al. Nov 2020 B2
10827164 Perreault et al. Nov 2020 B2
10831943 Santarone Nov 2020 B2
10835296 Elimelech Nov 2020 B2
10838206 Fortin-Deschenes et al. Nov 2020 B2
10839629 Jones Nov 2020 B2
10839956 Beydoun et al. Nov 2020 B2
10841556 Casas Nov 2020 B2
10842002 Chang Nov 2020 B2
10842461 Johnson et al. Nov 2020 B2
10849691 Zucker Dec 2020 B2
10849693 Lang Dec 2020 B2
10849710 Liu Dec 2020 B2
10861236 Geri et al. Dec 2020 B2
10865220 Ebetino Dec 2020 B2
10869517 Halpern Dec 2020 B1
10869727 Yanof et al. Dec 2020 B2
10872472 Watola Dec 2020 B2
10877262 Luxembourg Dec 2020 B1
10877296 Lindsey Dec 2020 B2
10878639 Douglas et al. Dec 2020 B2
10893260 Trail et al. Jan 2021 B2
10895742 Schneider Jan 2021 B2
10895743 Dausmann Jan 2021 B2
10895906 West et al. Jan 2021 B2
10898151 Harding et al. Jan 2021 B2
10908420 Lee et al. Feb 2021 B2
10921595 Rakshit Feb 2021 B2
10921613 Gupta et al. Feb 2021 B2
10928321 Rawle Feb 2021 B2
10928638 Ninan Feb 2021 B2
10929670 Troy et al. Feb 2021 B1
10935815 Castaneda Mar 2021 B1
10935816 Ban et al. Mar 2021 B2
10936537 Huston Mar 2021 B2
10939973 Dimaio et al. Mar 2021 B2
10939977 Messinger et al. Mar 2021 B2
10941933 Ferguson Mar 2021 B2
10946108 Zhang et al. Mar 2021 B2
10950338 Douglas Mar 2021 B2
10951872 Casas Mar 2021 B2
10964095 Douglas Mar 2021 B1
10964124 Douglas Mar 2021 B1
10966768 Poulos Apr 2021 B2
10969587 McDowall et al. Apr 2021 B2
10993754 Kuntz et al. May 2021 B2
11000335 Dorman May 2021 B2
11002994 Jiang et al. May 2021 B2
11006093 Hegyi May 2021 B1
11013550 Rioux et al. May 2021 B2
11013560 Lang May 2021 B2
11013562 Marti et al. May 2021 B2
11013573 Chang May 2021 B2
11013900 Malek et al. May 2021 B2
11016302 Freeman et al. May 2021 B2
11019988 Fiebel et al. Jun 2021 B2
11027027 Manning et al. Jun 2021 B2
11029147 Abovitz et al. Jun 2021 B2
11030809 Wang Jun 2021 B2
11041173 Zhang et al. Jun 2021 B2
11045663 Mori et al. Jun 2021 B2
11049293 Chae et al. Jun 2021 B2
11049476 Fuchs et al. Jun 2021 B2
11050990 Casas Jun 2021 B2
11057505 Dharmatilleke Jul 2021 B2
11058390 Douglas Jul 2021 B1
11061257 Hakim Jul 2021 B1
11064904 Kay et al. Jul 2021 B2
11065062 Frushour et al. Jul 2021 B2
11067387 Marell et al. Jul 2021 B2
11071497 Hallack et al. Jul 2021 B2
11079596 Hua et al. Aug 2021 B2
11087039 Duff et al. Aug 2021 B2
11090019 Siemionow et al. Aug 2021 B2
11097129 Sakata et al. Aug 2021 B2
11099376 Steier et al. Aug 2021 B1
11103320 Leboeuf et al. Aug 2021 B2
D930162 Cremer et al. Sep 2021 S
11109762 Steier et al. Sep 2021 B1
11112611 Kessler et al. Sep 2021 B1
11122164 Gigante Sep 2021 B2
11123604 Fung Sep 2021 B2
11129562 Roberts et al. Sep 2021 B2
11132055 Jones et al. Sep 2021 B2
11135015 Crawford et al. Oct 2021 B2
11135016 Frielinghaus et al. Oct 2021 B2
11137610 Kessler et al. Oct 2021 B1
11141221 Hobeika et al. Oct 2021 B2
11153549 Casas Oct 2021 B2
11153555 Healy et al. Oct 2021 B1
11163176 Karafin et al. Nov 2021 B2
11164324 Liu et al. Nov 2021 B2
11166006 Hegyi Nov 2021 B2
11169380 Manly et al. Nov 2021 B2
11172990 Lang Nov 2021 B2
11179136 Kohli et al. Nov 2021 B2
11180557 Noelle Nov 2021 B2
11181747 Kessler et al. Nov 2021 B1
11185891 Cousins et al. Nov 2021 B2
11187907 Osterman et al. Nov 2021 B2
11202682 Staunton et al. Dec 2021 B2
11207150 Healy et al. Dec 2021 B2
11217028 Jones et al. Jan 2022 B2
11224483 Steinberg et al. Jan 2022 B2
11224763 Takahashi et al. Jan 2022 B2
11227417 Berlinger et al. Jan 2022 B2
11231787 Isaacs et al. Jan 2022 B2
11243404 McDowall et al. Feb 2022 B2
11244508 Kazanzides et al. Feb 2022 B2
11253216 Crawford et al. Feb 2022 B2
11253323 Hughes et al. Feb 2022 B2
11257190 Mao et al. Feb 2022 B2
11257241 Tao Feb 2022 B2
11263772 Siemionow et al. Mar 2022 B2
11269401 West et al. Mar 2022 B2
11272151 Casas Mar 2022 B2
11278359 Siemionow et al. Mar 2022 B2
11278413 Lang Mar 2022 B1
11280480 Wilt et al. Mar 2022 B2
11284846 Graumann et al. Mar 2022 B2
11291521 Im Apr 2022 B2
11294167 Ishimoda Apr 2022 B2
11297285 Pierce Apr 2022 B2
11300252 Nguyen Apr 2022 B2
11300790 Cheng et al. Apr 2022 B2
11304621 Merschon et al. Apr 2022 B2
11304759 Kovtun et al. Apr 2022 B2
11307402 Steier et al. Apr 2022 B2
11308663 Alhrishy et al. Apr 2022 B2
11311341 Lang Apr 2022 B2
11317973 Calloway et al. May 2022 B2
11337763 Choi May 2022 B2
11348257 Lang May 2022 B2
11350072 Quiles Casas May 2022 B1
11350965 Yilmaz et al. Jun 2022 B2
11351006 Aferzon et al. Jun 2022 B2
11354813 Piat et al. Jun 2022 B2
11360315 Tu et al. Jun 2022 B2
11373342 Stafford et al. Jun 2022 B2
11382699 Wassall et al. Jul 2022 B2
11382700 Calloway et al. Jul 2022 B2
11382712 Elimelech et al. Jul 2022 B2
11382713 Healy et al. Jul 2022 B2
11389252 Gera Jul 2022 B2
11393229 Zhou et al. Jul 2022 B2
11399895 Soper et al. Aug 2022 B2
11402524 Song et al. Aug 2022 B2
11406338 Tolkowsky Aug 2022 B2
11412202 Hegyi Aug 2022 B2
11423554 Borsdorf et al. Aug 2022 B2
11430203 Navab et al. Aug 2022 B2
11432828 Lang Sep 2022 B1
11432931 Lang Sep 2022 B2
11443428 Petersen et al. Sep 2022 B2
11443431 Flossmann et al. Sep 2022 B2
11452568 Lang Sep 2022 B2
11452570 Tolkowsky Sep 2022 B2
11460915 Frielinghaus et al. Oct 2022 B2
11461936 Freeman et al. Oct 2022 B2
11461983 Jones et al. Oct 2022 B2
11464580 Kemp et al. Oct 2022 B2
11464581 Calloway Oct 2022 B2
11475625 Douglas Oct 2022 B1
11478214 Siewerdsen et al. Oct 2022 B2
11483532 Quiles Casas Oct 2022 B2
11488021 Sun et al. Nov 2022 B2
11490986 BEn-Yishai Nov 2022 B2
11510750 Dulin et al. Nov 2022 B2
11513358 McDowall et al. Nov 2022 B2
11527002 Govari Dec 2022 B2
11528393 Garofolo et al. Dec 2022 B2
11544031 Harviainen Jan 2023 B2
11573420 Sarma et al. Feb 2023 B2
11589927 Oezbek et al. Feb 2023 B2
11627924 Alexandroni et al. Apr 2023 B2
11644675 Manly et al. May 2023 B2
11648016 Hathaway et al. May 2023 B2
11651499 Wang et al. May 2023 B2
11657518 Ketcha et al. May 2023 B2
11666458 Kim et al. Jun 2023 B2
11669984 Siewerdsen et al. Jun 2023 B2
11686947 Loyola et al. Jun 2023 B2
11699236 Avital et al. Jul 2023 B2
11712582 Miyazaki et al. Aug 2023 B2
11715210 Haslam et al. Aug 2023 B2
11719941 Russell Aug 2023 B2
11730389 Farshad et al. Aug 2023 B2
11733516 Edwin et al. Aug 2023 B2
11734901 Jones et al. Aug 2023 B2
11744657 Leboeuf et al. Sep 2023 B2
11750794 Benishti et al. Sep 2023 B2
11766296 Wolf et al. Sep 2023 B2
11798178 Merlet Oct 2023 B2
11801097 Crawford et al. Oct 2023 B2
11801115 Elimelech et al. Oct 2023 B2
11808943 Robaina et al. Nov 2023 B2
11815683 Sears et al. Nov 2023 B2
11826111 Mahfouz Nov 2023 B2
11832886 Dorman Dec 2023 B2
11838493 Healy et al. Dec 2023 B2
11839433 Schaewe et al. Dec 2023 B2
11839501 Takahashi et al. Dec 2023 B2
11864934 Junio et al. Jan 2024 B2
11885752 St-Aubin et al. Jan 2024 B2
11892647 Hung et al. Feb 2024 B2
11896445 Gera et al. Feb 2024 B2
11900620 Lalys et al. Feb 2024 B2
11914155 Zhu et al. Feb 2024 B2
11918310 Roh et al. Mar 2024 B1
11922631 Haslam et al. Mar 2024 B2
11941814 Crawford et al. Mar 2024 B2
11944508 Cowin et al. Apr 2024 B1
11948265 Gibby et al. Apr 2024 B2
11950968 Wiggermann Apr 2024 B2
11957420 Lang Apr 2024 B2
11961193 Pelzl et al. Apr 2024 B2
11963723 Vilsmeier et al. Apr 2024 B2
11972582 Yan et al. Apr 2024 B2
11974819 Finley et al. May 2024 B2
11974887 Elimelech et al. May 2024 B2
11977232 Wu et al. May 2024 B2
11980429 Wolf et al. May 2024 B2
11980506 Wolf et al. May 2024 B2
11980507 Elimelech et al. May 2024 B2
11980508 Elimelech et al. May 2024 B2
11983824 Avisar et al. May 2024 B2
12002171 Jones et al. Jun 2024 B2
12010285 Quiles Casas Jun 2024 B2
12014497 Hong et al. Jun 2024 B2
12019314 Steines et al. Jun 2024 B1
12033322 Laaksonen et al. Jul 2024 B2
12044856 Gera et al. Jul 2024 B2
12044858 Gera et al. Jul 2024 B2
12056830 Cvetko et al. Aug 2024 B2
12059281 Weingarten et al. Aug 2024 B2
12063338 Quiles Casas Aug 2024 B2
12063345 Benishti et al. Aug 2024 B2
12069233 Benishti et al. Aug 2024 B2
12076158 Geiger et al. Sep 2024 B2
12076196 Elimelech et al. Sep 2024 B2
12079385 Ben-Yishai et al. Sep 2024 B2
20020082498 Wendt et al. Jun 2002 A1
20030059097 Abovitz et al. Mar 2003 A1
20030117393 Sauer et al. Jun 2003 A1
20030130576 Seeley et al. Jul 2003 A1
20030156144 Morita Aug 2003 A1
20030210812 Khamene et al. Nov 2003 A1
20030225329 Rossner et al. Dec 2003 A1
20040019263 Jutras et al. Jan 2004 A1
20040030237 Lee et al. Feb 2004 A1
20040138556 Cosman Jul 2004 A1
20040152955 McGinley et al. Aug 2004 A1
20040171930 Grimm et al. Sep 2004 A1
20040238732 State et al. Dec 2004 A1
20050017972 Poole et al. Jan 2005 A1
20050024586 Teiwes et al. Feb 2005 A1
20050119639 McCombs et al. Jun 2005 A1
20050154296 Lechner et al. Jul 2005 A1
20050203367 Ahmed et al. Sep 2005 A1
20050203380 Sauer et al. Sep 2005 A1
20050215879 Chuanggui Sep 2005 A1
20050267358 Tuma et al. Dec 2005 A1
20060072124 Smetak et al. Apr 2006 A1
20060134198 Tawa et al. Jun 2006 A1
20060147100 Fitzpatrick Jul 2006 A1
20060176242 Jaramaz et al. Aug 2006 A1
20070018975 Chuanggui et al. Jan 2007 A1
20070058261 Sugihara et al. Mar 2007 A1
20070100325 Jutras et al. May 2007 A1
20070183041 McCloy et al. Aug 2007 A1
20070273610 Baillot Nov 2007 A1
20080002809 Bodduluri Jan 2008 A1
20080007645 McCutchen Jan 2008 A1
20080035266 Danziger Feb 2008 A1
20080085033 Haven et al. Apr 2008 A1
20080159612 Fu et al. Jul 2008 A1
20080183065 Goldbach Jul 2008 A1
20080221625 Hufner et al. Sep 2008 A1
20080253527 Boyden et al. Oct 2008 A1
20080262812 Arata et al. Oct 2008 A1
20080287728 Mostafavi et al. Nov 2008 A1
20090005961 Grabowski et al. Jan 2009 A1
20090018437 Cooke Jan 2009 A1
20090024127 Lechner et al. Jan 2009 A1
20090036902 DiMaio et al. Feb 2009 A1
20090062869 Claverie et al. Mar 2009 A1
20090099445 Burger Apr 2009 A1
20090123452 Madison May 2009 A1
20090227847 Tepper et al. Sep 2009 A1
20090285366 Essenreiter et al. Nov 2009 A1
20090300540 Russell Dec 2009 A1
20100076305 Maier-Hein et al. Mar 2010 A1
20100094308 Tatsumi et al. Apr 2010 A1
20100106010 Rubner et al. Apr 2010 A1
20100114110 Taft et al. May 2010 A1
20100138939 Bentzon et al. Jun 2010 A1
20100149073 Chaum et al. Jun 2010 A1
20100172567 Prokoski Jul 2010 A1
20100210939 Hartmann et al. Aug 2010 A1
20100266220 Zagorchev et al. Oct 2010 A1
20100274124 Jascob et al. Oct 2010 A1
20110004259 Stallings et al. Jan 2011 A1
20110098553 Robbins et al. Apr 2011 A1
20110105895 Kornblau et al. May 2011 A1
20110216060 Weising et al. Sep 2011 A1
20110245625 Trovato et al. Oct 2011 A1
20110254922 Schaerer et al. Oct 2011 A1
20110306873 Shenai et al. Dec 2011 A1
20120014608 Watanabe Jan 2012 A1
20120068913 Bar-Zeev et al. Mar 2012 A1
20120078236 Schoepp Mar 2012 A1
20120109151 Maier-Hein et al. May 2012 A1
20120143050 Heigl Jun 2012 A1
20120155064 Waters Jun 2012 A1
20120162452 Liu Jun 2012 A1
20120182605 Hall et al. Jul 2012 A1
20120201421 Hartmann et al. Aug 2012 A1
20120216411 Wevers et al. Aug 2012 A1
20120224260 Healy et al. Sep 2012 A1
20120238609 Srivastava et al. Sep 2012 A1
20120289777 Chopra et al. Nov 2012 A1
20120306850 Balan et al. Dec 2012 A1
20120320100 Machida et al. Dec 2012 A1
20130002928 Imai Jan 2013 A1
20130009853 Hesselink et al. Jan 2013 A1
20130038632 Dillavou et al. Feb 2013 A1
20130050258 Liu et al. Feb 2013 A1
20130050833 Lewis et al. Feb 2013 A1
20130057581 Meier Mar 2013 A1
20130079829 Globerman et al. Mar 2013 A1
20130083009 Geisner et al. Apr 2013 A1
20130106833 Fun May 2013 A1
20130135734 Shafer et al. May 2013 A1
20130135738 Shafer et al. May 2013 A1
20130190602 Liao et al. Jul 2013 A1
20130195338 Xu et al. Aug 2013 A1
20130209953 Arlinsky et al. Aug 2013 A1
20130212453 Gudai et al. Aug 2013 A1
20130234914 Fujimaki Sep 2013 A1
20130234935 Griffith Sep 2013 A1
20130237811 Mihailescu et al. Sep 2013 A1
20130245461 Maier-Hein et al. Sep 2013 A1
20130249787 Morimoto Sep 2013 A1
20130249945 Kobayashi Sep 2013 A1
20130265623 Sugiyama et al. Oct 2013 A1
20130267838 Fronk et al. Oct 2013 A1
20130278631 Border et al. Oct 2013 A1
20130278635 Maggiore Oct 2013 A1
20130300637 Smits et al. Nov 2013 A1
20130300760 Sugano et al. Nov 2013 A1
20130342571 Kinnebrew et al. Dec 2013 A1
20130345718 Crawford et al. Dec 2013 A1
20140031668 Mobasser et al. Jan 2014 A1
20140049629 Siewerdsen et al. Feb 2014 A1
20140088402 Xu Mar 2014 A1
20140088990 Nawana et al. Mar 2014 A1
20140104505 Koenig Apr 2014 A1
20140105912 Noelle Apr 2014 A1
20140114173 Bar-Tal et al. Apr 2014 A1
20140142426 Razzaque et al. May 2014 A1
20140168261 Margolis et al. Jun 2014 A1
20140176661 Smurro et al. Jun 2014 A1
20140177023 Gao et al. Jun 2014 A1
20140189508 Granchi et al. Jul 2014 A1
20140198129 Liu et al. Jul 2014 A1
20140218291 Kirk Aug 2014 A1
20140240484 Kodama et al. Aug 2014 A1
20140243614 Rothberg et al. Aug 2014 A1
20140256429 Kobayashi et al. Sep 2014 A1
20140266983 Christensen Sep 2014 A1
20140268356 Bolas et al. Sep 2014 A1
20140270505 McCarthy Sep 2014 A1
20140275760 Lee et al. Sep 2014 A1
20140285404 Takano et al. Sep 2014 A1
20140285429 Simmons Sep 2014 A1
20140300632 Laor Oct 2014 A1
20140300967 Tilleman et al. Oct 2014 A1
20140301624 Barckow et al. Oct 2014 A1
20140303491 Shekhar et al. Oct 2014 A1
20140320399 Kim et al. Oct 2014 A1
20140333899 Smithwick Nov 2014 A1
20140336461 Reiter et al. Nov 2014 A1
20140340286 Machida et al. Nov 2014 A1
20140361956 Mikhailov et al. Dec 2014 A1
20140371728 Vaughn Dec 2014 A1
20150005772 Anglin et al. Jan 2015 A1
20150018672 Blumhofer et al. Jan 2015 A1
20150031985 Reddy et al. Jan 2015 A1
20150043798 Carrell et al. Feb 2015 A1
20150070347 Hofmann et al. Mar 2015 A1
20150084990 Laor Mar 2015 A1
20150150641 Daon et al. Jun 2015 A1
20150182293 Yang et al. Jul 2015 A1
20150192776 Lee et al. Jul 2015 A1
20150209119 Theodore et al. Jul 2015 A1
20150230873 Kubiak et al. Aug 2015 A1
20150230893 Huwais Aug 2015 A1
20150261922 Nawana et al. Sep 2015 A1
20150277123 Chaum et al. Oct 2015 A1
20150282735 Rossner Oct 2015 A1
20150287188 Gazit et al. Oct 2015 A1
20150287236 Winne et al. Oct 2015 A1
20150297314 Fowler et al. Oct 2015 A1
20150305828 Park et al. Oct 2015 A1
20150310668 Ellerbrock Oct 2015 A1
20150338652 Lim et al. Nov 2015 A1
20150338653 Subramaniam et al. Nov 2015 A1
20150350517 Duret et al. Dec 2015 A1
20150351863 Plassky et al. Dec 2015 A1
20150363978 Maimone et al. Dec 2015 A1
20150366620 Cameron et al. Dec 2015 A1
20160015878 Graham et al. Jan 2016 A1
20160022287 Nehls Jan 2016 A1
20160030131 Yang et al. Feb 2016 A1
20160054571 Tazbaz et al. Feb 2016 A1
20160086380 Vayser et al. Mar 2016 A1
20160103318 Du et al. Apr 2016 A1
20160125603 Tanji May 2016 A1
20160133051 Aonuma et al. May 2016 A1
20160143699 Tanji May 2016 A1
20160153004 Zhang et al. Jun 2016 A1
20160163045 Penney et al. Jun 2016 A1
20160175064 Steinle et al. Jun 2016 A1
20160178910 Gudicell et al. Jun 2016 A1
20160191887 Casas Jun 2016 A1
20160223822 Harrison et al. Aug 2016 A1
20160228033 Rossner Aug 2016 A1
20160246059 Halpin et al. Aug 2016 A1
20160249989 Devam et al. Sep 2016 A1
20160256223 Haimerl et al. Sep 2016 A1
20160275684 Elenbaas et al. Sep 2016 A1
20160297315 Gonzalez et al. Oct 2016 A1
20160302870 Wilkinson et al. Oct 2016 A1
20160324580 Esterberg Nov 2016 A1
20160324583 Kheradpir et al. Nov 2016 A1
20160339337 Ellsworth et al. Nov 2016 A1
20170014119 Capote et al. Jan 2017 A1
20170024634 Miao et al. Jan 2017 A1
20170027650 Merck et al. Feb 2017 A1
20170031163 Gao et al. Feb 2017 A1
20170031179 Guillot et al. Feb 2017 A1
20170045742 Greenhalgh et al. Feb 2017 A1
20170065364 Schuh et al. Mar 2017 A1
20170068119 Antaki et al. Mar 2017 A1
20170076501 Jagga et al. Mar 2017 A1
20170086941 Marti et al. Mar 2017 A1
20170112586 Dhupar Apr 2017 A1
20170164919 Lavallee et al. Jun 2017 A1
20170164920 Lavallee et al. Jun 2017 A1
20170178375 Benishti et al. Jun 2017 A1
20170220224 Kodali et al. Aug 2017 A1
20170239015 Sela et al. Aug 2017 A1
20170245944 Crawford et al. Aug 2017 A1
20170251900 Hansen et al. Sep 2017 A1
20170252109 Yang et al. Sep 2017 A1
20170258526 Lang Sep 2017 A1
20170281283 Siegler et al. Oct 2017 A1
20170312032 Amanatullah et al. Nov 2017 A1
20170322950 Han et al. Nov 2017 A1
20170348055 Salcedo et al. Dec 2017 A1
20170348061 Joshi et al. Dec 2017 A1
20170366773 Kiraly et al. Dec 2017 A1
20170367766 Mahfouz Dec 2017 A1
20170367771 Tako et al. Dec 2017 A1
20170372477 Penney et al. Dec 2017 A1
20180003981 Urey Jan 2018 A1
20180018791 Guoyi Jan 2018 A1
20180021597 Berlinger et al. Jan 2018 A1
20180028266 Barnes et al. Feb 2018 A1
20180036884 Chen et al. Feb 2018 A1
20180049622 Ryan et al. Feb 2018 A1
20180055579 Daon et al. Mar 2018 A1
20180071029 Srimohanarajah et al. Mar 2018 A1
20180078316 Schaewe et al. Mar 2018 A1
20180082480 White et al. Mar 2018 A1
20180092667 Heigl et al. Apr 2018 A1
20180092698 Chopra et al. Apr 2018 A1
20180092699 Finley Apr 2018 A1
20180116732 Lin et al. May 2018 A1
20180116741 Garcia et al. May 2018 A1
20180117150 O'Dwyer et al. May 2018 A1
20180120106 Sato May 2018 A1
20180133871 Farmer May 2018 A1
20180153626 Yang et al. Jun 2018 A1
20180182150 Benishti et al. Jun 2018 A1
20180185100 Weinstein et al. Jul 2018 A1
20180185113 Gregerson et al. Jul 2018 A1
20180193097 McLachlin et al. Jul 2018 A1
20180200002 Kostrzewski et al. Jul 2018 A1
20180247128 Alvi et al. Aug 2018 A1
20180262743 Casas Sep 2018 A1
20180303558 Thomas Oct 2018 A1
20180311011 Van et al. Nov 2018 A1
20180317803 Ben-Yishai et al. Nov 2018 A1
20180318035 McLachlin et al. Nov 2018 A1
20180368898 Divincenzo et al. Dec 2018 A1
20190000372 Gullotti et al. Jan 2019 A1
20190000564 Navab et al. Jan 2019 A1
20190015163 Abhari et al. Jan 2019 A1
20190018235 Ouderkirk et al. Jan 2019 A1
20190038362 Nash et al. Feb 2019 A1
20190038365 Soper et al. Feb 2019 A1
20190043238 Benishti et al. Feb 2019 A1
20190043392 Abele Feb 2019 A1
20190046272 Zoabi et al. Feb 2019 A1
20190046276 Inglese et al. Feb 2019 A1
20190053851 Siemionow et al. Feb 2019 A1
20190069971 Tripathi et al. Mar 2019 A1
20190080515 Geri et al. Mar 2019 A1
20190105116 Johnson et al. Apr 2019 A1
20190130792 Rios et al. May 2019 A1
20190142519 Siemionow et al. May 2019 A1
20190144443 Jackson et al. May 2019 A1
20190175228 Elimelech et al. Jun 2019 A1
20190192230 Siemionow et al. Jun 2019 A1
20190200894 Jung et al. Jul 2019 A1
20190201106 Siemionow et al. Jul 2019 A1
20190205606 Zhou et al. Jul 2019 A1
20190216537 Eltorai et al. Jul 2019 A1
20190251692 Schmidt-Richberg et al. Aug 2019 A1
20190251694 Han et al. Aug 2019 A1
20190254753 Johnson et al. Aug 2019 A1
20190273916 Benishti et al. Sep 2019 A1
20190310481 Blum et al. Oct 2019 A1
20190333480 Lang Oct 2019 A1
20190369660 Wen et al. Dec 2019 A1
20190369717 Frielinghaus et al. Dec 2019 A1
20190378276 Flossmann et al. Dec 2019 A1
20190387351 Lyren et al. Dec 2019 A1
20200015895 Frielinghaus et al. Jan 2020 A1
20200019364 Pond Jan 2020 A1
20200020249 Jarc et al. Jan 2020 A1
20200038112 Amanatullah et al. Feb 2020 A1
20200043160 Mizukura et al. Feb 2020 A1
20200078100 Weinstein et al. Mar 2020 A1
20200085511 Oezbek et al. Mar 2020 A1
20200088997 Lee et al. Mar 2020 A1
20200100847 Siegler et al. Apr 2020 A1
20200117025 Sauer Apr 2020 A1
20200129058 Li et al. Apr 2020 A1
20200129136 Harding et al. Apr 2020 A1
20200129262 Verard et al. Apr 2020 A1
20200129264 Oativia et al. Apr 2020 A1
20200133029 Yonezawa Apr 2020 A1
20200138518 Lang May 2020 A1
20200138618 Roszkowiak et al. May 2020 A1
20200143594 Lal et al. May 2020 A1
20200146546 Chene et al. May 2020 A1
20200151507 Siemionow et al. May 2020 A1
20200156259 Ruiz et al. May 2020 A1
20200159313 Gibby et al. May 2020 A1
20200163723 Wolf et al. May 2020 A1
20200163739 Messinger et al. May 2020 A1
20200178916 Lalys et al. Jun 2020 A1
20200184638 Meglan et al. Jun 2020 A1
20200186786 Gibby et al. Jun 2020 A1
20200188028 Feiner et al. Jun 2020 A1
20200188034 Lequette et al. Jun 2020 A1
20200201082 Carabin Jun 2020 A1
20200229877 Siemionow et al. Jul 2020 A1
20200237256 Farshad et al. Jul 2020 A1
20200237459 Racheli et al. Jul 2020 A1
20200237880 Kent et al. Jul 2020 A1
20200242280 Pavloff et al. Jul 2020 A1
20200246074 Lang Aug 2020 A1
20200246081 Johnson et al. Aug 2020 A1
20200264451 Blum et al. Aug 2020 A1
20200265273 Wei et al. Aug 2020 A1
20200275988 Johnson et al. Sep 2020 A1
20200281554 Trini et al. Sep 2020 A1
20200288075 Bonin et al. Sep 2020 A1
20200294233 Merlet Sep 2020 A1
20200297427 Cameron et al. Sep 2020 A1
20200305980 Lang Oct 2020 A1
20200315734 El Amm Oct 2020 A1
20200321099 Holladay et al. Oct 2020 A1
20200323460 Busza et al. Oct 2020 A1
20200323609 Johnson et al. Oct 2020 A1
20200327721 Siemionow et al. Oct 2020 A1
20200330179 Ton Oct 2020 A1
20200337780 Winkler Oct 2020 A1
20200341283 McCracken et al. Oct 2020 A1
20200352655 Freese Nov 2020 A1
20200355927 Marcellin-Dibon et al. Nov 2020 A1
20200360091 Murray et al. Nov 2020 A1
20200375666 Stephen Dec 2020 A1
20200377493 Heiser et al. Dec 2020 A1
20200377956 Vogelstein et al. Dec 2020 A1
20200388075 Kazanzides et al. Dec 2020 A1
20200389425 Bhatia et al. Dec 2020 A1
20200390502 Holthuizen et al. Dec 2020 A1
20200390503 Casas et al. Dec 2020 A1
20200402647 Domracheva et al. Dec 2020 A1
20200409306 Gelman et al. Dec 2020 A1
20200410687 Siemionow et al. Dec 2020 A1
20200413031 Khani et al. Dec 2020 A1
20210004956 Book et al. Jan 2021 A1
20210009339 Morrison et al. Jan 2021 A1
20210015560 Boddington et al. Jan 2021 A1
20210015583 Avisar et al. Jan 2021 A1
20210022599 Freeman et al. Jan 2021 A1
20210022808 Lang Jan 2021 A1
20210022811 Mahfouz Jan 2021 A1
20210022828 Elimelech et al. Jan 2021 A1
20210029804 Chang Jan 2021 A1
20210030374 Takahashi et al. Feb 2021 A1
20210030511 Wolf et al. Feb 2021 A1
20210038339 Yu et al. Feb 2021 A1
20210049825 Wheelwright et al. Feb 2021 A1
20210052348 Stifter et al. Feb 2021 A1
20210056687 Hibbard et al. Feb 2021 A1
20210065911 Goel et al. Mar 2021 A1
20210077195 Saeidi et al. Mar 2021 A1
20210077210 Itkowitz et al. Mar 2021 A1
20210080751 Lindsey et al. Mar 2021 A1
20210090344 Geri et al. Mar 2021 A1
20210093391 Poltaretskyi et al. Apr 2021 A1
20210093392 Poltaretskyi et al. Apr 2021 A1
20210093400 Quaid et al. Apr 2021 A1
20210093417 Liu Apr 2021 A1
20210104055 Ni et al. Apr 2021 A1
20210107923 Jackson et al. Apr 2021 A1
20210109349 Schneider et al. Apr 2021 A1
20210109373 Loo et al. Apr 2021 A1
20210110517 Flohr et al. Apr 2021 A1
20210113269 Vilsmeier et al. Apr 2021 A1
20210113293 Silva et al. Apr 2021 A9
20210121238 Palushi et al. Apr 2021 A1
20210137634 Lang May 2021 A1
20210141887 Kim et al. May 2021 A1
20210150702 Claessen et al. May 2021 A1
20210157544 Denton May 2021 A1
20210160472 Casas May 2021 A1
20210161614 Elimelech et al. Jun 2021 A1
20210162287 Xing et al. Jun 2021 A1
20210165207 Peyman Jun 2021 A1
20210169504 Brown Jun 2021 A1
20210169578 Calloway et al. Jun 2021 A1
20210169581 Calloway et al. Jun 2021 A1
20210169605 Calloway et al. Jun 2021 A1
20210186647 Elimelech et al. Jun 2021 A1
20210196404 Wang Jul 2021 A1
20210211640 Bristol et al. Jul 2021 A1
20210223577 Zhang et al. Jul 2021 A1
20210225006 Grady et al. Jul 2021 A1
20210227791 De et al. Jul 2021 A1
20210231301 Hikmet et al. Jul 2021 A1
20210235061 Hegyi Jul 2021 A1
20210248822 Choi et al. Aug 2021 A1
20210274281 Zhang et al. Sep 2021 A1
20210278675 Klug et al. Sep 2021 A1
20210282887 Wiggermann Sep 2021 A1
20210290046 Nazareth et al. Sep 2021 A1
20210290336 Wang Sep 2021 A1
20210290394 Mahfouz Sep 2021 A1
20210295108 Bar Sep 2021 A1
20210295512 Knoplioch et al. Sep 2021 A1
20210298795 Bowling et al. Sep 2021 A1
20210298835 Wang Sep 2021 A1
20210306599 Pierce Sep 2021 A1
20210311322 Belanger et al. Oct 2021 A1
20210314502 Liu Oct 2021 A1
20210315636 Akbarian et al. Oct 2021 A1
20210315662 Freeman et al. Oct 2021 A1
20210325684 Ninan et al. Oct 2021 A1
20210332447 Lubelski et al. Oct 2021 A1
20210333561 Oh et al. Oct 2021 A1
20210341739 Cakmakci et al. Nov 2021 A1
20210341740 Cakmakci et al. Nov 2021 A1
20210346115 Dulin et al. Nov 2021 A1
20210349677 Baldev et al. Nov 2021 A1
20210364802 Uchiyama et al. Nov 2021 A1
20210369226 Siemionow et al. Dec 2021 A1
20210371413 Thurston et al. Dec 2021 A1
20210373333 Moon Dec 2021 A1
20210373344 Loyola et al. Dec 2021 A1
20210378757 Bay et al. Dec 2021 A1
20210382310 Freeman et al. Dec 2021 A1
20210386482 Gera et al. Dec 2021 A1
20210389590 Freeman et al. Dec 2021 A1
20210400247 Casas Dec 2021 A1
20210401533 Im Dec 2021 A1
20210402255 Fung Dec 2021 A1
20210405369 King Dec 2021 A1
20220003992 Ahn Jan 2022 A1
20220007006 Healy et al. Jan 2022 A1
20220008135 Frielinghaus et al. Jan 2022 A1
20220038675 Hegyi Feb 2022 A1
20220039873 Harris Feb 2022 A1
20220051484 Jones et al. Feb 2022 A1
20220054199 Sivaprakasam et al. Feb 2022 A1
20220061921 Crawford et al. Mar 2022 A1
20220071712 Wolf et al. Mar 2022 A1
20220079675 Lang Mar 2022 A1
20220087746 Lang Mar 2022 A1
20220113810 Isaacs et al. Apr 2022 A1
20220117669 Nikou et al. Apr 2022 A1
20220121041 Hakim Apr 2022 A1
20220133484 Lang May 2022 A1
20220142730 Wolf et al. May 2022 A1
20220155861 Myung et al. May 2022 A1
20220159227 Quiles Casas May 2022 A1
20220179209 Cherukuri Jun 2022 A1
20220192776 Gibby et al. Jun 2022 A1
20220193453 Miyazaki et al. Jun 2022 A1
20220201274 Achilefu et al. Jun 2022 A1
20220245400 Siemionow et al. Aug 2022 A1
20220245821 Ouzounis Aug 2022 A1
20220257206 Hartley et al. Aug 2022 A1
20220269077 Adema et al. Aug 2022 A1
20220270263 Junio Aug 2022 A1
20220287676 Steines et al. Sep 2022 A1
20220292786 Pelzl et al. Sep 2022 A1
20220295033 Quiles Casas Sep 2022 A1
20220296315 Sokhanvar et al. Sep 2022 A1
20220304768 Elimelech et al. Sep 2022 A1
20220351385 Finley et al. Nov 2022 A1
20220353487 Hegyi Nov 2022 A1
20220358759 Cork et al. Nov 2022 A1
20220370152 Lavallee et al. Nov 2022 A1
20220387130 Spaas et al. Dec 2022 A1
20220392085 Finley et al. Dec 2022 A1
20220397750 Zhou et al. Dec 2022 A1
20220398752 Yoon et al. Dec 2022 A1
20220398755 Herrmann Dec 2022 A1
20220405935 Flossmann et al. Dec 2022 A1
20230004013 McCracken et al. Jan 2023 A1
20230009793 Gera et al. Jan 2023 A1
20230025480 Kemp et al. Jan 2023 A1
20230027801 Qian et al. Jan 2023 A1
20230032731 Hrndler et al. Feb 2023 A1
20230034189 Gera et al. Feb 2023 A1
20230050636 Yanof et al. Feb 2023 A1
20230053120 Jamali et al. Feb 2023 A1
20230073041 Samadani et al. Mar 2023 A1
20230085387 Jones et al. Mar 2023 A1
20230087783 Dulin et al. Mar 2023 A1
20230100078 Toporek et al. Mar 2023 A1
20230123621 Joshi et al. Apr 2023 A1
20230126207 Wang Apr 2023 A1
20230129056 Hemingway et al. Apr 2023 A1
20230131515 Oezbek et al. Apr 2023 A1
20230149083 Lin et al. May 2023 A1
20230162493 Worrell et al. May 2023 A1
20230165640 Dulin et al. Jun 2023 A1
20230169659 Chen et al. Jun 2023 A1
20230196582 Grady et al. Jun 2023 A1
20230200917 Calloway et al. Jun 2023 A1
20230236426 Manly et al. Jul 2023 A1
20230236427 Jiannyuh Jul 2023 A1
20230260142 Chatterjee et al. Aug 2023 A1
20230290037 Tasse et al. Sep 2023 A1
20230295302 Bhagavatheeswaran et al. Sep 2023 A1
20230306590 Jazdzyk et al. Sep 2023 A1
20230316550 Hiasa Oct 2023 A1
20230326011 Cutforth et al. Oct 2023 A1
20230326027 Wahrenberg Oct 2023 A1
20230329801 Elimelech et al. Oct 2023 A1
20230334664 Lu et al. Oct 2023 A1
20230335261 Reicher et al. Oct 2023 A1
20230359043 Russell Nov 2023 A1
20230363832 Mosadegh et al. Nov 2023 A1
20230371984 Leuthardt et al. Nov 2023 A1
20230372053 Elimelech et al. Nov 2023 A1
20230372054 Elimelech et al. Nov 2023 A1
20230377171 Hasler et al. Nov 2023 A1
20230377175 Seok Nov 2023 A1
20230379448 Benishti et al. Nov 2023 A1
20230379449 Benishti et al. Nov 2023 A1
20230386022 Tan et al. Nov 2023 A1
20230386067 De et al. Nov 2023 A1
20230386153 Rybnikov et al. Nov 2023 A1
20230394791 Wang et al. Dec 2023 A1
20230397349 Capelli et al. Dec 2023 A1
20230397957 Crawford et al. Dec 2023 A1
20230410445 Elimelech et al. Dec 2023 A1
20230419496 Wuelker et al. Dec 2023 A1
20230420114 Scholler et al. Dec 2023 A1
20240008935 Wolf et al. Jan 2024 A1
20240016549 Johnson et al. Jan 2024 A1
20240016572 Elimelech et al. Jan 2024 A1
20240020831 Johnson et al. Jan 2024 A1
20240020840 Johnson et al. Jan 2024 A1
20240020862 Johnson et al. Jan 2024 A1
20240022704 Benishti et al. Jan 2024 A1
20240023946 Wolf et al. Jan 2024 A1
20240041530 Lang Feb 2024 A1
20240041558 Siewerdsen et al. Feb 2024 A1
20240045491 Sourov Feb 2024 A1
20240058064 Weiser et al. Feb 2024 A1
20240062387 Frantz et al. Feb 2024 A1
20240103271 Zare Seisan Mar 2024 A1
20240103282 Law et al. Mar 2024 A1
20240111163 Law et al. Apr 2024 A1
20240122560 Junio et al. Apr 2024 A1
20240126087 Gera et al. Apr 2024 A1
20240127559 Rybnikov et al. Apr 2024 A1
20240127578 Hiasa Apr 2024 A1
20240129451 Healy et al. Apr 2024 A1
20240130826 Elimelech et al. Apr 2024 A1
20240134206 Gera et al. Apr 2024 A1
20240144497 Cvetko et al. May 2024 A1
20240156532 Weiman et al. May 2024 A1
20240177445 Galeotti et al. May 2024 A1
20240177458 Zhang et al. May 2024 A1
20240180634 Mikus Jun 2024 A1
20240184119 Lee et al. Jun 2024 A1
20240185509 Kovler et al. Jun 2024 A1
20240202926 Crawford et al. Jun 2024 A1
20240202927 Haslam et al. Jun 2024 A1
20240212111 Genghi et al. Jun 2024 A1
20240233131 Westerhoff et al. Jul 2024 A1
20240245463 Vilsmeier et al. Jul 2024 A1
20240245474 Weiman et al. Jul 2024 A1
20240248530 Gibby et al. Jul 2024 A1
20240252252 Lang Aug 2024 A1
20240261036 Finley et al. Aug 2024 A1
20240261058 Gera et al. Aug 2024 A1
20240265645 Papar Aug 2024 A1
20240266033 Freeman et al. Aug 2024 A1
20240268922 Calloway et al. Aug 2024 A1
20240273740 Gibby et al. Aug 2024 A1
20240281979 Schrempf et al. Aug 2024 A1
20240296527 Nett et al. Sep 2024 A1
20240303832 Chen et al. Sep 2024 A1
20240307101 Gera et al. Sep 2024 A1
20240312012 Li et al. Sep 2024 A1
20240341861 Wolf et al. Oct 2024 A1
20240341910 Wolf et al. Oct 2024 A1
20240341911 Elimelech et al. Oct 2024 A1
Foreign Referenced Citations (187)
Number Date Country
3022448 Feb 2018 CA
3034314 Feb 2018 CA
101379412 Mar 2009 CN
101379412 Mar 2009 CN
103106348 May 2013 CN
111915696 Nov 2020 CN
112489047 Mar 2021 CN
112489047 Mar 2021 CN
202004011567 Nov 2004 DE
102004011567 Sep 2005 DE
102014008153 Oct 2014 DE
202022103168 Jun 2022 DE
0933096 Aug 1999 EP
1640750 Mar 2006 EP
1757974 Feb 2007 EP
2119397 Nov 2009 EP
2134847 Dec 2009 EP
2557998 Feb 2013 EP
2823463 Jan 2015 EP
2868277 May 2015 EP
2134847 Jun 2015 EP
2891966 Jul 2015 EP
2963616 Jan 2016 EP
3028258 Jun 2016 EP
3034607 Jun 2016 EP
3037038 Jun 2016 EP
3069318 Sep 2016 EP
3076660 Oct 2016 EP
2891966 Jan 2017 EP
3121789 Jan 2017 EP
3123970 Feb 2017 EP
2654749 May 2017 EP
3175815 Jun 2017 EP
3216416 Sep 2017 EP
2032039 Oct 2017 EP
3224376 Oct 2017 EP
3247297 Nov 2017 EP
3256213 Dec 2017 EP
3306567 Apr 2018 EP
3320874 May 2018 EP
2030193 Jul 2018 EP
2225723 Feb 2019 EP
2619622 Feb 2019 EP
3034607 Mar 2019 EP
2892558 Apr 2019 EP
3494903 Jun 2019 EP
2635299 Jul 2019 EP
3505050 Jul 2019 EP
3224376 Aug 2019 EP
2875149 Dec 2019 EP
3593227 Jan 2020 EP
3634294 Apr 2020 EP
3206583 Sep 2020 EP
3711700 Sep 2020 EP
2625845 Mar 2021 EP
3789965 Mar 2021 EP
3076660 Apr 2021 EP
3858280 Aug 2021 EP
3593227 Sep 2021 EP
3913423 Nov 2021 EP
3789965 Dec 2021 EP
3634294 Jan 2022 EP
3952331 Feb 2022 EP
3960235 Mar 2022 EP
3635683 Jul 2022 EP
3602492 Nov 2022 EP
4173590 May 2023 EP
3533031 Aug 2023 EP
4252695 Oct 2023 EP
3195257 Nov 2023 EP
3405909 Nov 2023 EP
4270313 Nov 2023 EP
4287120 Dec 2023 EP
3488381 Feb 2024 EP
3834768 Feb 2024 EP
3903714 Feb 2024 EP
4336450 Mar 2024 EP
3814984 Apr 2024 EP
4115389 Apr 2024 EP
3752981 May 2024 EP
4375948 May 2024 EP
4383203 Jun 2024 EP
2507314 Apr 2014 GB
262864 Mar 2019 IL
2004-237092 Aug 2004 JP
2008-507361 Mar 2008 JP
2009-514571 Apr 2009 JP
2021-525186 Sep 2021 JP
10-2014-0120155 Oct 2014 KR
20140120155 Oct 2014 KR
0334705 Apr 2003 WO
03034705 Apr 2003 WO
2006002559 Jan 2006 WO
2007051304 May 2007 WO
2007115826 Oct 2007 WO
2008103383 Aug 2008 WO
2010067267 Jun 2010 WO
2010074747 Jul 2010 WO
2012061537 May 2012 WO
2012101286 Aug 2012 WO
2013112554 Aug 2013 WO
2014014498 Jan 2014 WO
2014024188 Feb 2014 WO
2014037953 Mar 2014 WO
2014113455 Jul 2014 WO
2014125789 Aug 2014 WO
2014167563 Oct 2014 WO
2014174067 Oct 2014 WO
2015058816 Apr 2015 WO
WO2015061752 Apr 2015 WO
WO2015109145 Jul 2015 WO
2016151506 Sep 2016 WO
2018052966 Mar 2018 WO
2018073452 Apr 2018 WO
WO2018200767 Apr 2018 WO
2018206086 Nov 2018 WO
2019083431 May 2019 WO
2019135209 Jul 2019 WO
2019161477 Aug 2019 WO
2019195926 Oct 2019 WO
2019211741 Nov 2019 WO
WO2019210353 Nov 2019 WO
2020109903 Jun 2020 WO
2020109904 Jun 2020 WO
2021019369 Feb 2021 WO
WO2021017019 Feb 2021 WO
WO2021023574 Feb 2021 WO
WO2021046455 Mar 2021 WO
WO2021048158 Mar 2021 WO
WO2021021979 Apr 2021 WO
WO2021061459 Apr 2021 WO
WO2021062375 Apr 2021 WO
WO2021073743 Apr 2021 WO
WO2021087439 May 2021 WO
WO2021091980 May 2021 WO
WO2021112918 Jun 2021 WO
2021130564 Jul 2021 WO
WO2021137752 Jul 2021 WO
WO2021141887 Jul 2021 WO
WO2021145584 Jul 2021 WO
WO2021154076 Aug 2021 WO
2021188757 Sep 2021 WO
2021255627 Dec 2021 WO
WO2021183318 Dec 2021 WO
WO2021257897 Dec 2021 WO
WO2021258078 Dec 2021 WO
WO2022009233 Jan 2022 WO
2022053923 Mar 2022 WO
2022079565 Apr 2022 WO
2023003952 Jan 2023 WO
2023281395 Jan 2023 WO
2023011924 Feb 2023 WO
2023007418 Feb 2023 WO
2023021448 Feb 2023 WO
2023021450 Feb 2023 WO
2023021451 Feb 2023 WO
2023026229 Mar 2023 WO
2023047355 Mar 2023 WO
2023072887 May 2023 WO
2023088986 May 2023 WO
2023158878 Aug 2023 WO
2023159104 Aug 2023 WO
2023161848 Aug 2023 WO
2023163933 Aug 2023 WO
2023175244 Sep 2023 WO
2023186996 Oct 2023 WO
2023202909 Oct 2023 WO
2023205212 Oct 2023 WO
2023205896 Nov 2023 WO
2023209014 Nov 2023 WO
2023229415 Nov 2023 WO
2023232492 Dec 2023 WO
2023240912 Dec 2023 WO
2024001140 Jan 2024 WO
2024002620 Jan 2024 WO
2024013642 Jan 2024 WO
2024018368 Jan 2024 WO
2024046760 Mar 2024 WO
2024052136 Mar 2024 WO
2024077077 Apr 2024 WO
2024121060 Jun 2024 WO
2024132609 Jun 2024 WO
2024145341 Jul 2024 WO
2024160896 Aug 2024 WO
2024165508 Aug 2024 WO
2024173251 Aug 2024 WO
2024186811 Sep 2024 WO
Non-Patent Literature Citations (33)
Entry
US 11,395,705 B2, 09/2022, Lang (withdrawn)
Van Ooijen et al., “Noninvasive Coronary Imaging Using Electron Beam CT: Surface Rendering Versus vol. Rendering,” Computers in Radiology, AJR, vol. 180, pp. 223-226, Jan. 2003.
Zhang et al., “Medical Volume Rendering Techniques,” Independent Research, Spring 2014, arXiv:1802.07710v1, pp. 1-33, Feb. 21, 2018.
Romero, “Volume Ray Casting Techniques and Applications Using General Purpose Computations on Graphics Processing Units,” Thesis/Dissertation Collections, Rochester Institute of Technology, RIT Scholar Works, pp. 1-140, Jun. 2009.
Liberadzki et al., “Structured-Light-Based System for Shape Measurement of the Human Body in Motion,” Sensors, vol. 18, pp. 1-19, year 2018.
WEBSTER (ed.), “Structured Light Techniques and Applications,” Wiley Encyclopedia of Electrical and Electronics Engineering, pp. 1-24, year 2016.
Lorensen et al., “Marching Cubes: A High Resolution 3D Surface Construction Algorithm,” ACM SIGGRAPH '87, Computer Graphics, vol. 21, No. 4, pp. 163-169, Jul. 1987.
Wikipedia, “Marching Cubes,” pp. 1-4, last edited Sep. 4, 2021.
Milletari et al., “V-Net: fully Convolutional Neural Networks for Volumetric Medical Image Segmentation,” arXiv:1606.04797v1, pp. 1-11, Jun. 15, 2016.
Lumus Ltd., “DK-32 See-through Wearable Display Development Kit” , Rehovot, Israel, pp. 1-2, Dec. 24, 2013.
Liao et al., ‘3-D Augmented Reality for MRI-Guided Surgery Using Integral Videography Autostereoscopic Image Overlay’, IEEE Transactions on Biomedical Engineering, vol. 57, No. 6, pp. 1476-1486, Feb. 17, 2010.
Hainich et al., “Near-Eye displays”, Chapter 10 of Displays: Fundamentals and Applications, CRC press, pp. 439-504, Jul. 5, 2011.
Brainlab—Image Registration Options Enhanced Visualization Leveraging More Data , pp. 1-4, Feb. 2019.
Liu et al., “Marker orientation in fiducial registration”, Medical Imaging 2003: Image Processing, Proceedings of SPIE vol. 5032, pp. 1176-1185, 2003.
Fingas, “Fraunhofer iPad app guides liver surgery through augmented reality”, pp. 1-6, Aug. 22, 2013.
Sagitov et al., “Comparing Fiducial Marker Systems in the Presence of Occlusion”, International Conference on Mechanical, System and Control Engineering (ICMSC), pp. 1-6, 2017.
International Application # PCT/IB2021/055242 Search Report dated Oct. 7, 2021.
Mitrasinovic et al., “Clinical and surgical applications of smart glasses”, pp. 381-401, Technology and Health Care, issue 23, year 2015.
Martin-Gonzalez et al., “Head-mounted virtual loupe with sight-based activation for surgical applications”, IEEE symposium on mixed and augmented reality, pp. 207-208, Oct. 19-22, 2009.
Figl et al., “A fully automated calibration method for an optical see-through head-mounted operating microscope with variable zoom and focus”, pp. 1492-1499, IEEE transactions on medical imaging, vol. 24, No. 11, Nov. 2005.
Medithinq Co. Ltd., “Metascope: world's first wearable scope”, pp. 1-7, Jan. 2023.
Martin-Gonzalez et al., “Sight-based magnification system for surgical applications”, pp. 26-30, Conference proceedings of Bildverarbeitung für die Medizin, year 2010.
Burstrom et al., “Frameless patient tracking with adhesive optical skin markers for augmented reality surgical navigation in spine surgery”, SPINE, vol. 45, No. 22, pp. 1598-1604, year 2020.
Suenaga et al., “Vision-based markerless registration using stereo vision and an augmented reality surgical navigation system: a pilot study”, BMC Medical Imaging, pp. 1-11, year 2015.
Mayfield Clinic, “Spinal Fusion: Lateral Lumbar Interbody Fusion (LLIF)”, pp. 1-6, Jan. 2021.
Qian et al., “AR-Loupe: Magnified Augmented Reality by Combining an Optical See-Through Head-Mounted Display and a Loupe”, pp. 2550-2562, IEEE Transactions on Visualization and Computer Graphics, vol. 28, No. 7, Jul. 2022.
Kazanzides et al., “Systems and Methods for Augmented Reality Magnifying Loupe”, case ID 15944, pp. 1-2, Nov. 26, 2020.
16 Augmented Reality Glasses of 2021 (with Features), in Back to News, Dated May 6, 2022, accessed at https://web.archive. org/web/20221127195438/https://circuitstream.com/blog/16-augmented-reality-glasses-of-2021-with-features-breakdowns/.
Everysight, Installing your RX Adaptor, accessed Mar. 13, 2024 at https://support.everysight.com/hc/en-us/articles/115000984571-Installing-your-RX-Adaptor.
Everysight, Raptor User Manual, copyright 2017, in 46 pages.
Frames Direct, InSpatialRx Prescription Insert, Prescription Insert for Magic Leap 1, accessed Mar. 8, 2024 at https://www.framesdirect.com/inspatialrx-prescription-insert. html.
Reddit, Notice on Prescription Lenses for Nreal Glasses, accessed Mar. 13, 2024 at https://www.reddit.com/r/nreal/comments/x1fte5/notice_on_prescription_lenses_for_nreal_glasses/.
Vuzix Blades, Prescription Lens Installation Guide, copyright 2020.
Related Publications (1)
Number Date Country
20230329799 A1 Oct 2023 US
Continuation in Parts (1)
Number Date Country
Parent 16901026 Jun 2020 US
Child 18008980 US