Endoscopic ultrasonography simulation

Information

  • Patent Grant
  • 9501955
  • Patent Number
    9,501,955
  • Date Filed
    Monday, May 20, 2002
    22 years ago
  • Date Issued
    Tuesday, November 22, 2016
    8 years ago
Abstract
A method and a system to simulate ultrasound images and hence for simulating the evaluation of a subject with such images as a diagnostic tool. For the purpose of simulating ultrasound images, these images may be considered to be planar slices of a given geometry. These slices are generated from the surface geometry of the portion of the subject which would receive the ultrasound waves if the ultrasound diagnostic procedure was actually be performed. Simulating such ultrasound images may therefore be performed by determining such images from surface models, and then rendering those ultrasound images at the desired position. The rendering of such simulated ultrasound images also optionally and preferably includes the simulation of the “grainy” quality of such images.
Description
RELATED PATENT APPLICATIONS

This application is a National Phase Application of PCT/IL02/00393, International Filing Date 20 May 2002 which claims priority from Israel Patent Application No. 143255, filed 20 May 2001.


FIELD OF THE INVENTION

The present invention is of a method for simulating ultrasound images, and in particular, for such a method for simulating ultrasound images which are produced during a medical diagnostic procedure.


BACKGROUND OF THE INVENTION

ERCP (Endoscopic Retrograde Cholangio Pancreatography) is a minimally invasive procedure which allows the doctor to perform necessary treatments such as enlarging a bile duct opening, removing gallstones lodged in the bile duct, inserting a stent (drain) in the duct or taking a biopsy specimen.


A flexible fiber optic tube, called a duodenoscope, is passed through the mouth, esophagus and stomach into the duodenum, which is the first part of the small intestine. The duodenoscope is a thin, flexible tube with a tiny video camera and light at one side of the tip. The papilla, which is an opening where the bile and pancreatic ducts empty into the duodenum, is visually identified. A small plastic tube, or cannula, is passed through the duodenoscope into the papilla. X-ray dye is injected through the cannula into the ducts. X-rays are then taken to study the ducts. Any necessary treatments can be performed at this time.


This treatment scheme is somewhat similar to gastro-endoscopy, in which an endoscope is inserted through the rectum in order to examine the large colon, for example. In both cases, students are taught to perform such procedures according to the traditional model for medical education, in which students observe and assist more experienced physicians. Unfortunately, such observation alone cannot provide the necessary training for such complicated medical procedures. Students may also perform procedures on animals and human cadavers, neither of which replicates the visual and tactile sensations of a live human patient. Thus, traditional medical training is not adequate for modern technologically complex medical procedures.


In an attempt to provide more realistic medical training for such procedures, simulation devices have been developed which attempt to replicate the tactile sensations and/or visual feedback for these procedures, in order to provide improved medical training without endangering human patients. An example of such a simulation device is disclosed in U.S. Pat. No. 5,403,191, in which the disclosed device is a box containing simulated human organs. Various surgical laparoscopic procedures can be performed on the simulated organs. Visual feedback is provided by a system of mirrors. However, the system of both visual and tactile feedback is primitive in this device, and does not provide a true representation of the visual and tactile sensations which would accompany such surgical procedures in a human patient. Furthermore, the box itself is not a realistic representation of the three-dimensional structure of a human patient. Thus, the disclosed device is lacking in many important aspects and fails to meet the needs of a medical simulation device.


Attempts to provide a more realistic experience from a medical simulation device are disclosed in PCT Patent Application Nos. WO 96/16389 and WO 95/02233. Both of these applications disclose a device for providing a simulation of the surgical procedure of laparoscopy. Both devices include a mannequin in the shape of a human torso, with various points at which simulated surgical instruments are placed. However, the devices are limited in that the positions of the simulated surgical instruments are predetermined, which is not a realistic scenario. Furthermore, the visual feedback is based upon a stream of video images taken from actual surgical procedures. However, such simple rendering of video images would result in inaccurate or unrealistic images as portions of the video data would need to be removed for greater processing speed. Alternatively, the video processing would consume such massive amounts of computational time and resources that the entire system would fail to respond in a realistic time period to the actions of the student. At the very minimum, a dedicated graphics workstation would be required, rather than a personal computer (PC). Thus, neither reference teaches or discloses adequate visual processing for real time visual feedback of the simulated medical procedure.


Similarly, U.S. Pat. No. 4,907,973 discloses a device for simulating the medical procedure of flexible gastro-endoscopy. The disclosed device also suffers from the deficiencies of the above-referenced prior art devices, in that the visual feedback system is based upon rendering of video data taken from actual duodenoscopic procedures. As noted previously, displaying such data would either require massive computational resources, or else would simply require too much time for a realistic visual feedback response. Thus, the disclosed device also suffers from the deficiencies of the prior art.


A more useful and efficient medical simulation device for minimally invasive therapeutic procedures such as endoscopy is disclosed in PCT Application No. WO 99/38141, by the present inventors and incorporated by reference as if fully set forth herein. The disclosed medical simulation device provides real time, accurate and realistic visual feedback of general endoscopic procedures, as well as realistic tactile feedback, so that the visual and tactile systems are accurately linked for the simulation as for an actual medical procedure.


Duodenoscopy of the bilio-pancreatic system would also benefit from such realistic simulation, involving both visual and tactile feedback which are provided in an accurate manner. Bilio-pancreatic duodenoscopic procedures feature many of the same principles as gastro-endoscopy, since for both types of endoscopic procedures, an instrument is inserted into a body orifice, and must then be guided through a tubular organ without direct visual feedback. In addition, the physician performing the procedure must be able to correctly interpret both the indirect visual feedback provided through a video monitor, as well as the tactile feedback through the instrument itself. Therefore, both types of duodenoscopy require the physician to receive “hands-on” manual training for the correct performance of the procedure.


As part of the simulation of this procedure, the students should also be able to practice using the diagnostic tool of ultrasound. Ultrasonography is an optional adjunct to a large number of medical procedures, such as gastro-endoscopy and duodenoscopy. In addition, ultrasonography is itself an important medical procedure, which is used for example in order to safely visualize the fetus during pregnancy. Unfortunately, the background art does not feature suitable methods for accurately and efficiently simulating ultrasound images.


SUMMARY OF THE INVENTION

The background art does not teach or suggest a method and a system to efficiently and realistically simulate ultrasound images. Nor does the background art teach or suggest such a system and method for providing such images as part of the simulation of an overall medical procedure.


The present invention overcomes the deficiencies of the background art by providing a method and a system to simulate ultrasound images and hence for simulating the evaluation of a subject with such images as a diagnostic tool. For the purpose of simulating ultrasound images, these images may be considered to be planar slices of a given geometry. These slices are generated from the surface geometry of the portion of the subject which would receive the ultrasound waves if the ultrasound diagnostic procedure was actually being performed. Simulating such ultrasound images may therefore be performed by determining such images from surface models, and then rendering those ultrasound images at positions given by the tip of the scope. The rendering of such simulated ultrasound images also optionally and preferably includes the simulation of the “grainy” quality of such images.


According to the present invention, there is provided a method for simulating an ultrasound image of at least one anatomical feature of a subject, comprising: determining a surface geometry of the at least one anatomical feature; generating at least one planar slice of the surface geometry according to a location of a simulated source of ultrasound waves; and rendering the ultrasound image according to the at least one planar slice.


Hereinafter, the term “simulated medical procedure” refers to the simulation of the medical procedure as performed through the system and method of the present invention. Hereinafter, the term “actual medical procedure” refers to the performance of the medical procedure on an actual, living human patient with an actual medical instrument, such that the medical procedure is “real” rather than “simulated”. Hereinafter, the term “corresponding actual organ” refers to the “real” organ of a human being or other mammal which is being simulated by the simulated organ of the present invention.


Hereinafter, the term “subject” refers to the human or lower mammal upon which the method and system of the present invention are performed or operated. Hereinafter, the term “student” refers to any human using the system of the present invention, being trained according to the present invention or being taught according to the present invention including, but not limited to, students attending medical school or a university, a medical doctor, a trained pancreatic specialist or other trained medical specialist.





BRIEF DESCRIPTION OF THE DRAWINGS

The foregoing and other objects, aspects and advantages will be better understood from the following detailed description of a preferred embodiment of the invention with reference to the drawings, wherein:



FIG. 1 is a flowchart of an exemplary method of the present invention;



FIG. 2 is a flow diagram for determining the output texture for ultrasound images according to the present invention;



FIG. 3 shows exemplary ultrasound texture which is simulated according to the present invention;



FIG. 4 shows labeled exemplary organ slices, visualized through ultrasound, which are simulated according to the present invention; and



FIG. 5 shows an exemplary final ultrasound image, simulated according to the method of the present invention.



FIG. 6 is a flowchart showing an exemplary method of the present invention.





DETAILED DESCRIPTION OF PREFERRED EMBODIMENTS

The present invention is related to a method and a system for simulating ultrasound images and hence for simulating the evaluation of a subject with such images as a diagnostic tool. Preferably, but not necessarily, the ultrasound images are simulated as though they had been collected during the evaluation of a subject with gastro-endoscopy and/or duodenoscopy, although of course ultrasound images could optionally be simulated as though they had been collected during substantially any medical procedure.


For the purpose of simulating ultrasound images, these images may be considered to be planar slices of a given geometry. These slices are generated from the surface geometry of the portion of the subject which would receive the ultrasound waves if the ultrasound diagnostic procedure was actually being performed. This surface geometry may optionally be modeled as polygonal models or other surface representation, optionally and more preferably by using common 3D graphic accelerator functionality. Thus, a real-time performance to this simulation may optionally be obtained.


Simulating such ultrasound images may therefore be performed by determining such images from surface models, and then rendering those ultrasound images at positions given by the tip of the scope. Such rendering is similar to that currently performed for modeling regular endoscopy procedures, and is also known as a “movement model”. One example of such a movement model is given in PCT Application No. WO 99/38141, filed on Jan. 15, 1999, which is hereby incorporated by reference as if fully set forth herein. This reference also describes an exemplary system for simulating a medical procedure, which could optionally be used with the present invention, for example. Of course, other systems and/or devices could also optionally be used with the method of the present invention.


The present invention differs from such methods of modeling visual data for gastro-endoscopic procedures in that additional areas of the subject, including additional organs and/or organ portions, must also be rendered, since ultrasound waves penetrate beyond the wall of the organ which actually contains the ultrasound probe.


The rendering of such simulated ultrasound images also optionally and preferably includes the simulation of the “grainy” quality of such images. This quality is inherent to the method of ultrasound itself, as it involves sending sound waves to the desired area of the subject and then listening to the echoes, in order to measure the time required for the echo to return. Of course, at least a portion of the actual signal would be attenuated, since not all of the sound waves are reflected by the tissue. The intensity of the echo is also measured to provide the “echoic property” of the object, in which white spots in an image are hyper-echoic, and black spots are hypo-echoic. The graininess of the image occurs since the sound waves reflect in all directions, so minor echoes that are reflected locations other than the organ being examined also return to the probe, causing static interference with the signal being measured. Such static interference then creates the grainy image.


The principles and operation of a method and a system according to the present invention for the simulation of the medical diagnostic procedure of ultrasound, may be better understood with reference to the drawings and the accompanying description, it being understood that these drawings are given for illustrative purposes only and are not meant to be limiting.


As shown in FIG. 6, the overall method of rendering a simulated ultrasound image described in the invention is seen in flowchart form. The first step is to determine the surface geometry of an anatomical object. Next (step 2), a viewpoint above the surface geometry and z clip plane are chosen. Utilizing the above data, the software then generates a planar slice image (step 3) of the three dimensional surface in accordance with the location of a simulated source of ultrasound waves. Additional surface geometries are determined now if necessary by repeating step 1. Finally, (Step 4) the simulated ultrasound image is rendered through the processing of the planar slice image through known techniques such as bending and blurring of the image, and may be in accordance with selected textures from an actual ultrasound procedure.


As shown with regard to FIG. 1, an exemplary method of the present invention for rendering ultrasound images is described. Although the explanation is given with regard to a simulation of a linear array probe, in which the plane of the scan is parallel to the shaft of the endoscope probe, the system could easily be used to simulate other types of US (ultrasound) scans, for example a radial scan in which the plane of the scan is perpendicular to the shaft of the endoscope. One of ordinary skill in the art could easily alter the simulation process in order to accommodate these differences between scanning devices.


First, a technique called “stenciling” is used to render a slice of an object from a specific plane into a texture (which is basically a bitmap or a digital picture). Stenciling is known in the art of real time graphics, and is also used to simulate shadows and create other real-time computer graphic effects, in conjunction with a stencil buffer. This technique is performed by positioning the viewpoint of a virtual “camera” above the desired plane, in the direction of the plane normal. Positioning of a virtual camera or viewpoint for generating three-dimensional computer graphics is achieved by giving the coordinates of the viewpoint in relation to the coordinates of the object being viewed, by using a standard coordinate system (such as that of the X, Y & Z Axes for example).


The near Z clip plane is then set to fit the cutting plane. The near Z clip plane is the distance from the camera at which objects begin to be seen, such that objects closer than that distance are clipped or not shown. The requirement for the near Z clip plane in the background art is due to mathematical inaccuracies in computers, as objects cannot to come infinitely close to the camera. However, for the purposes of the present invention, this parameter is preferably set at a specific location in order to cause the object to be cut at a specific plane, in order to be able to render the ultrasound image at that plane. This parameter could optionally be set in any type of three-dimensional (3D) rendering software to cause the object to be so clipped. The general function of a near Z clipping plane is to remove all the geometry that is closer to the camera than to that plane.


Next, both sides of the surface of the whole object are preferably rendered. All the surfaces of the object are rendered, even those which are not seen such as those which are removed from view, for example in the event that a near Z clipping plane has been applied. When the object is rendered, the stencil buffer associated with that software is preferably used to count the amount of pixels written to the buffer. If an even amount of pixels are written, the slice is not visible, and if an odd number of pixels are written, the slice is visible, as shown with regard to FIG. 2. This effect occurs because both sides are rendered, but with different parameters. For those portions of the object which should not be displayed (such as the back part of an object, which could not be seen from the perspective of the user), each time a pixel is written to the image buffer, the stencil value for this pixel is incremented. For those portions which should be displayed, the value in the stencil buffer is decreased by one. The stencil buffer is therefore a bitmap of 0 and 1 values, in which values of “1” indicate where this object intersects the ultrasound image plane.


A texture is then preferably selected which resembles the texture of the ultrasound image for the desired object, which could optionally be extracted from real ultrasound images of the organ or other tissue, such as images of the liver for example. Next, this texture is rendered by using the previously defined information from each slice as the alpha channel. The alpha channel is another color channel which is added frequently in 3D graphics and computer graphics generally, in addition to the common Red/Green/Blue channels (or color components). The alpha channel is related to the amount of transparency of the color. The data of the fourth channel is then preferably applied to all graphic components containing color. With regard to FIGS. 2 and 3, the alpha channel is used to determine which areas of the simulated images below are transparent.


With regard to the texture, in order to add visual realism to those portions of image in the previously described stencil buffer, texture is preferably added to those portions of the image having values of “1” in that buffer. For the purposes of this description, the texture is preferably added in grayscale, although of course colors could optionally be added.


These steps are preferably performed for each object separately until all visible object slices are rendered, resulting in a simulated image as shown below with regard to FIG. 4. The image is then preferably “bent” and blurred to simulate the poor resolution of true ultrasound images. The process of “bending” the image is more preferably performed mathematically, while blurring may optionally and more preferably be performed through image processing which also involves a mathematical function, and which is also very well known in the art.


All the work done previously is rendered to a surface that is not visible to the user, such that there is now a rectangular picture that represents slices of all the objects that intersected the US plane. If it is taken into consideration that the US probe is not straight, but is in fact an array of sensors in the shape of an an arc, the resulting image should be in the shape of a sector circumscribed by an arc parallel to the base arc of the sector. To obtain this image the rectangular picture is rendered onto a template in the shape of a sector as described above, in which the template is a polygonal model that describes the ultimate overall shape of the image. The result of these processes is shown in the exemplary image of FIG. 5.


After these processes have been performed, the relevant anatomy has been modeled onto surface models of the anatomical feature, such as the organ. Each object has been rendered with the associated ultrasound properties, such as how the tissue responds to ultrasound waves in terms of echoing, and with a sample of the correct filling texture for simulation of the ultrasound image. For example, since moving the endoscope or other tool which holds the probe also alters the probe's position, the position of the cutting plane may be altered with each movement. In turn, the slices generated by the cutting plane are changed, such that a new ultrasound image is formed. The final simulated ultrasound image is then displayed to the user (see FIG. 5 for an example).


As previously described, one of ordinary skill in the art could easily adjust the method of the present invention in order to accommodate different types of ultrasound scanning technologies. For example, for a radial scanning the device, the image would only have “blur” added, but would not be “bent”, because the plane of scanning is perpendicular to the device. Also, the image would be round rather than square.


It will be appreciated that the above descriptions are intended only to serve as examples, and that many other embodiments are possible within the spirit and the scope of the present invention.

Claims
  • 1. A computer-implemented method for simulating an ultrasound image of at least one anatomical feature of a subject, the method comprising: modeling, by a computer, a surface geometry of the at least one anatomical feature as a polygonal representation within a polygonal surface model;based on the surface geometry, generating, by the computer, an image of at least one planar slice of said surface geometry based on a location of a simulated source of ultrasound waves and the polygonal surface geometry of a portion of the subject which would receive ultrasound waves if an ultrasound diagnostic procedure was actually being performed;based on the image of said at least one planar slice, rendering, by the computer, the simulated ultrasound image according to said image of said at least one planar slice as part of a simulation of an endoscopic medical procedure, wherein rendering the simulated ultrasound image includes simulating echoes reflected from tissue of the anatomical feature, and adding to the image a filling texture; anddisplaying, by the computer, the rendered image to a user.
  • 2. The method of claim 1, wherein said generating the image is performed by using a stencil buffer by: positioning a viewpoint for collecting the simulated ultrasound image above said planar slice; andsetting a near Z clip plane to fit a cutting plane for cutting said planar slice.
  • 3. The method of claim 2, wherein said generating with said stencil buffer is further performed by: setting a value for pixels for said planar slice to a positive value, and a value of pixels for the remainder of the anatomical feature to a zero value; andrendering said pixels.
  • 4. The method of claim 2, wherein said generating with said stencil buffer is further performed by: determining at least one intersection of the anatomical feature with said planar slice; andsetting a value for said pixels in said planar slice to a positive value for said stencil buffer.
  • 5. The method of claim 1, wherein the simulated ultrasound image is a grayscale image and wherein said filling texture is collected from an actual ultrasound medical procedure.
  • 6. The method of claim 1, wherein the simulated ultrasound image is a black and white image, and wherein said filling texture is collected from an actual ultrasound medical procedure.
  • 7. The method of claim 5, wherein said generating the image is performed by using a stencil buffer and wherein said rendering with said filling texture is only performed for pixels in said stencil buffer having a positive value.
  • 8. The method of claim 7, wherein said pixels having said positive value are white pixels.
  • 9. The method of claim 5, wherein the simulated ultrasound image is created by rendering said filling texture on said planar slice image.
  • 10. The method of claim 9, wherein rendering said filling texture further comprises at least one of bending and blurring said simulated image.
  • 11. The method of claim 10, wherein said bending further comprises: rendering said simulated image onto a template, said template comprising a polygonal model constructed according to a shape of an image from an actual ultrasound medical procedure.
Priority Claims (1)
Number Date Country Kind
143255 May 2001 IL national
PCT Information
Filing Document Filing Date Country Kind 371c Date
PCT/IL02/00393 5/20/2002 WO 00 8/6/2004
Publishing Document Publishing Date Country Kind
WO02/094080 11/28/2002 WO A
US Referenced Citations (391)
Number Name Date Kind
1959490 Mistelski May 1934 A
3024539 Rider Mar 1962 A
3263824 Jones et al. Aug 1966 A
3406601 Clifford Oct 1968 A
3490059 Paulsen et al. Jan 1970 A
3517446 Corlyon et al. Jun 1970 A
3520071 Abrahamson et al. Jul 1970 A
3573444 Kawabata et al. Apr 1971 A
3579842 Scher May 1971 A
3704529 Cioppa Dec 1972 A
3722108 Chase Mar 1973 A
3739276 Dornberger Jun 1973 A
3775865 Rowan Dec 1973 A
3789518 Chase Feb 1974 A
3795061 Sarnoff et al. Mar 1974 A
3795150 Eckhardt Mar 1974 A
3814145 Gott et al. Jun 1974 A
3861065 Courtenay et al. Jan 1975 A
3875488 Crocker et al. Apr 1975 A
3919691 Noll Nov 1975 A
3945593 Schanzer Mar 1976 A
3991490 Markman Nov 1976 A
4024473 Edge et al. May 1977 A
4024873 Antoshkiw et al. May 1977 A
4033331 Guss et al. Jul 1977 A
4078317 Wheatley et al. Mar 1978 A
4089494 Anderson et al. May 1978 A
4115755 Cotton Sep 1978 A
4136554 Larson Jan 1979 A
4148014 Burson Apr 1979 A
4162582 McGraw et al. Jul 1979 A
4177984 Douglas et al. Dec 1979 A
4182054 Wise et al. Jan 1980 A
4183249 Anderson Jan 1980 A
4227319 Guy et al. Oct 1980 A
4236685 Kissel Dec 1980 A
4250636 Horwitz Feb 1981 A
4250887 Dardik et al. Feb 1981 A
4262549 Schwellenbach Apr 1981 A
4264312 Cianci Apr 1981 A
4276702 Horwitz Jul 1981 A
4307539 Klein Dec 1981 A
4333070 Barnes Jun 1982 A
4334216 Lacroix Jun 1982 A
4360345 Hon Nov 1982 A
4398889 Lam et al. Aug 1983 A
4427388 Hope Jan 1984 A
4436188 Jones Mar 1984 A
4459113 Boscaro Gatti et al. Jul 1984 A
4464117 Foerst Aug 1984 A
4478407 Manabe Oct 1984 A
4481001 Graham et al. Nov 1984 A
4504233 Galus et al. Mar 1985 A
4513235 Acklam et al. Apr 1985 A
4545390 Leary Oct 1985 A
4550617 Fraignier et al. Nov 1985 A
4551101 Neumann Nov 1985 A
4573452 Greenberg Mar 1986 A
4599070 Hladky et al. Jul 1986 A
4604016 Joyce Aug 1986 A
4605373 Rosen Aug 1986 A
4632341 Repperger et al. Dec 1986 A
4642055 Saliterman Feb 1987 A
4646742 Packard et al. Mar 1987 A
4654648 Herrington et al. Mar 1987 A
4655673 Hawkes Apr 1987 A
4659313 Kuster et al. Apr 1987 A
4667182 Murphy May 1987 A
4688983 Lindbom Aug 1987 A
4706006 Solomon Nov 1987 A
4708650 Holewinski et al. Nov 1987 A
4708656 de Vries et al. Nov 1987 A
4712101 Culver Dec 1987 A
4713007 Alban Dec 1987 A
4726772 Amplatz Feb 1988 A
4733214 Andresen Mar 1988 A
4742815 Ninan et al. May 1988 A
4748984 Patel Jun 1988 A
4751662 Crosbie Jun 1988 A
4757302 Hatakeyama et al. Jul 1988 A
4769763 Trieb et al. Sep 1988 A
4775289 Kazerooni Oct 1988 A
4782327 Kley et al. Nov 1988 A
4786892 Kubo et al. Nov 1988 A
4789340 Zikria Dec 1988 A
4794384 Jackson Dec 1988 A
4795296 Jau Jan 1989 A
4797104 Laerdal et al. Jan 1989 A
4803413 Kendig et al. Feb 1989 A
4820162 Ross Apr 1989 A
4823634 Culver Apr 1989 A
4825875 Ninan et al. May 1989 A
4839838 LaBiche et al. Jun 1989 A
4857881 Hayes Aug 1989 A
4860215 Seraji Aug 1989 A
4865423 Doi Sep 1989 A
4867685 Brush et al. Sep 1989 A
4868549 Affinito et al. Sep 1989 A
4870964 Bailey, Jr. et al. Oct 1989 A
4874998 Hollis, Jr. Oct 1989 A
H703 Repperger et al. Nov 1989 H
4879556 Duimel Nov 1989 A
4881324 Khinchuk Nov 1989 A
4885565 Embach Dec 1989 A
4887966 Gellerman Dec 1989 A
4891764 McIntosh Jan 1990 A
4896554 Culver Jan 1990 A
4907796 Roel-Rodriguez Mar 1990 A
4907970 Meenen Mar 1990 A
4907973 Hon Mar 1990 A
4909232 Carella Mar 1990 A
4912638 Pratt Mar 1990 A
4930770 Baker Jun 1990 A
4934694 McIntosh Jun 1990 A
4940234 Ishida et al. Jul 1990 A
4949119 Moncrief et al. Aug 1990 A
4955654 Tsuchihashi et al. Sep 1990 A
4961138 Gorniak Oct 1990 A
4961267 Herzog Oct 1990 A
4964097 Wang et al. Oct 1990 A
4975546 Craig Dec 1990 A
4982618 Culver Jan 1991 A
4982918 Kaye Jan 1991 A
4998916 Hammerslag et al. Mar 1991 A
5004391 Burdea Apr 1991 A
5007300 Siva Apr 1991 A
5009598 Bennington Apr 1991 A
5018922 Yoshinada et al. May 1991 A
5019761 Kraft May 1991 A
5021982 Crosbie et al. Jun 1991 A
5022384 Freels et al. Jun 1991 A
5033352 Kellogg et al. Jul 1991 A
5044956 Behensky et al. Sep 1991 A
5048508 Storz Sep 1991 A
5055051 Duncan Oct 1991 A
5057078 Foote et al. Oct 1991 A
5062594 Repperger Nov 1991 A
5072361 Davis et al. Dec 1991 A
5077769 Franciose Dec 1991 A
5078152 Bond et al. Jan 1992 A
5086296 Clark Feb 1992 A
5103404 McIntosh Apr 1992 A
5104328 Lounsbury Apr 1992 A
5112228 Zouras May 1992 A
5116051 Moncrief et al. May 1992 A
5116180 Fung et al. May 1992 A
5125843 Holloway Jun 1992 A
5126948 Mitchell et al. Jun 1992 A
5135488 Foote et al. Aug 1992 A
5139261 Openiano Aug 1992 A
5142931 Menahem Sep 1992 A
5143505 Burdea et al. Sep 1992 A
5146566 Hollis, Jr. et al. Sep 1992 A
5149270 McKeown Sep 1992 A
5151856 Halmann et al. Sep 1992 A
5153716 Smith Oct 1992 A
5158459 Edelberg Oct 1992 A
5167159 Lucking Dec 1992 A
5171299 Heitzmann et al. Dec 1992 A
5177473 Drysdale Jan 1993 A
5180351 Ehrenfried Jan 1993 A
5181181 Glynn Jan 1993 A
5184306 Erdman et al. Feb 1993 A
5184319 Kramer Feb 1993 A
5185561 Good et al. Feb 1993 A
5186629 Rohen Feb 1993 A
5189355 Larkins et al. Feb 1993 A
5191320 MacKay Mar 1993 A
5193963 McAffee et al. Mar 1993 A
5196017 Silva et al. Mar 1993 A
5197003 Moncrief et al. Mar 1993 A
5203563 Loper, III Apr 1993 A
5204600 Kahkoska Apr 1993 A
5209131 Baxter May 1993 A
5209661 Hildreth et al. May 1993 A
5212473 Louis May 1993 A
5215523 Williams et al. Jun 1993 A
5220260 Schuler Jun 1993 A
5222893 Hardesty Jun 1993 A
5223776 Radke et al. Jun 1993 A
5228356 Chuang Jul 1993 A
5230339 Charlebois Jul 1993 A
5240417 Smithson et al. Aug 1993 A
5243266 Kasagami et al. Sep 1993 A
5246007 Frisbie et al. Sep 1993 A
5247432 Ueda Sep 1993 A
5252068 Gryder Oct 1993 A
5252070 Jarrett Oct 1993 A
5257462 Buttermann Nov 1993 A
5259626 Ho Nov 1993 A
5259894 Sampson Nov 1993 A
5264768 Gregory et al. Nov 1993 A
5265034 Breckenridge et al. Nov 1993 A
5269519 Malone Dec 1993 A
5275174 Cook Jan 1994 A
5275565 Moncrief Jan 1994 A
5279309 Taylor Jan 1994 A
5279563 Brucker et al. Jan 1994 A
5280265 Kramer et al. Jan 1994 A
5283970 Aigner Feb 1994 A
5286203 Fuller et al. Feb 1994 A
5295694 Levin Mar 1994 A
5296846 Ledley Mar 1994 A
5296871 Paley Mar 1994 A
5305203 Raab Apr 1994 A
5309140 Everett, Jr. et al. May 1994 A
5311422 Loftin et al. May 1994 A
5313230 Venolia et al. May 1994 A
5313568 Wallace et al. May 1994 A
5314339 Aponte May 1994 A
5317689 Nack et al. May 1994 A
5318533 Adams et al. Jun 1994 A
5324260 O'Neill et al. Jun 1994 A
5327790 Levin et al. Jul 1994 A
5334027 Wherlock Aug 1994 A
5335557 Yasutake Aug 1994 A
5344354 Wiley Sep 1994 A
5353242 Crosbie et al. Oct 1994 A
5354162 Burdea et al. Oct 1994 A
5355148 Anderson Oct 1994 A
5364271 Aknin et al. Nov 1994 A
5366376 Copperman et al. Nov 1994 A
5368478 Andreiko et al. Nov 1994 A
5368484 Copperman et al. Nov 1994 A
5368487 Medina Nov 1994 A
5368565 DeLong Nov 1994 A
5370535 Prendergast Dec 1994 A
5379663 Hara Jan 1995 A
5382885 Salcudean et al. Jan 1995 A
5384460 Tseng Jan 1995 A
5385549 Lampropoulos et al. Jan 1995 A
5389865 Jacobus et al. Feb 1995 A
5396267 Bouton Mar 1995 A
5397308 Ellis et al. Mar 1995 A
5397323 Taylor et al. Mar 1995 A
5399091 Mitsumoto Mar 1995 A
5402801 Taylor Apr 1995 A
5403191 Tuason Apr 1995 A
5412189 Cragun May 1995 A
5412880 Raab May 1995 A
5414337 Schuler May 1995 A
5423754 Cornelius et al. Jun 1995 A
5425644 Szinicz Jun 1995 A
5425709 Gambale Jun 1995 A
5428748 Davidson et al. Jun 1995 A
5429140 Burdea et al. Jul 1995 A
5430665 Jin et al. Jul 1995 A
5436640 Reeves Jul 1995 A
5437290 Bolger et al. Aug 1995 A
5445166 Taylor Aug 1995 A
5451924 Massimino et al. Sep 1995 A
5459382 Jacobus et al. Oct 1995 A
5461711 Wang et al. Oct 1995 A
5467441 Stone et al. Nov 1995 A
5467763 McMahon et al. Nov 1995 A
5470232 Kelso et al. Nov 1995 A
5473235 Lance et al. Dec 1995 A
5482051 Reddy et al. Jan 1996 A
5492530 Fischell et al. Feb 1996 A
5506605 Paley Apr 1996 A
5512919 Araki Apr 1996 A
5515078 Greschler et al. May 1996 A
5524637 Erickson Jun 1996 A
5541831 Thomas Jul 1996 A
5542672 Meredith Aug 1996 A
5542676 Howe, Jr. et al. Aug 1996 A
5547382 Yamasaki et al. Aug 1996 A
5548694 Frisken Aug 1996 A
5553198 Wang et al. Sep 1996 A
5559412 Schuler Sep 1996 A
5565840 Thorner et al. Oct 1996 A
5575761 Hajianpour Nov 1996 A
5577981 Jarvik Nov 1996 A
5580249 Jacobsen et al. Dec 1996 A
5584701 Lampotang et al. Dec 1996 A
5587937 Massie et al. Dec 1996 A
5591924 Hilton Jan 1997 A
5592401 Kramer Jan 1997 A
5599301 Jacobs et al. Feb 1997 A
5600348 Bartholow et al. Feb 1997 A
5607157 Nagashima Mar 1997 A
5607308 Copperman et al. Mar 1997 A
5609485 Bergman et al. Mar 1997 A
5609607 Hechtenberg et al. Mar 1997 A
5616030 Watson Apr 1997 A
5623582 Rosenberg Apr 1997 A
5625551 Mitarai et al. Apr 1997 A
5625576 Massie et al. Apr 1997 A
5629594 Jacobus et al. May 1997 A
5631861 Kramer May 1997 A
5631973 Green May 1997 A
5643087 Marcus et al. Jul 1997 A
5651775 Walker et al. Jul 1997 A
5657429 Wang et al. Aug 1997 A
5661253 Aoki Aug 1997 A
5661667 Rueb et al. Aug 1997 A
5666473 Wallace Sep 1997 A
5669818 Thorner et al. Sep 1997 A
5676157 Kramer Oct 1997 A
5680590 Parti Oct 1997 A
5684722 Thorner et al. Nov 1997 A
5691898 Rosenberg et al. Nov 1997 A
5694013 Stewart et al. Dec 1997 A
5695500 Taylor et al. Dec 1997 A
5701140 Rosenberg et al. Dec 1997 A
5709219 Chen et al. Jan 1998 A
5716016 Iwade et al. Feb 1998 A
5720619 Fisslinger Feb 1998 A
5722403 McGee et al. Mar 1998 A
5724264 Rosenberg et al. Mar 1998 A
5731804 Rosenberg Mar 1998 A
5736978 Hasser et al. Apr 1998 A
5739811 Rosenberg et al. Apr 1998 A
5742278 Chen et al. Apr 1998 A
5749853 O'Donnell et al. May 1998 A
5755577 Gillio May 1998 A
5758649 Iwashita et al. Jun 1998 A
5766016 Sinclair et al. Jun 1998 A
5768413 Levin et al. Jun 1998 A
5769640 Jacobus et al. Jun 1998 A
5771181 Moore et al. Jun 1998 A
5776050 Chen et al. Jul 1998 A
5776126 Wilk et al. Jul 1998 A
5781172 Engel et al. Jul 1998 A
5782764 Werne Jul 1998 A
5797900 Madhani et al. Aug 1998 A
5800179 Bailey Sep 1998 A
5805140 Rosenberg et al. Sep 1998 A
5806521 Morimoto et al. Sep 1998 A
5807377 Madhani et al. Sep 1998 A
5808665 Green Sep 1998 A
5810007 Holupka et al. Sep 1998 A
5821920 Rosenberg et al. Oct 1998 A
5831408 Jacobus et al. Nov 1998 A
5841889 Seyed-Bolorforosh Nov 1998 A
5844392 Peurach et al. Dec 1998 A
5882206 Gillio Mar 1999 A
5889670 Schuler et al. Mar 1999 A
5889672 Schuler et al. Mar 1999 A
5930741 Kramer Jul 1999 A
5945978 Holmes Aug 1999 A
5956484 Rosenberg et al. Sep 1999 A
5986643 Harvill et al. Nov 1999 A
5999185 Kato et al. Dec 1999 A
6004134 Marcus et al. Dec 1999 A
6019724 Gronningsaeter et al. Feb 2000 A
6024576 Bevirt et al. Feb 2000 A
6037927 Rosenberg Mar 2000 A
6038488 Barnes et al. Mar 2000 A
6042555 Kramer et al. Mar 2000 A
6050962 Kramer et al. Apr 2000 A
6059506 Kramer May 2000 A
6062865 Bailey May 2000 A
6084587 Tarr et al. Jul 2000 A
6088017 Tremblay et al. Jul 2000 A
6104379 Petrich et al. Aug 2000 A
6110130 Kramer Aug 2000 A
6111577 Zilles et al. Aug 2000 A
6117078 Lysyansky et al. Sep 2000 A
6120465 Guthrie et al. Sep 2000 A
6148280 Kramer Nov 2000 A
6157193 Renz et al. Dec 2000 A
6160489 Perry et al. Dec 2000 A
6162174 Friemel Dec 2000 A
6162190 Kramer Dec 2000 A
6193657 Drapkin Feb 2001 B1
6195592 Schuler Feb 2001 B1
6210168 Aiger et al. Apr 2001 B1
6219032 Rosenberg et al. Apr 2001 B1
6222523 Harvill et al. Apr 2001 B1
6239784 Holmes May 2001 B1
6264609 Herrington et al. Jul 2001 B1
6275213 Tremblay et al. Aug 2001 B1
6318146 Madsen et al. Nov 2001 B1
6323837 Rosenberg Nov 2001 B1
6346940 Fukunaga Feb 2002 B1
6413229 Kramer et al. Jul 2002 B1
6428490 Kramer et al. Aug 2002 B1
6497672 Kramer Dec 2002 B2
6544041 Damadian Apr 2003 B1
RE38242 Engel et al. Sep 2003 E
6694163 Vining Feb 2004 B1
6876891 Schuler et al. Apr 2005 B1
6885361 Harvill et al. Apr 2005 B1
6939138 Chosack et al. Sep 2005 B2
7215326 Rosenberg May 2007 B2
20020072814 Schuler et al. Jun 2002 A1
20020168618 Anderson et al. Nov 2002 A1
20020176619 Love Nov 2002 A1
20030112922 Burdette et al. Jun 2003 A1
20040034301 Falco Feb 2004 A1
Foreign Referenced Citations (33)
Number Date Country
0 147 516 Mar 1988 EP
0 265 011 Apr 1988 EP
0 393 683 Oct 1990 EP
0 456 103 Nov 1991 EP
0 489 469 Jun 1992 EP
0 316 763 Aug 1992 EP
0 567 215 Oct 1993 EP
0 571 827 Dec 1993 EP
0 624 861 Nov 1994 EP
0 626 634 Nov 1994 EP
0 623 066 Jul 1997 EP
0 632 709 Mar 2002 EP
2592514 Dec 1985 FR
2 195 808 Apr 1988 GB
2 252 656 Aug 1992 GB
2 288 686 Oct 1995 GB
03-98080 Apr 1991 JP
WO9106935 May 1991 WO
WO 9111775 Aug 1991 WO
WO 9304625 Mar 1993 WO
WO 9308517 Apr 1993 WO
WO 9314483 Jul 1993 WO
WO 9318475 Sep 1993 WO
WO 9425948 Nov 1994 WO
WO 9502233 Jan 1995 WO
WO 9502233 Jan 1995 WO
WO9510080 Apr 1995 WO
WO 9532459 Nov 1995 WO
WO 9616389 May 1996 WO
WO 9616389 May 1996 WO
WO9628800 Sep 1996 WO
WO 9938141 Jul 1999 WO
WO 9938141 Jul 1999 WO
Non-Patent Literature Citations (160)
Entry
Yoshitaka, Adachi et al. Intermediate Interpretation for Stiff Virtual Objects Proceedings of the Virtual Reality Annual International Symposium (VRAIS' 95) Technical Research Center, Suzuki Motor Corporation, Yokohama, Japan.
M.L. Agronin. The Design of a Nine-String Six-Degree-of-Freedom Force-Feedback Joystick for Telemanipulation. pp. 340-348 Jet Propulsion Laboratory—California, LA.
Peter K Allen et al. Acquisition and Interpretation of 3-D Sensor Data from Touch Dept. of Computer Science, Columbia University, NY CH2813-4/89/0000/0033/$01.00 1989 IEEE pp. 33-40.
Fumihito Arai et al. Intelligent Assistance in Operation of Active Catheter for Minimum Invasive Surgery. Nagoya University—Nagoya, Japan Kinjo University—Nagoya, Japan IEEE International Workshop on Robot and Human Communication 0-7803-2002-6/94 $4.00 1994 IEEE.
Daniel Bachofen et al. Enhancing the Visual Realism of Hysteroscopy Simulation Book Series: Studies in Health Technology and Informatics—Book Medicine meets Virtual Reality 14: Accelerating Change in Health Care: Next Medical Toolkit vol. 119/2005 pp. 31-36.
J. Baille et al. Use of Computer Graphics Simulation for Teaching of Flexible Sigmoidoscopy. Duke University Medical Center, North Carolina, USA. Endoscopy 3 vol. May 23, 1991 pp. 126-129.
David Baraff An Introduction to Physically Based Modeling: Ridged Body Simulation II—Nonpenetration Constraints Robotics Institute Carnegie Mellon Institute pp. 32-68 1997.
Adelstein et al. ASME Symposium 1992 Design and Implementation of a Force Reflecting Manipuladum for Manual Control Research. CA, USA.
J.H. Anderson et al. Da Vinci: A Vascular Catheterization and Interventional Radiology-Based Training and Patient Pretreatment Planning Simulator (Abstract) JVIR Supplement, Journal of Vascular and Interventional Radiology, vol. 7, No. 1, Part 2. Jan.-Feb. 1996 Washington, US.
J. Batter and F. Brooks, Jr. Grope-1: A Computer Display to the Sense of Feel 1972 North Carolina, USA.
M. Bostrom et al. Design of an Interactive Lumbar Puncture Simulator With Tactile Feedback IEEE Neutral Network Counsel Virtual Reality Annual International Symposium Conference Sep. 18-22, 1993; Seattle, Washington U.S.
M. Bostrom Design of Hardware for Simulating Lumbar Puncture with Force Feedback Thayer School of Engineering, Dartmouth College. Mar. 17, 1993.
F. P. Brooks, et al. Project Grope—Haptic Displays for Scientific Visualization ACM, Computer Graphics, vol. 24, No. 4. Aug. 1990—Chapel Hill NC, USA.
G. Burdea and T. Speeter Nasa Portable Dextrous Force Feedback Master for Robot Telemanipulation (PDMFF) pp. 153-161 Technical Report Server (NTRS), NJ ,USA.
Burdea et al. A Distributed Virtual Environment with Dextrous Force Feedback Informatique '92, International Conference Interface to Real and Virtual Worlds, Rutgers University EC2 Conference Mar. 23-27, 1992, NJ, USA.
J. Capowski, Remote Manipulators as a Computer Input Device University Microfilms, A XEROX Company, Ann Arbor, Michigan UMI Dissertation Services. 1971—Michigan USA.
Cover et al. Interactively Deformable Models for Surgery Simulation (Object Modeling) Computer Graphics & Applications IEEE pp. 68-75 Atlanta, GA, USA.
J. S. Denson and S. Abrahamson A Computer-Controlled Patient Simulator Apr. 21, 1969—vol. 208, No. 3 pp. 504-508 LA, USA.
D. Gillies and C. Williams, London UK An Interactive Graphic Simulator for the Teaching of Fibrendoscopic Techniques Eurographics '87 Elsevier Science Publishers B.V North Holland pp. 127-138.
Gillies, Haritsis and Williams Computer Simulation for Teaching Endoscopic Procedures Endoscopy, Supplement II, vol. 24, Jul. 1992. pp. 455-550.
A. Haritsis D. Gillies Ch. Williams (Eurographics) Realistic Generation and Real Time Animation of Images of the Human Colon Computer Graphics Forum vol. II No. 3, conference issue—Sep. 7-11, 1992. NNC Blackwell.
Haritsis 1992 (Hellenic) A.Haritsis D. Gillies Ch. Williams Computer Simulation: New Horizons in Endoscopy Teaching Hellenic Journal of Gastroenterology 1992 pp. 54-63 London UK.
G. Higgins, et al. Higgins 1995 (Surg. Tech. Int'l IV) Virtual Reality Surgery: Implementation of a Coronary Angioplasty Training Simulator. University Medical Press, San Francisco, 1995. pp. 379-383.
D. Hon Ixion's Laparoscopic Surgical Skills Simulator Symposium: Medicine Meets Virtual Reality II Jan. 27-30, 1994 San Diego, USA.
D. Hon Ixion's Realistic Medical Simulations Virtual Reality World, vol. 2, No. 4 Jul./Aug. 1994 pp. 58-62.
H. Iwata Artificial Reality with Force-feedback: Development of Desktop Virtual Space with Compact Master Manipulator. ACM SIGGRPAH 1990 Computer Graphics & Interactive Techniques vol. 24, No. 4. pp. 165-170 Aug. 6-10, 1990.
B.G Jackson L.B Rosenberg Force Feedback and Medical Simulation IOS Press and Ohmsha Jan. 19-22, 1995 pp. 147-151—CA, USA.
P.J. Kilpatrick Kilpatrick Thesis 1976 pp. 11-27 The Use of a Kinesthetic Supplement in an Interactive Graphics System. The University of North Carolina, USA.
Kotoku et al. A Force Display System for Virtual Environments and its Evaluation International Workshop on Robot and Human Communication IEEE Sep. 1-3, 1992 pp. 246-251—Ibaraki, Japan.
U.G. Kuhnapfel Realtime Graphical Computer Simulation for Endoscopic Surgery Symposium: Medicine Meets Virtual Reality II Jan. 27-30, 1994 San Diego, CA, USA.
U.G.Kuhnapfel et al. Endo surgery simulations with KISMET: a flexible tool for surgical instrument design, operation room planning and VR technology based abdominal surgery training. Virtual Reality World '95, Conference Stuttgart, Germany Computerwoche Verlag, 1995. pp. 165-171.
B. Marcus Feedback Technology and Virtual Environments pp. 87-95 Jul. 1-3, 1992—1992 International Conference on Artificial Reality and Telexistence (ICAT 1992) pp. 87-95.
Mark et al. Adding Force Feedback to Graphics Systems: Issues and Solutions Aug. 4-9, 1996 ACM SIGGRAPH 1996 Computer Graphics Proceedings, Annual Conference Chapel Hill. North Carolina, Usa.
T.H. Massie Design of a Three Degree of Freedom Force-Reflecting Haptic Interface MIT, USA Thesis—pp. 6-38 May 18, 1993 Submitted May 17, 1993.
K.T. McGovern et al. The Virtual Clinic™, A Virtual Reality Surgical Simulator Symposium: Medicine Meets Virtual Reality II pp. 151-157 Jan. 27-30, 1994 San-Diego CA, USA.
D. Meglan Making Surgical Simulation Real ACM SIGGRAPH Computer Graphics pp. 37-39 Nov. 1996 Rockville, MD, USA.
Meglan et al. The Teleos Virtual Environment Toolkit for Simulation-Based Surgical Education Interactive Technology and the New Paradigm for Healthcare Proceeding of MMVR 3, IOS Press and Ohmsha pp. 346-351. Jan. 17-20, 1996 San-Diego CA, USA.
J. R. Merril The Future of Virtual Reality, Medicine, and the Information Superhighway Journal of Knowledge Engineering & Technology, vol. 7, No. 1 Spring 1994 pp. 33-35 MD, USA.
Merril et al. Photorealistic Interactive Three-Dimensional Graphics in Surgical Simulation Interactive Technology and the New Paradigm for Healthcare Proceeding of MMVR 3, IOS Press and Ohmsha pp. 244-252 Jan. 19-22, 1995 San Diego, USA.
Merril et al. Surgical Simulation Using Virtual Reality Technology: Design, Implementation, and Implications. Surgical Technology International III 1994 pp. 53-60. Published by Universal Medical Press, CA, USA.
Merril et al. Virtual Heart Surgery—Trade Show and Medical Education 1994 Virtual Reality World pp. 55-57 Jul./Aug. 1994 MD, USA.
Merril et al Cyber Surgery—Cutting Costs, Sewing Benefits The Virtual Reality Special Report, Miller Freedman Inc. Summer 1994 pp. 39-42 MD, USA.
Minsky et al. Feeling and Seeing: Issues in Force Display ACM 1990 pp. 235-243 CA, USA.
M.D. Noar N. Soehendra Endoscopy Simulation Training Devices Endoscopy 1992, vol. 24 pp. 159-166 Georg Thieme Vering Stuttgart. New York.
M.D. Noar Robotics Interactive Simulation of RCP Sphincterotomy and EGD, Endoscopy 1992, vol. 24, pp. 539-541 Supplement 2 Georg Thieme Vering Stuttgart. New York.
A. M. Noll Man-Machine Tactile Communication Polytechnic Institute of Brooklyn, Jun. 1971, pp. 1V-X111 and 1-87.
Ernest M. Otani Software Tools for Dynamic and Kinematic Modeling of Human Emotion Department of Computer & Information Science Technical Reports (CIS) University of Pennsylvania, Jul. 1989, pp. 1-74.
M. Ouh-Young Force Display in Molecular Docking UNC, The University of North Carolina at Chapel Hill 1990, pp. 1-369.
J. Peifer, et al. Medicine Meets Virtual Reality, Health Care in the Information Age Applied Virtual Reality for Simulation of Endoscopic Retrograde Cholangio-Pancreatography IOM Press, Proceedings of Medicine Meets Virtual Reality 4, San Diego, California, Jan. 17-20, 1996, pp. 36-42.
S. Pieper et al. Stereoscopic Displays and Applications II Virtual environment system for simulation of leg surgery SPIE vol. 1457, Feb. 25-27, 1991, pp. 188-197.
S. Pieper et al. Interactive Graphics for Plastic Surgery: A task-level analysis and Implementation 1992 ACM Computer Graphics Special Issue on 1992 Symposium on Interactive 3D Graphics, Cambridge, MA Mar. 29-Apr. 1, 1992, pp. 127-134.
D. Popa Simulation of Lumbar Puncture Procedure using Force Feedback in Virtual Environments Thayer School of Engineering, Dartmouth College, Hanover, New Hampshire, Jun. 1994, pp. 1-134.
Preminger et al. Medicine Meets Virtual Reality, Health Care in the Information Age Virtual Reality Surgical Simulation in Endoscopic Urologic Surgery IOM Press, Proceedings of Medicine Meets Virtual Reality 4, San Diego, California, Jan. 17-20, 1996, Chapter 19, pp. 157-163.
L.B Rosenberg, B.G Jackson Foot-Based Interfaces to Virtual Environments Using the Immersion Interface Box (TM) Virtual Reality and Persons With Disabilities, Second Annual International Conference, Jun. 8-10, 1994, pp. 145-148.
L.B Rosenberg “Virtual Fixtures”—Perceptual overlays enhance operator performance in telepresence tasks Stanford University, Aug. 1994. pp. 1-214.
M. A. Russo The Design and Implementation of a Three Degree of Freedom of Freedom Force Output Joystick MIT, May 11, 1990. pp. 1-131.
Salisbury et al. Haptic Rendering: Programming Touch Interaction with Virtual Objects Symposium on Interactive 3D Graphics, 1995 ACM, pp. 123-130.
S. S. Saliterman A Computerized Simulator for Critical-Care Training: New Technology for Medical Education Scientific session of the Mayo Medical School Alumni Society , Nov. 4, 1989, pp. 968-978.
B. Schmult et al. Application Areas for a Force-Feedback Joystick DSC vol. 49. Advances in Robotics, Mechatronics, and Haptic Interfaces ASME 1993, pp. 47-54.
Singh et al. Design of an Interactive Lumbar Puncture Simulator With Tactile Feedback IEEE International Conference on Robotics and Automation, May 8-13, 1994, pp. 1734-1739.
M. Stanley and J. Colgate Computer Simulation of Interacting Dynamic Mechanical Systems using Distributed Memory Parallel Processors ASME Winter Annual Meeting, Symposium on Haptic Interfaces for Virtual Environment and Teleoperator Systems, Nov. 8-13, 1992, pp. 55-61.
Sharon A. Stansfield Visually-Guided Haptic Object Recognition University of Pennsylvania 1987 UMI, pp. 1-216.
I. Sutherland The Ultimate Display for Production Proceedings of the IFIP Congress 1965, pp. 506-508.
D. Terzopoulos and D. Metaxas Dynamic 3D Models with Local and Global Deformations: Deformable Superquadrics IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 13, No. 7, Aug. 30, 1990, pp. 703-714.
Williams, Baillie, Gillies, Borislow and Cotton Teaching Gastrointestinal Endoscopy by Computer Simulation: a Prototype for Colonoscopy and ERCP Gastrointestinal Endoscopy vol. 36, No. 1., 1990, pp. 49-54.
C. M. Winey III Computer Simulated Visual and Tactile Feedback as an aid to Manipulator and Vehicle Control MIT, Jul. 31, 1981, pp. 1-132.
O.C Zienkiewicz The Finite Element Method McGraw-Hill Book Company (UK) Limited, 1977, pp. 677-757.
Beth A. Marcus Hands On: Haptic Feedback in Surgical Simulation Exos, Inc., Jan. 27-30, 1994, pp. SIMB 004163-SIMB 004174.
Virtual Reality and Medicine the Cutting Edge SIG Advanced Applications, Inc. Conference and Exhibition, Sep. 8-11, 1994, The New York Hilton.
Daane et al. A $100 Surgical Simulator for the IBM PC Interactive Technology and the New Paradigm for Healthcare Jan. 1995—pp. 79-80.
Higgins U.S. Army Medical Research Grant Annual Report entitled “Medical Simulation for Trauma Management” Grant No. DAMD 17-94-J-4470.
Strutz et al. 3-D Guided Endoscopic Surgery of Paranasal Sinusese Surgical Technology International IV, Oct. 1995, pp. 195-197.
Stone Haptic Human-Computer Interaction—Haptic Feedback: A Brief History from Telepresence to Virtual Reality Haptic Human-Computer Interaction, First International Workshop, Glasgow, UK Proceedings. Aug. 31-Sep. 1, 2000.
Loftin et al. A Virtual Environment for Laparoscopic Surgical Training Medicine Meets Virtual Reality II: Interactive Technology & Healthcare, Jan. 1994.
Durrani et al. Advanced Endoscopic Imaging: 3-D Laparoscopic Endoscopy Surgical Technology International III, Oct. 1994.
Johnston et al. Assessing a Virtual Reality Surgical Skills Simulator Stud Health Technol Inform. 1996; 29:608-17.
Sheridan Automatica the Journal of IFAC the International Federation of Automatic Control / Telerobotics. Automatica, vol. 25, No. 4, pp. 487-507.
Barfield et al Virtual Environments and Advanced Interface Design 1995 pp. 358-414.
Bejczy et al. Controlling Remote Manipulators Through Kinesthetic Coupling Computers in Mechanical Engineering Jul. 1983, pp. 48-60.
Beer-Gable Computer Assisted Training in Endoscopy (C.A.T.E.): From a Simulator to a Learning Station.Endoscopy 1992; 24:suppl. 2: pp. 534-538.
Kuenhapfel et al. CAD-Based Graphical Computer Simulation in Endoscopic Surgery Institute fur Angewandte Informatik, Kernforschumgszentr urn Karlsruhe Germany, Oct. 1994.
Campos et al. A Robotic Haptic System Architecture University of Pennsylvania, Dept. of Computer & Information Science Technical Reprot No. MS-CIS-00-51 1990.
Merril et al. Changing the Focus of Surgical Training Virtual Reality World, Mar./Apr. 1995, pp. 56-60.
Szabo et al. Choreographed Instrument Movements During Laparoscopic Surgery: Needle Driving, Knot Tying, and Anastomosis Techniques. Medicine Meets Virtual Reality II; Interactive Technology & Healthcare, Jan. 1994. pp. 216-217.
Dumay Cybersurgery Medicine Meets Virtual Reality II. Interactive Technology & Healthcare, Jan. 1994.
Greenleaf DataGlove and Datasuit: Virtual Reality Technology Applied to the Measurement of Human Movement. Medicine Meets Virtual Reality II. Interactive Technology & Healthcare, Jan. 1994, pp. 63-69.
Burdea et al. Dextrous Telerobotics with Force Feedback—An Overview, Part 1: Human Factors Rutgers—The State University of New Jersey, Dept. of Electrical and Computer Engineering, Robotica (1991) vol. 9, pp. 171-178.
Online reference dated May 31, 1995, updates chapter 13 of the AutoCAD Release 13 Developer's Guide dated Apr. 14, 1995.
Brochure: Dynacath Simulator Dynacath Minneapolis, MN.
Kuhn et al. Karlsruhe Endoscopic Surgery Trainer A “Virtual Reality” based Training System for Minimally Invasive Surgery.
Christensen Bringing Telematics Into Health Care in the European Communities Medicine Meets Virtual Reality II. Interactive Technology & Healthcare, Jan. 1994, 21-23.
Marcus et al. Exos Research on Master Controllers for Robotic Devices NASA Technical Reports Server NTRS, pp. 238-245 Jan. 1, 1992.
Merril VR for Medical Training and Trade Show “Fly-Pape”: Virtual Reality World, May/Jun. 1994, pp. 53-57.
Burdea Force and Touch Feedback for Virtual Reality Electrical and Computer-Engineering Dept. Rutgers—The State University of New Jersey.
Baumann et al. Force Feedback for Virtual Reality Based Minimally Invasive Surgery Simulator Medicine Meets Virtual Reality IV: Health Care in the Information Age, Jan. 1996.
Burdea Chapter 7 Force and Touch Feedback for Virtual Reality Physical Modeling, John Wiley & Sons, Inc., pp. 168-223.
Jason Fritz Haptic Rendering Techniques for Scientific Visualization Jason P. Fritz Thesis at University of Delaware Fall 1996.
Marcus Hands On: Haptic Feedback in Surgical Simulation Exos Inc. 2 A Gill St. Woburn, MA.
Rosenberg et al. A Haptic Interface for Virtual Simulation of Endoscopic Surgery Medicine Meets Virtual Reality IV: Health Care in the Information Age, Jan. 1996 pp. 371-387.
Haritsis et al. Realistic Generation and Real Time Animation of Images of the Human Colon Computer Graphics Forum, vol. II, No. 3, pp. C367-380.
Ho et al. IGES and PDES, The Current Status of Product Data Exchange Status Dept. of Computer Science, Univ. of Mo-Rolla, Rolla MO, 1988 IEEE, pp. 210-216.
Hooper The Interactive Assembly and Computer Animation of Reconfigurable Robotic Systems Mechanical Engineering Dept. The University of Texas at Austin. 1990.
Rosenberg Louis B. Human Interface Hardware for Virtual Laparoscopic Surgery Interactive Technology and the New Paradigm for Health Care Immersion Corp. Santa Clara, CA. Chapter49, pp. 322—Jan. 1995.
Funda et al. Image-guided Command and Control of a Surgical Robot Medicine Meets Virtual Reality II. Interactive Technology & Healthcare, Jan. 1994.
Jaramaz et al. Integrating Finite Element Analysis Into Pre-Operative Surgical Planning and Simulation of Total Joint Replacement Surgery Medicine Meets Virtual Reality II. Interactive Technology & Healthcare, Jan. 1994, pp. 34-37.
Merck & Co. An Introduction to The Robotic Endoscopy Simulator 1989.
Hastreiter et al. Intuitive and Interactive Manipulation of 3D Datasets by Integrating Texture Mapping Based Volume Rendering into the Open Inventor Class Hierarchy Lehrstuhl fur Graphische Datenverarbeitung (IMMD9) Universitat Erlangen.
Issacs et al. Controlling Dynamic Simulation with Kinematic Constraints, Behavior Functions and Inverse Dynamics Computer Graphics vol. 21, No. 4, pp. 215-224.
Filerman et al. Issues in the Design of Tactile Input Devices for Mechanical CAD Systems Massachusetts Institute of Technology, Artificial Intelligence Laboratory 1989.
Artificial Reality with Force-feed back: Development of Desktop Virtual Space with Compact Master Manipulator Iwata Computer Graphics vol. 24, No. 4, pp. 165-170.
Hon Ixion's Laparoscopic Surgical Skills Simulator Medicine Meets Virtual Reality II. Interactive Technology & Healthcare, Jan. 1994.
Kilpatrick The use of a Kinesthetic Supplement in an Interactive Graphics System Xerox University Microfilms 1976.
Kuhnapfel et al. Endosurgery Simulations with KISMET Virtual Reality World, pp. 165-171 1995.
Immersion Corporation Laparoscopic Impulse Engine Impulse Engine 2000™ Software Development Kit (Ver. 1.0)(Immersion) Immersion Corporation—Version 1.0 Mar. 1995.
McKensie et al. Lasers in Surgery and Medicine Wessex Regional Medical Physics Service and Department of Otolaryngology, vol. 29, No. 6, pp. 619-641 1984.
Massie et al. The Phantom Haptic Interface: A Device for Probing Virtual Objects Proceedings of the ASME Winter Annual Meeting, Symposium on Haptic Interfaces for Virtual Environment and Teleoperator Systems, Chicago, IL Nov. 1994.
McAfee et al Teleoperator Subsystem-Telerobot Demonstrator: Force Reflecting Hand Controller Equipment Manual Jet Propulsion Laboratory, California Institute of Technology, pp. 3-11.
Poston et al. The Medical Reality Sculptor Medicine Meets Virtual Reality II. Interactive Technology & Healthcare, Jan. 1994. pp. 174-176.
Satava Medical Virtual Reality: The Current Status of the Future Medicine Meets Virtual Reality IV: Health Care in the Information Age, Jan. 1996.
Merril et al. Virtual Reality for Trade Shows and Individual Physician Training Virtual Reality Systems, pp. 40-44 Spring 2004.
Metaxas et al. Dynamic Deformation of Solid Primitives with Constraints Computer Graphics Proceedings, Annual Conference Series, University of Toronto. pp. 309-312.
Flynn Virtual Reality and Virtual Spaces Find a Niche in Real Medicine; Simulated Surgery on a Computer—This Won't Hurt. New York Times Jun. 5, 1995.
Massimo et al. One Handed Tracking in Six Degrees of Freedom IEEE International Conference on Systems, MAN and Cybernetics, Massachusetts Institute of Technology, Man-Machine Systems Laboratory, vol. 1 of III.
Hannaford et al. Performance Evaluation of a Six-Axis Generalized Force-Reflecting Teleoperator IEEE May/Jun. 1991, vol. 21, No. 3 pp. 620-633.
Merril Presentation Material: Medicine Meets Virtual Reality II. Interactive Technology & Healthcare, Jan. 1994.
Immersion Human Interface Corporation Immersion Probe and Personal Digitizer Programmer's Technical Reference Manual: Immersion Probe and Personal Digitizer May 19, 1994.
Durlach Psychophysical Considerations in the Design of Human-Machine Interfaces for Teleoperator and Virtual-Environment Systems Medicine Meets Virtual Reality II. Interactive Technology & Healthcare, Jan. 1994 pp. 45-47.
Hubner et al. Real-Time Volume Visualization of Medical Image Data for Diagnostic and Navigational Purposes in Computer Aided Surgery Proc., Computer Assisted Radiology, CAR'96 Paris, pp. 751-756 Jun. 26-29, 1996.
Merril et al. Revealing the Mysteries of the Brain with VR Virtual Reality Special Report, Winter 1994, pp. 61-65.
Neisius et al. Robotic Telemanipulator for Laparoscopy 1995 IEEE-EMBC and CMBEC Theme 5: Neuromuscular Systems/ Biomechamics, pp. 1199-1200.1995.
Sato et at. Space Interface Device for Artificial Reality—SPIDAR System and Computers in Japan, vol. 23, No. 12, pp. 44-54.
Galyean et al. Sculpting: An Interactive Volumetric Modeling Technique Computer Graphics, vol. 25, No. 4.
Medical World News Virtual Reality Shapes Surgeon's Skills Medical World News, Feb. 1994, pp. 26-27.
Hon Tactile and Visual Simulation: A Realistic Endoscopy Experience Medicine Meets Virtual Reality: Discovering Applications for 3-D Multi-Media Interactive Technology in the Health Sciences, Jun. 4-7, 1992.
Johnson Tactile Feedback Enhancement to Laparoscopic Tools Medicine Meets Virtual Reality II. Interactive Technology & Healthcare, Jan. 1994.
Fischer et al. Tactile Feedback for Endoscopic Surgery Interactive Technology and the New Paradigm for Healthcare, Jan. 1995.
Peine et al. A Tactile Sensing and Display System for Surgical Applications Interactive Technology and the New Paradigm for Healthcare, Jan. 1995 pp. 283-288.
Computer Procedures for Finite Element Analysis Taylor Computer Procedures, The Finite Element Method, McCraw Hill, pp. 677-757 Computer Procedures for Finite Element Analysis.
Hunter et al. Teleoperated Microsurgical Robot and Associated Virtual Environment Medicine Meets Virtual Reality II. Interactive Technology & Healthcare, Jan. 1994.
Holler et al. Teleprescence Systems for Application in Minimally Invasive Surgery Medicine Meets Virtual Reality II. Interactive Technology & Healthcare, Jan. 1994.
Sheridan Telerobotics Automatica vol. 25, No. 4, pp. 487-507.
Satava Virtual Reality Surgical Simulator: The First Steps Medicine Meets Virtual Reality: Discovering Applications for 3-D Multi-Media Interactive Technology in the Health Sciences—Jun. 4-7, 1992.
Frolich et al. The Responsive Workbench: A Virtual Working Environment for Physicians Interactive Technology and the New Paradigm for Healthcare, Jan. 1995, pp. 118-119.
Doyle et al. The Virtual Embryo: VR Applications in Human Developmental Anatomy Medicine Meets Virtual Reality II: Interactive Technology & Healthcare, Jan. 1994, pp. 38-41.
Baille Gastrointestinal Endoscopy: Time for Change Scott Med J. Feb. 1989; 34 (1): 389-90.
Song et al. Tissue Cutting in Virtual Environments Interactive Technology and the New Paradigm for Healthcare, Jan. 1995.
Gyeong-Jae et al. Tissue Cutting in Virtual Environments Interactive Technology and the New Paradigm for Healthcare, Jan. 1995 359-364J.
Sukthankar Towards Virtual Reality of “Tissue Squeezing”: A Feasibility Study Medicine Meets Virtual Reality II: Interactive Technology & Healthcare, Jan. 1994, pp. 182-186.
Adachi Touch and Trace on the Free-Form Surface of Virtual Object Proceedings of IEEE Virtual Reality Annual International Symposium—Sep. 18-22, 1993 Seattle, WA pp. 162-168.
Trevidi et al. Developing Sensor-Based Robotic System Using Virtual Reality Concepts Proceedings for the ANS Fifth Topical Meeting on Robotics and Robotic Systems Knoxsville, TN/Apr. 25-30, vol. 1, pp. 165-172.
CH Products CH Products Virtual Pilot Control Yoke 1993.
Henderson “Virtual Realities” as Instructional Technology Journal of Interactive Instruction Development, pp. 24-30.
Hoffman Virtual Reality and the Medical Curriculum: Integrating Extant and Emerging Technologies Medicine Meets Virtual Reality II: Interactive Technology & Healthcare, Jan. 1994 pp. 73-76.
Burdea et al. Virtual Reality Technology Chap. 6, pp. 221-242. Wiley-Interscience 2003.
Iwata et al. Volume Haptization IEEE 1993, pp. 16-18.
Anon. VR in Medicine VR News; Apr. 1996 vol. 5, Issue 3.
Ota et al. Virtual Reality in Surgical Education ComputBiol Med., Mar. 1995, 25(2): 127-37.
MacDonald et al. Virtual Reality Technology Applied to Anesthesiology Interactive Technology and the New Paradigm for Healthcare, Jan. 1995.
Bell et al. The Virtual Reality Modeling Language, version 1.0 Specification 1996.
Merril Why I Simulate Surgery . . . Virtual Reality World, Nov./Dec. 1994, pp. 54-57.
Related Publications (1)
Number Date Country
20040253572 A1 Dec 2004 US