Client-server visualization system with hybrid data processing

Information

  • Patent Grant
  • 12170073
  • Patent Number
    12,170,073
  • Date Filed
    Tuesday, April 25, 2023
    a year ago
  • Date Issued
    Tuesday, December 17, 2024
    4 days ago
Abstract
The invention comprises a system of client-server visualization with hybrid data processing, having a server digital data processor, that allows for server side rendering and processing image data, and client digital data processors simultaneously connected to the server, which receives messages from the clients, creates rendered images of data sets or other data processing results and sends those rendered images and results to the clients for display or further processing. Performing certain image rendering operations on either the server or the client according to which is better suited for the tasks requested by the user at any point in time, and possibly adjusting this division of work dynamically, improves rendering speed and application responsiveness on the clients.
Description
SUMMARY OF THE INVENTION

The invention comprises, in one aspect, a system including one or more client computers that are coupled to, and can simultaneously connect with, a render server that renders and/or otherwise processes image data (including, by way of non-limiting example, 2D, 3D, and 4D medical or microscopy image data). The client computers generate messages that cause the render server to render images (or to perform other data processing tasks) and to return the results to the client computers for display or further processing. Rendering speed and application responsiveness on the client computers is improved by performing certain image rendering operations on either the server or the client, e.g., depending on which is better suited for the tasks requested by the user at any point in time, optionally, adjusting this division of work dynamically. We refer to this as client-server visualization with hybrid data processing.


In a related aspect, the invention comprises a system as described above, by way of example, wherein at least one of the client computers comprises local processing resources such as, for example, a central processing unit (CPU), a graphics processing unit (GPU), and/or a graphics library.


Such a client computer can include applications software or other functionality that generates requests for rendering one or more aspects of an image, including, by way of non limiting example, an image aspect (e.g., representing an acquired or synthesized image) and an overlay graphics aspect (e.g., representing textual or other graphical elements to be presented with the acquired/synthesized image). For example, the request to render an image aspect can include image data from a CT scan, and the request to render the overlay graphics aspect can include text (such as patient data), a ruler, a registration “cross-hair”, and so forth).


Such a client computer can, further, include a render module that responds to multiple requests by the applications software (or other functionality) by effecting processing of at least one aspect of at least one of the images using the local processing resources and messaging the render server to render (or otherwise process) the other aspect(s) of that and/or other images. Thus, continuing the example, the render module can respond to requests from the applications program by (i) rendering patient-identifying text (i.e., the overlay graphics aspect) of an image using the local CPU (or GPU) and (ii) messaging the render server to render CT scan data (the image aspect) of that image.


Related aspects of the invention provide systems as described above, by way of example, wherein the render module combines aspects of an image rendered utilizing the local resources with aspects rendered by the render server. To this end, the render module can paste into a buffer the image (or other) aspect of an image returned by the render server and can add to that buffer overlay graphics (or other) aspects rendered by local resources. The render module can make that buffer available for further processing and/or display, e.g., on a monitor or other display device.


Further aspects of the invention provide systems as described above, by way of example, wherein, in addition to (or instead of) image and overlay graphics aspects, one or more requests generated by the application can be for other aspects of an image, e.g., a perspective aspect (e.g., indicating a vantage point of a viewer), and so forth. Thus, for example, requests can be for an image comprising data from a CT scan (i.e., the image aspect), along with a specified viewer vantage point or virtual camera position (the perspective aspect) and, possibly, by way of further example, additionally having patient-identifying text (the overlay graphics aspect).


As above, the client computer render module can respond to such a requests by processing those for some aspects of the image using local processing resources, while messaging the render server to process those for others. Thus, continuing the example above, the render module can respond to the requests by (i) messaging the render server to compute a slice from CT scan data, (ii) obtaining that slice from the render server and rendering it, using a local GPU, from the specified vantage point and, optionally, (iii) combining it with locally rendered text. Such re-rendering may be effected, for example, in response to user requests to zoom or pan an image (or to adjust window/level settings for such an image).


Further aspects of the invention provide systems as described above, by way of example, in which the render server comprises a module that simultaneously processes image data in response to interleaved requests from one or more client computers. A related aspect of the invention provides such a system in which the render server includes one or more central processing units that process image data in response to such interleaved requests. A still further aspect of the invention provides such a system in which the render server module maintains requests in queues on the render server. Those requests may be received directly from the client digital data processors or may be generated as a result of messages received from them.


Further aspects of the invention provide systems employing combinations of the features described above.


Still further aspects of the invention provide methods for graphics processing paralleling the features described above.


These and other aspects of the invention are evident in the drawings and in the description that follows.





BRIEF DESCRIPTION OF THE DRAWINGS

A further appreciation of the invention may be attained by reference to the drawings, in which:



FIG. 1 illustrates a system according to one practice of the invention;



FIG. 2 depicts a computed tomography (CT) scan, including both image and overlay graphics aspects, of the type suitable for rendering by a system according to the invention;



FIG. 3 is a data-flow diagram depicting rendering of an image of the type shown in FIG. 2 in a system according to the invention;



FIGS. 4A-4B are data-flow diagrams depicting alternate allocations of rendering tasks between client and server computers in rendering of an image of the type shown in FIG. 2 in a system according to the invention; and



FIG. 5 is a data-flow diagram depicting an allocation of rendering tasks between client and server computers in rendering an image of a movie, video or other “moving picture” sequence in a system according to the invention





DETAILED DESCRIPTION OF THE ILLUSTRATED EMBODIMENT

The construction and operation of the illustrated embodiment may be more fully appreciated by reference to commonly assigned U.S. patent application Ser. No. 12/275,421, filed Nov. 21, 2008 by Westerhoff et al., entitled “multi-User multi-GPU Render Server Apparatus and Methods”, which issued as U.S. Pat. No. 8,319,781 on Nov. 27, 2012 (hereinafter, the “Related Application”), a non-provisional claiming the benefit of filing of U.S. Provisional Patent Application Ser. No. 60/989,881, entitled “multi-User multi-GPU Render Server Apparatus and Methods,” the teachings of both which are incorporated by reference herein.


Overview


FIG. 1 depicts a system 10 according to one practice of the invention. A render server, or server digital data processor, 11 is connected via one or more network interfaces 12, 13 and network devices, such as switches or hubs 14, 15, to one or more networks 22, 23—which can be WANs (wide area networks), LANs (local area networks), or other types of networks known in the art. One or more client computers, or client digital data processors, 16-21 are coupled for communications with render server 11 via networks 22, 23.


In the illustrated embodiment, software running on client computers 16-21 allows them to establish a network connection to the render server 11 on which server software is running. As a user interacts with one of the client computers, the software on that computer messages the render server 11, which renders or otherwise processes images (or partial images) in response and returns those images (or partial images) and/or processing results to the respective client computer for further processing and/or display.


The components illustrated in FIG. 1 may be of the type generally known in the art, as adapted in accord with the teachings hereof. Thus, by way of non-limiting example, illustrated render server 11 and client computers 16-21 may comprise conventional workstations, personal computers and other digital data processing apparatus of the type available in the marketplace, as adapted in accord with the teachings hereof.


The make-up of a client computer of the illustrated embodiment is shown, by way of example, in the break-out of FIG. 1. As illustrated, client computer 18 includes local processing resources such as, by way of non-limiting example, CPU 18a, RAM 18b, I/O section 18c and, optionally graphics processing unit 18d, all configured and operated in the conventional manner known in the art, as adapted in accord with the teachings hereof. One or more of the other client computers 16-17, 19-21 may be similarly configured. As further shown in the drawing, one or more of the client computers of the illustrated embodiment can include monitors for display of images rendered in accord with the teachings hereof. The client computers may also include other output devices, instead or in addition, for display of rendered images (e.g., plotters, printers, and so forth), as well as keyboards, mice and other input devices.


Preferably, server digital data processor 11 is constructed and operated in the manner of server 11 illustrated and described in the Related Application (the teachings of which are incorporated herein by reference), as further adapted in accord with the teachings hereof. This includes, by way of non-limiting example, the construction and operations shown and discussed in FIGS. 2 and 13-17 and the accompanying text of the Related Application, which figures and text are also incorporated herein by reference. Thus, for example, a preferred render server 11 includes one or more host systems, each equipped with one or more central processing units (CPUs) that are coupled to host memory and to one or more graphics (GPU) boards. Each graphics board can, in turn, include on-board memory (referred to as graphics memory) and a graphics processing unit (GPU). In order to render an image, e.g., in response to a messaged request from a client computer, a CPU on the server 11 (i) causes a respective portion of a data set to be transferred from host memory (or an external storage device) to the graphics memories on the server and (ii) issues commands to one or more of the GPUs on the server. The resulting image generated by the GPU(s) on the server is transferred to host memory on the server and, then, via network interfaces 39, 40, to the requesting client computer. A further appreciation of these and other aspects of operation of a preferred render server 11 may be attained by reference to aforementioned incorporated-by-reference Related Application.


It will be appreciated that the system 10 of FIG. 1 illustrates just one configuration of digital data processing devices with which the invention may be practiced. Other embodiments may, for example, utilize greater or fewer numbers of client computers, networks, networking apparatus (e.g., switches or hubs) and so forth. Moreover, it will be appreciated that the invention may be practiced with additional server digital data processors. Still further, it will be appreciated that the server digital data processor 11 may, itself, function, at least in part, in the role of a client computer (e.g., generating and servicing its own requests and/or generating requests for servicing by other computers) and vice versa.


Server Software and Client Software


Operation of the system 10 of the illustrated embodiment in regards relevant hereto is controlled by software running on render server 11 (“Server Software”) and software running on one or more of the client computers 16-21 (“Client Software”), e.g., client computer 18.


The Client Software handles data processing tasks, image rendering, network communication with render server 11 and client-side display, as discussed elsewhere herein. The make-up of Client Software of the illustrated embodiment is shown, by way of example, in the break-out of FIG. 1 and its operation is detailed in the sections that follow. As illustrated, the Client Software 18e includes an operating system 18e1, a graphics subsystem 18e2 (optionally, among other subsystems) and a graphics application 18e3 (optionally, among other applications). Although the discussion in the sections of text that follow largely focuses on operation of the Client Software 18e of client computer 18, it will be appreciated the Client Software of one or more of the other client computers 16-17, 19-21 may be similarly constructed and operated.


The operating system 18e1 is constructed and operated in the conventional manner of operating systems of client devices of the type shown herein, as adapted in accord with the teachings hereof.


The graphics application 18e3 provides an interface by which the user interacts with a data set that he/she wishes to visualize and/or otherwise process. This includes, for example, allowing the user to choose a data set, to choose render parameters such as color or data window or the view point or camera position when visualizing (e.g., rotating) the data set. In these regards, the graphics application 18e3 operates in the manner of conventional graphics applications of the type known in the art.


The graphics subsystem 18e2 is responsible for handling image rendering requests generated by the graphics application 18e3 and functions at the interface between that application and the client computer's operating system 18e1 and hardware. In the illustrated embodiment, it includes a graphics library 18e2a with functions that are invoked directly and indirectly by the graphics application's rendering requests. In the foregoing regards, the graphics subsystem 18e2 can be constructed and operated in the conventional manner of graphics subsystems known in the art, as adapted in accord with the teachings hereof.


The illustrated graphics subsystem 18e2 also includes a render module 18e2b that is operated in accord with the teachings hereof and that effects processing of requests made by the graphics application 18e3 such that requests to render some aspects of a given image are resolved (i.e., rendered) using the local processing resources (such as, by way of example, CPU 18a, graphics processing unit 18d and/or graphics library 18e2a) while those for other aspects of that same image are resolved by messaging the render server for rendering or other processing by it. In the discussion that follows, operations generally attributed to the “Client Software” (e.g., of client computer 18) are performed by the render module 18e2b, unless otherwise evident from context.


The Server Software operates in connection with the Client Software, e.g., 18e, running on a client computer, e.g., 18, to render or otherwise process data sets designated by a user of that client computer. Thus, as the user interacts with a data set (and, more particularly, for example, requests rendering of a data set), the Client Software, e.g., 18e, (and, more particularly, the render module 18e2b) on the respective client computer, e.g., 18, sends messages to the Server Software which, in turn, generate images, partial images or image data and returns them to the client computer (and, more particularly, to the render module 18e2b) for display and/or further processing. In addition to performing these rendering operations, the Server Software oversees network communication, data management, and other data processing tasks (for example, filtering).


Consistent with the remarks above, the Server Software is constructed and operated in the manner described in the incorporated-by-reference Related Application, as further adapted in accord with the teachings hereof. This includes, by way of non-limiting example, the construction and operations shown and discussed in FIGS. 2 and 13-17 and the accompanying text of the Related Application, which figures and text are also incorporated herein by reference.


Thus, though not a requirement of the invention, the Client Software, e.g., 18e, and Server Software of the illustrated embodiment can cooperate in the manner described in incorporated-by-reference Related Application and, more particularly, by way of non-limiting example, in the manner described in connection with FIG. 13 and the accompanying text thereof, which figures and text are also incorporated herein by reference. Thus, for example, the Server Software can listen for incoming network connections and, when a client computer attempts to make such a connection, can exchange authentication credentials and can check whether sufficient resources are available on the server 11 before accepting or rejecting the connection. Once a network connection is established, the Server Software can listen for incoming messages. These may include: (i) requests for a list of data sets available on the server-potentially along with some filter criteria, (ii) requests to load a data set for subsequent rendering, (iii) requests to render a data set with specified rendering parameters and a specified resolution level, (iv) requests to terminate a given connection, (v) requests to apply a filter (for example noise removal or sharpening) etc. Received messages can be processed immediately or can be added to a queue for later processing. In the typical case in which a client computer sends a render request, the Server Software can transfer the data set in question (or portions of it) into graphics memories on the server 11, can issue commands to GPUs of the server 11 to create a rendered image in those graphics memories, and can transfer the rendered image back into host memory on server 11 for subsequent processing and network transfer back to the requesting client computer. The Server Software can, furthermore, prioritize requests added to the queue of pending requests, alter requests in the queue (e.g., remove requests which are obsoleted, break down requests into multiple smaller ones), and/or determine resources used to process a request. Though not a requirement of the instant invention, the Server Software can service such rendering requests from multiple client computers concurrently. As noted, the Client Software, e.g., 18e, and Server Software of other embodiments may cooperate other than in the manner described above.


The Client Software, e.g., 18e, and, particularly, for example, its render module 18e2b, can improve rendering speed and application responsiveness by rendering some image aspects locally and messaging the server 11 to render (or otherwise process) other image aspects. For example, in response to requests by the graphics application 18e3 executing on a client computer, e.g., 18, to render aspects of an image of the type shown in FIG. 2—an image that includes an image aspect including acquired or synthesized images and an overlay graphics aspect representing textual or other graphical elements to be presented with the acquired/synthesized image—the render module 18e2b can message the server 11 to render the image aspect and can utilize the CPU and/or GPU of the respective client computer, e.g., 18, to render the overlay graphics aspect. Alternatively, the render module 18e2b can cause such render requests by the graphics application 18e3 to be processed entirely locally or, still further, can message the render server to effect rendering entirely on that server. Thus, over the course of multiple user requests and, in turn, multiple image rendering requests by the graphics application 18e3 (i.e., in response to those user requests), the Client Software, e.g., 18e, running on a given client computer, e.g., 18, effects rendering of at least one aspect of at least one of those images on the local processing resources (such as, by way of example, CPU 18a, graphics processing unit 18d and/or graphics library 18e2a) and messages the server 11 to render (or otherwise process) the other aspect(s) of those images.


In the illustrated embodiment, decisions on whether to resolve given render requests locally (i.e., to use local resources to render an aspect of an image in response to a given request by an application) or to message the render server (i.e., to render or otherwise process an aspect of an image in response to the given request) are generally made on a request-by-request basis. However, as will be appreciated, even a single render request can result in rendering using both local resources and the render server.


More particularly, decisions on whether and how to divide responsibility for rendering (e.g., between the local resources and the render server) are made according to which (i.e., the local resources or the render server) is better suited for the requisite rendering tasks, e.g., at the point in time rendering is requested.


To these ends, the render module 18e2b has access to the internal state of the graphics application 18e3 (e.g., as discerned from the calls it makes to the aforementioned graphics library 18e2a), as well as other information necessary to determine how to allocate rendering and compute tasks between the respective client computer (e.g., 18) and the render server 11, e.g., so as to avoid inefficient utilization of the server on account, for example, of unnecessary network roundtrips. That “other information” includes, for example, (i) the capabilities of the local processing resources (such as, by way of example, CPU 18a, graphics processing unit 18d and/or graphics library 18e2a) of the client computer, e.g., 18, itself, (ii) the load on those resources, (iii) the throughput of the network connecting the client and server computers, (iv) the image rendering capabilities of the render server, (v) the load on the render server, (vi) the locale of data being rendered. The latter (e.g., the capabilities of, and load on, the network and/or render server) may be communicated by the render server to the client computer's render module 18e2b, e.g., in response to a query made by the latter, and/or may be discerned by the render module 18e2b based on the rapidity with which the render server responds to image-rendering messages submitted to it by the client computer.


By way of brief digression, FIG. 2 is an example of an image—here, an image that forms part of a user interface 28—of the type generated by the graphics application 18e3 on a client computer, e.g., 18, with rendering by the local processing resources and/or the render server. That user interface 28 includes an image 30, shown here as a central frame within the user interface. The image includes an image aspect comprising an MPR image of a patient's legs 32. It also includes an overlay graphics aspect comprising patient data text 34, a scout rectangle 36 for designating cropping regions, a measurement line 38, and an elliptical annotation 40.


Of course, the invention is not limited to use with images that form user interfaces, nor with images of the type shown in FIG. 2, nor to images having two aspects. Indeed, in the latter regard, the invention can be used with images that have only a single aspect or that have three or more aspects. Without loss of generality and for sake of convenience an image having two or more aspects (e.g., image 30 of FIG. 2) is sometimes referred to herein as a “composite image” and, at other times, simply, as an “image.”



FIG. 3 is a data-flow diagram depicting one example of how the Client Software and, particularly, the render module 18e2b, can allocate rendering in response to requests by application 18e3 in regard to a (composite) image 30 of the type shown in FIG. 2. The particular allocation shown in FIG. 3A can be made, for example, on grounds that (a) the image aspect is relatively unchanging static (barring, for example, a new “slice” selection by the user) and is based on data already present on the server, and (b) the overlay graphics aspect is dynamic (e.g., changing as the user alters UT elements by moving the mouse, strikes keys on the keyboard, etc.).



FIG. 3, which shows time proceeding from left to right, depicts the Client Software and, particularly, the render module 18e2b, messaging the server 11 with a render request (see, step 50) in response to a render request pertaining to image aspects of a composite image by application 18e3. In response, the server 11 queues the request (step 54), computes the respective image slice (step 58) and renders the image aspect (step 60), which it subsequently returns to the render module 18e2b (step 62), which can paste the image into a back buffer (not shown) or other storage area. Having once received the image aspect rendered by the server, the render module 18e2b renders the overlay graphics aspect (step 64) of the composite image, e.g., onto the back buffer, again, in response to a render request by the application 18e3, and displays the resulting back buffer image (step 66) and/or makes it available to the Client Software (e.g., application 18e3) for further processing. When adding the overlay graphics aspect to the back buffer, the render module 183e2b can insure that those graphics match the state of the application 18e3 (and, specifically, for example, the state of the image aspects) at the time server was messaged to render the image aspects.


As indicated in FIG. 3, after the resulting image is rendered, the render module 18e2b can re-render the overlay graphics aspects, e.g., in response to the user moving the mouse, striking keys on the keyboard or otherwise taking steps that effect the overlay graphics, without calling on the server 11 to re-generate the image aspects. (See, the notation “start here if ‘overlay graphics’ changed.”) Conversely, if the user selects another slice to render, the render module 18e2b can re-message the server 11 to re-render the image aspect of the composite image and can re-render the overlay graphics aspects (See the notation “start over if slice orientation changed.”)



FIGS. 4A-4B are data-flow diagrams of another example of how the Client Software and, particularly, the render module 18e2b, can allocate rendering in response to requests by application 18e3 in regard to a composite image 30 of the type shown in FIG. 2—here, in connection with an image whose perspective aspect (i.e., zoom factor, panning position, or window/level setting) is changed, e.g., by user request.


As indicated by their use of similar elemental designations, FIGS. 4A-4B generally depict a data-flow like that shown in FIG. 3 and discussed above. The particular allocation of labor (as between the local resources and the render server) reflected in FIGS. 4A-4B is based on the grounds discussed above in connection with FIG. 3, as well as on the potentially changing perspective aspect of the composite image being rendered. Particularly, FIG. 4A shows such an allocation where the application 18e3 makes less frequent changes to zoom factor, panning position, window/level setting, or other perspective aspects. FIG. 4B, on the other hand, shows the allocation where the application 18e3 makes more frequent changes.


Referring to FIG. 4A, the Client Software and, particularly, the render module 18e2b, messages the server 11 with render requests (see, steps 50, 52) in response to render request pertaining to image and perspective aspects of the composite image by application 18e3. In response, the server 11 queues the requests (steps 54, 56), computes the respective image slice (step 58) and renders the image and perspective aspects (step 60), which it subsequently returns to the render module 18e2b (step 62) for temporary storage in a back buffer, or otherwise. Having once received the image aspect rendered by the server, the render module 18e2b renders the overly graphics aspect (step 64), e.g., on top of the back buffer, again, in response to a render request by the application 18e3, and displays the resulting image (step 66) and/or makes it available to the Client Software (e.g., application 18e3) for further processing. As above, when adding the overlay graphics aspect to the back buffer, the render module 183e2b can insure that those graphics match the state of the application 18e3 (and, specifically, for example, the state of the image aspects) at the time server was messaged to render the image aspects.


As above, after the resulting image is rendered, the render module 18e2b can re-render the overlay graphics aspects, e.g., in response to the user moving the mouse, striking keys on the keyboard or otherwise taking steps that effect the overlay graphics, without calling on the server 11 to re-render the image and perspective aspects. (See, the notation “start here if overlay graphics changed.”). Conversely, in response if the user selects another orientation (e.g., another slice to render and/or changes the view perspective), the render module 18e2b re-messages the server 11 to re-render both the image and perspective aspects (See, the notation “start over if slice orientation changed.”)


If the user repeatedly modifies the perspective, the render module 18e2b can allocate processing of both the perspective aspect and the overlay graphics aspect to the local processing resources to avoid repeated network roundtrips. This is reflected in FIG. 4B.


To this end, and with reference to FIG. 4B, before the given client computer, e.g., 18, can process the perspective aspect, it must request the slice from server 11. Once having obtained the image slice, the client computer can process perspective aspects and overlay graphics aspects for that image slice. Only, if the slice orientation changes, does the render module 18e2b have to message server 11 to compute the new slice; otherwise, the render module 18e2b can perform perspective and/or overlay graphics rendering using image slice data previously provided by the server.


More particularly, referring to FIG. 4B, the Client Software and, particularly, the render module 18e2b, messages the server 11 with a render request (see, step 50) in response to image render requests by application 18e3. In response, the server 11 queues the request (step 54), computes the respective image slice (see step 58) and sends it back to the client (step 61). Having once received the slice generated by the server 11, the render module 18e2b renders the corresponding image aspect from the perspective (e.g., zoom, pan, window level setting) chosen by the user or otherwise requested by the application 18e3 (step 63). The render module 18e2b, further, renders the overly graphics aspect (step 64), again, in response to a render request by the application 18e3, and displays the resulting image (step 66).


As indicated in FIG. 4, after the resulting image is rendered, the render module 18e2b can re-render the perspective and overlay graphics aspects, e.g., in response to the user (and application 18e3) changing the zoom, pan, window level-setting, without calling on the server 11 to re-generate the slice. (See the notation “start here if zoom, pan, w/l changed”). Moreover, the render module 18e2b can re-render the overlay graphics aspects, e.g., in response to the user (and application 18e3) moving the mouse, striking keys on the keyboard or otherwise taking steps that effect the overlay graphics, without calling on the server 11 to regenerate the slice or using local resources to re-render the image aspects. Conversely, in response if the user selects another slice to render, the render module 18e2b can re-message the server 11 to re-render the slice and, thereby, to restart the rendering process (See the notation “start over if slice orientation changed.”)


Though FIGS. 3 and 4B illustrates rendering of image aspects on the server 11 and overlay graphics and/or perspective aspects on a given client computer, e.g., 18, it will be appreciated that the render module 18e2b can cause the entirety of aspects of an image to be rendered on the server (or, conversely, on the client), e.g., as illustrated in FIG. 5. Such might be the case, by way of non-limiting example, in rendering a movie, video or other image sequence including both image aspects and overlay graphics aspects, where the resulting file is to be stored on the server, possibly, for transfer to other servers. In such an instance, by way of non-limiting example, the Client Software and, particularly, the render module 18e2b, can message the server 11 with a render request (see, step 70) in response to a request from application 18e3, e.g., to render an image from the movie sequence. In response, the server 11 queues the request (step 72), computes the respective image slice (step 74), renders the slice image (step 76) and the overlay (step 78), stores the resulting image (step 80) and sends a copy back to the client for display (step 82). At the same time, the server can send a copy of the image to a DICOM node for archival storage (step 84).


Described above are methods and systems meeting the desired objects, among others. It will be appreciated that the embodiments shown and described herein are merely examples of the invention and that other embodiments, incorporating changes therein may fall within the scope of the invention.

Claims
  • 1. A method for rendering images comprising: receiving at a server a first rendering request for a first image from a client computer at a first time, the server comprising a server digital data processor with access to a graphics processing unit and access to a back buffer, with a render module running on the server digital data processor, and the client computer comprising a local processing resource, where the render module in response to the first rendering request:(a) divides the first rendering request for the first image into at least a first perspective aspect and a first overlay aspect;(b) allocates the rendering of the first overlay aspect to the client computer;(c) renders the first perspective aspect using the graphics processing unit;(d) receives the first overlay aspect from the client computer;(e) combines the first perspective aspect and the first overlay aspect to generate a first composite image;(f) compares the first overlay aspect to insure that the first composite image matches the first image requested; and(g) stores the first composite image in the back buffer, where the first composite image is adapted for display on a display device, where the back buffer is available to be accessed by the client computer.
  • 2. The method of claim 1, where the first overlay aspect is a textual element.
  • 3. The method of claim 2, where the textual element comprises patient data.
  • 4. The method of claim 1, where the first overlay aspect is a graphical element.
  • 5. The method of claim 4, where the graphical element is selected from the group consisting of a ruler, a registration and a cross-hair’.
  • 6. The method of claim 1, further comprising where the render module stores the first perspective aspect separately in the back buffer.
  • 7. The method of claim 1, where the render module compares the first overlay aspect to insure that the first composite image matches the first image requested at the first time.
  • 8. The method of claim 1, further comprising where the render module compares the first perspective aspect to insure that the first composite image matches the first image requested at the first time.
  • 9. The method of claim 1, further comprising where the render module archives the first composite image in response to the first rendering request.
  • 10. The method of claim 1, further comprising where the render module in response to the first rendering request sends the first composite image to the client computer.
  • 11. The method of claim 1, further comprising where the render module displays the first composite image.
  • 12. A method for rendering images comprising: receiving at a server a first rendering request for a first image from a first client computer at a first time, the server comprising a server digital data processor with access to a graphics processing unit and access to a back buffer, with a render module running on the server digital data processor, and the first client computer comprising a first local processing resource;receiving at the server a second rendering request for a second image from a second client computer at a second time, and the second client computer comprising a second local processing resource;where the render module in response to the first rendering request:(a) divides the first rendering request for the first image into at least a first perspective aspect and a first overlay aspect;(b) allocates the rendering of the first overlay aspect to the first client computer;(c) renders the first perspective aspect using the graphics processing unit;(d) receives the first overlay aspect from the first client computer;(e) combines the first perspective aspect and the first overlay aspect to generate a first composite image adapted for display on a first display device, where the first composite image is stored in the back buffer, where the back buffer is available to be accessed by the first client computer;(f) divides the second rendering request for the second image into at least a second perspective aspect and a second overlay aspect;(g) allocates the rendering of the second overlay aspect to the second client computer;(h) renders the second perspective aspect using the graphics processing unit;(i) receives the second overlay aspect from the second client computer; and(j) combines the second perspective aspect and the second overlay aspect to generate a second composite image adapted for display on a second display device, where the second composite image is stored in the back buffer, where the back buffer is available to be accessed by the second client computer.
  • 13. The method of claim 12, where the second overlay aspect is a textual element.
  • 14. The method of claim 13, where the textual element comprises patient data.
  • 15. The method of claim 12, where the second overlay aspect is a graphical element.
  • 16. The method of claim 15, where the graphical element is selected from the group consisting of a ruler, a registration and a cross-hair’.
  • 17. The method of claim 12, further comprising where the render module stores the second perspective aspect separately in the back buffer.
  • 18. The method of claim 12, further comprising where the render module compares the second overlay aspect to insure that the second composite image matches the second image requested at the second time.
  • 19. The method of claim 12, further comprising where the render module compares the second perspective aspect to insure that the second composite image matches the second image requested at the second time.
  • 20. The method of claim 12, further comprising where the render module in response to the second rendering request archives the second composite image.
Parent Case Info

This application is a continuation of (1) U.S. application Ser. No. 17/574,986 entitled Client-Server Visualization System with Hybrid Data Processing, filed Jan. 13, 2022, which issued as U.S. Pat. No. 11,640,809 on May 2, 2023, which claims priority to (2) U.S. application Ser. No. 16/986,020 entitled Client-Server Visualization System with Hybrid Data Processing, filed Aug. 5, 2020, which issued as U.S. Pat. No. 11,244,650 on Feb. 8, 2022, which claims priority to (3) U.S. application Ser. No. 16/504,146 entitled Client-Server Visualization System with Hybrid Data Processing, filed Jul. 5, 2019, which issued as U.S. Pat. No. 10,762,872 on Sep. 1, 2020, which claims priority to (4) U.S. application Ser. No. 16/036,451 entitled Client-Server Visualization System with Hybrid Data Processing, filed Jul. 16, 2018, which issued as U.S. Pat. No. 10,380,970 on Aug. 13, 2019, which claims priority to and is a continuation of (5) U.S. application Ser. No. 15/450,888 entitled Client-Server Visualization System with Hybrid Data Processing, filed Mar. 6, 2017, which issued as U.S. Pat. No. 10,043,482 on Aug. 7, 2018, which claims priority to and is a continuation of (6) U.S. application Ser. No. 14/641,243 entitled Client-Server Visualization System with Hybrid Data Processing, filed Mar. 6, 2015, which issued as U.S. Pat. No. 9,595,242 on Mar. 14, 2017, which claims priority to and is a continuation of (7) U.S. application Ser. No. 12/275,834 entitled Client-Server Visualization System with Hybrid Data Processing, filed Nov. 21, 2008 which issued as U.S. Pat. No. 9,019,287 on Apr. 28, 2015, which claims the benefit of priority of (8) U.S. Provisional Patent Application Ser. No. 60/989,881 filed Nov. 23, 2007 and (9) U.S. Provisional Patent Application Ser. No. 60/989,913 filed Nov. 23, 2007, the teachings of each of (1)-(9) are incorporated herein by reference in their entireties and for all purposes. The invention pertains to digital data processing and, more particularly, by way of example, to the visualization of 3D and 4D image data. It has application to areas including, by way of non-limiting example, medical imaging, atmospheric studies, astrophysics, and geophysics. 3D and 4D image data is routinely acquired with computer tomographic scanners (CT), magnetic resonance imaging scanners (MRI), confocal microscopes, 3D ultrasound devices, positron emission tomographic scanners (PET) and other imaging devices. Medical imaging is just one example of a market that uses these devices. It is growing rapidly, with new CT scanners, for example, collecting ever greater amounts of data even more quickly than previous generation scanners. As this trend continues across many markets, the demand for better and faster visualization methods that allow users to interact with the image data in real-time will increase. Standard visualization methods fall within the scope of volume rendering techniques (VRT), shaded volume rendering techniques (sVRT), maximum intensity projection (MIP), oblique slicing or multiplanar reformats (MPR), axial/sagittal and coronal slice display, and thick slices (also called slabs). In the following, these and other related techniques are collectively referred to as “volume rendering.” In medical imaging, for example, volume rendering is used to display 3D images from 3D image datasets, where a typical 3D image dataset is a large number of 2D slice images acquired by a CT or MRI scanner and stored in a data structure. The rendition of such images can be quite compute intensive and therefore takes a long time on a standard computer, especially, when the data sets are large. Too long compute times can, for example, prevent the interactive exploration of data sets, where a user wants to change viewing parameters, such as the viewing position interactively, which requires several screen updates per second (typically 5-25 updates/second), thus requiring rendering times of fractions of a second or less per image. Several approaches have been taken to tackle this performance problem. For example, special-purchase chips have been constructed to implement volume rendering in hardware. Another approach is to employ texture hardware built into high-end graphics workstations or graphics super-computers, such as for example Silicon Graphics Onyx computers with Infinite Reality and graphics. More recently, standard graphics boards, such as NVIDIA's Geforce and Quadro FX series, as well as AMD/ATI's respective products, are also offering the same or greater capabilities as far as programmability and texture memory access are concerned. Typically hardware for accelerated volume rendering must be installed in the computer (e.g., workstation) that is used for data analysis. While this has the advantage of permitting ready visualization of data sets that are under analysis, it has several drawbacks. First of all, every computer which is to be used for data analysis needs to be equipped with appropriate volume-rendering hardware, as well as enough main memory to handle large data sets. Second the data sets often need to be transferred from a central store (e.g., a main enterprise server), where they are normally stored, to those local workstations prior to analysis and visualization, thus potentially causing long wait times for the user during transfer. Several solutions have been proposed in which data processing applications running on a server are controlled from a client computer, thus, avoiding the need to equip it with the full hardware needed for image processing/visualization and also making data transfer to the client unnecessary. Such solutions include Microsoft's Windows 2003 server (with the corresponding remote desktop protocol (RDP)), Citrix Presentation Server, VNC, or SGI's OpenGL Vizserver. However, most of these solutions do not allow applications to use graphics hardware acceleration. The SGI OpenGL Vizserver did allow hardware accelerated graphics applications to be run over the network: it allocated an InfiniteReality pipeline to an application controlled over the network. However that pipeline could then not be used locally any longer and was also blocked for other users. Thus effectively all that the Vizserver was doing was extending a single workplace to a different location in the network. The same is true for VNC. For general graphics applications (i.e., not specifically volume rendering applications), such as computer games, solutions have been proposed to combine two graphics cards on a single computer (i.e., the user's computer) in order to increase the rendering performance, specifically NVIDIA's SLI and AMD/ATI's Crossfire products. In these products, both graphics cards receive the exact same stream of commands and duplicate all resources (such as textures). Each of the cards then renders a different portion of the screen—or in another mode one of the cards renders every second image and the other card renders every other image. While such a solution is transparent to the application and therefore convenient for the application developers it is very limited too. Specifically the duplication of all textures effectively eliminates half of the available physical texture memory. An object of the invention is to provide digital data processing methods and apparatus, and more particularly, by way of example, to provide improved such methods and apparatus for visualization of image data. A further object of the invention is to provide methods and apparatus for rendering images. A still further object of the invention is to provide such methods and apparatus for rendering images as have improved real-time response to a user's interaction. Yet a still further object of the invention is to provide such methods and apparatus as allow users to interactively explore the rendered images.

US Referenced Citations (397)
Number Name Date Kind
2658310 Cook Nov 1953 A
3431200 Davis Mar 1969 A
3645040 Ort Feb 1972 A
4137868 Pryor Feb 1979 A
4235043 Harasawa et al. Nov 1980 A
4258661 Margen Mar 1981 A
4267038 Thompson May 1981 A
4320594 Raymond Mar 1982 A
4746795 Stewart et al. May 1988 A
4905148 Crawford Feb 1990 A
4910912 Lowrey, III Mar 1990 A
4928250 Greenberg et al. May 1990 A
4958460 Nielson et al. Sep 1990 A
4984160 Saint Felix et al. Jan 1991 A
5031117 Minor et al. Jul 1991 A
5091960 Butler Feb 1992 A
5121708 Nuttle Jun 1992 A
5128864 Waggener et al. Jul 1992 A
5218534 Trousset et al. Jun 1993 A
5235510 Yamada Aug 1993 A
5241471 Trousset et al. Aug 1993 A
5253171 Hsiao et al. Oct 1993 A
5274759 Yoshioka Dec 1993 A
5280428 Wu et al. Jan 1994 A
5287274 Saint Felix et al. Feb 1994 A
5293313 Cecil Mar 1994 A
5307264 Waggener et al. Apr 1994 A
5355453 Row et al. Oct 1994 A
5368033 Moshfeghi Nov 1994 A
5375156 Kuo-Petravic et al. Dec 1994 A
5412703 Goodenough et al. May 1995 A
5412764 Tanaka May 1995 A
5442672 Bjorkholm et al. Aug 1995 A
5452416 Hilton Sep 1995 A
5488700 Glassner Jan 1996 A
5560360 Filler Oct 1996 A
5594842 Kaufman et al. Jan 1997 A
5602892 Llacer Feb 1997 A
5633951 Moshfeghi May 1997 A
5633999 Clowes et al. May 1997 A
5640436 Kawai et al. Jun 1997 A
5671265 Andress Sep 1997 A
5744802 Muehllehner et al. Apr 1998 A
5774519 Lindstrom et al. Jun 1998 A
5790787 Scott et al. Aug 1998 A
5793374 Guenter et al. Aug 1998 A
5793879 Benn et al. Aug 1998 A
5813988 Alfano et al. Sep 1998 A
5821541 Tumer Oct 1998 A
5825842 Taguchi Oct 1998 A
5838756 Taguchi et al. Nov 1998 A
5841140 Mc Croskey et al. Nov 1998 A
5909476 Cheng et al. Jun 1999 A
5930384 Guillemaud et al. Jul 1999 A
5931789 Alfano et al. Aug 1999 A
5950203 Stakuis Sep 1999 A
5960056 Lai Sep 1999 A
5963612 Navab Oct 1999 A
5963613 Navab Oct 1999 A
5963658 Klibanov et al. Oct 1999 A
6002739 Heumann Dec 1999 A
6018562 Willson Jan 2000 A
6032264 Beffa et al. Feb 2000 A
6044132 Navab Mar 2000 A
6049390 Notredame Apr 2000 A
6049582 Navab Apr 2000 A
6072177 Mccroskey et al. Jun 2000 A
6088423 Krug et al. Jul 2000 A
6091422 Ouaknine et al. Jul 2000 A
6104827 Benn et al. Aug 2000 A
6105029 Maddalozzo, Jr. et al. Aug 2000 A
6108007 Shochet Aug 2000 A
6108576 Alfano et al. Aug 2000 A
6123733 Dalton Sep 2000 A
6175655 George Jan 2001 B1
6205120 Packer et al. Mar 2001 B1
6219061 Lauer et al. Apr 2001 B1
6226005 Laferriere May 2001 B1
6236704 Navab et al. May 2001 B1
6243098 Lauer et al. Jun 2001 B1
6249594 Hibbard Jun 2001 B1
6255655 Mc Croskey et al. Jul 2001 B1
6264610 Zhu Jul 2001 B1
6268846 Georgiev Jul 2001 B1
6278460 Myers et al. Aug 2001 B1
6282256 Grass et al. Aug 2001 B1
6289235 Webber et al. Sep 2001 B1
6304771 Yodh et al. Oct 2001 B1
6320928 Vaillant et al. Nov 2001 B1
6324241 Besson Nov 2001 B1
6377257 Borrel Apr 2002 B1
6377266 Baldwin Apr 2002 B1
6384821 Borrel May 2002 B1
6404843 Vaillant Jun 2002 B1
6415013 Hsieh et al. Jul 2002 B1
6470067 Harding Oct 2002 B1
6470070 Menhardt Oct 2002 B2
6473793 Dillon et al. Oct 2002 B1
6475150 Haddad Nov 2002 B2
6507633 Elbakri et al. Jan 2003 B1
6510241 Vaillant et al. Jan 2003 B1
6519355 Nelson Feb 2003 B2
6526305 Mori Feb 2003 B1
6557102 Wong et al. Apr 2003 B1
6559958 Motamed May 2003 B2
6591004 VanEssen et al. Jul 2003 B1
6615063 Ntziachristos et al. Sep 2003 B1
6633688 Nixon Oct 2003 B1
6636623 Nelson et al. Oct 2003 B2
6654012 Lauer et al. Nov 2003 B1
6658142 Kam et al. Dec 2003 B1
6664963 Zatz Dec 2003 B1
6674430 Kaufman et al. Jan 2004 B1
6697508 Nelson Feb 2004 B2
6707878 Claus et al. Mar 2004 B2
6718195 Van Der Mark et al. Apr 2004 B2
6731283 Navab May 2004 B1
6740232 Beaulieu May 2004 B1
6741730 Rahn et al. May 2004 B2
6744253 Stolarczyk Jun 2004 B2
6744845 Harding et al. Jun 2004 B2
6745070 Wexler et al. Jun 2004 B2
6747654 Laksono et al. Jun 2004 B1
6754299 Patch Jun 2004 B2
6765570 Cheung Jul 2004 B1
6765981 Heumann Jul 2004 B2
6768782 Hsieh et al. Jul 2004 B1
6770893 Nelson Aug 2004 B2
6771733 Katsevich Aug 2004 B2
6778127 Stolarczyk et al. Aug 2004 B2
6785409 Suri Aug 2004 B1
6798417 Taylor Sep 2004 B1
6807581 Starr et al. Oct 2004 B1
6825840 Gritz Nov 2004 B2
6825843 Allen et al. Nov 2004 B2
6923906 Oswald et al. Aug 2005 B2
6947047 Moy et al. Sep 2005 B1
6978206 Pu Dec 2005 B1
7003547 Hubbard Feb 2006 B1
7006101 Brown et al. Feb 2006 B1
7031022 Komori et al. Apr 2006 B1
7034828 Drebin et al. Apr 2006 B1
7039723 Hu May 2006 B2
7050953 Chiang et al. May 2006 B2
7054852 Cohen May 2006 B1
7058644 Patchet et al. Jun 2006 B2
7076735 Callegari Jul 2006 B2
7098907 Houston et al. Aug 2006 B2
7120283 Thieret Oct 2006 B2
7133041 Kaufman et al. Nov 2006 B2
7154985 Dobbs Dec 2006 B2
7167176 Sloan et al. Jan 2007 B2
7184041 Heng et al. Feb 2007 B2
7185003 Bayliss et al. Feb 2007 B2
7219085 Buck et al. May 2007 B2
7242401 Yang et al. Jul 2007 B2
7262770 Sloan et al. Aug 2007 B2
7274368 Keslin Sep 2007 B1
7299232 Stakutis et al. Nov 2007 B2
7315926 Fridella et al. Jan 2008 B2
7324116 Boyd et al. Jan 2008 B2
7339585 Verstraelen et al. Mar 2008 B2
7472156 Philbrick et al. Dec 2008 B2
7502869 Boucher et al. Mar 2009 B2
7506375 Kanda et al. Mar 2009 B2
7552192 Carmichael Jun 2009 B2
7609884 Stalling Oct 2009 B1
7693318 Stalling Apr 2010 B1
7701210 Ichinose Apr 2010 B2
7778392 Berman Aug 2010 B1
7876944 Stalling Jan 2011 B2
7889895 Nowinski Feb 2011 B2
7899516 Chen et al. Mar 2011 B2
7907759 Hundley Mar 2011 B2
7956612 Sorensen Jun 2011 B2
7983300 Vaughan et al. Jul 2011 B2
7991837 Tahan Aug 2011 B1
7995824 Yim Aug 2011 B2
8107592 Berman Jan 2012 B2
8189002 Westerhoff May 2012 B1
8319781 Westerhoff Nov 2012 B2
8369600 Can et al. Feb 2013 B2
8386560 Ma Feb 2013 B2
8392529 Westerhoff Mar 2013 B2
8508539 Vlietinck Aug 2013 B2
8538108 Shekhar Sep 2013 B2
8542136 Owsley et al. Sep 2013 B1
8548215 Westerhoff Oct 2013 B2
8701167 Kovalan Apr 2014 B2
8745413 Hahn Jun 2014 B2
8775510 Westerhoff Jul 2014 B2
8976190 Westerhoff Mar 2015 B1
9019287 Westerhoff Apr 2015 B2
9106609 Kovalan Aug 2015 B2
9167027 Westerhoff Oct 2015 B2
9299156 Zalis Mar 2016 B2
9355616 Westerhoff May 2016 B2
9438667 Kovalan Sep 2016 B2
9454813 Westerhoff Sep 2016 B2
9509802 Westerhoff Nov 2016 B1
9524577 Westerhoff Dec 2016 B1
9531789 Westerhoff Dec 2016 B2
9595242 Westerhoff Mar 2017 B1
9728165 Westerhoff Aug 2017 B1
9749245 Stalling Aug 2017 B2
9749389 Kovalan Aug 2017 B2
9860300 Westerhoff Jan 2018 B2
9898855 Westerhoff Feb 2018 B2
9904969 Westerhoff Feb 2018 B1
9984460 Westerhoff May 2018 B2
9984478 Westerhoff May 2018 B2
10038739 Westerhoff Jul 2018 B2
10043482 Westerhoff Aug 2018 B2
10070839 Westerhoff Sep 2018 B2
10311541 Westerhoff Jun 2019 B2
10320684 Stalling Jun 2019 B2
10373368 Westerhoff Aug 2019 B2
10380970 Westerhoff Aug 2019 B2
10395398 Westerhoff Aug 2019 B2
10430914 Westerhoff Oct 2019 B2
10448911 Erhard Oct 2019 B2
10452813 Sorenson Oct 2019 B2
10540803 Westerhoff Jan 2020 B2
10614543 Westerhoff Apr 2020 B2
10631812 Westerhoff Apr 2020 B2
10686868 Westerhoff Jun 2020 B2
10706538 Westerhoff Jul 2020 B2
10726955 Kovalan Jul 2020 B2
10762872 Westerhoff Sep 2020 B2
10818048 Zhao Oct 2020 B2
10867011 Sorenson Dec 2020 B2
10930397 Kovalan Feb 2021 B2
10970365 Sorenson Apr 2021 B2
10978184 Sorenson Apr 2021 B2
11075978 Westerhoff Jul 2021 B2
11183292 Stalling Nov 2021 B2
11315210 Westerhoff Apr 2022 B2
11328381 Westerhoff May 2022 B2
11514572 Westerhoff Nov 2022 B2
11599672 Stalling Mar 2023 B2
11620773 Westerhoff Apr 2023 B2
11640809 Westerhoff May 2023 B2
11763516 Westerhoff Sep 2023 B2
20010026848 Van Der Mark Oct 2001 A1
20020010551 Wang et al. Jan 2002 A1
20020016813 Woods et al. Feb 2002 A1
20020034817 Henry et al. Mar 2002 A1
20020049825 Jewett et al. Apr 2002 A1
20020080143 Morgan et al. Jun 2002 A1
20020089587 White et al. Jul 2002 A1
20020099290 Haddad Jul 2002 A1
20020099844 Baumann et al. Jul 2002 A1
20020120727 Curley et al. Aug 2002 A1
20020123680 Vaillant Sep 2002 A1
20020138019 Wexler Sep 2002 A1
20020150202 Harding Oct 2002 A1
20020150285 Nelson Oct 2002 A1
20020180747 Lavelle et al. Dec 2002 A1
20020184238 Chylla Dec 2002 A1
20020184349 Manukyan Dec 2002 A1
20030001842 Munshi Jan 2003 A1
20030031352 Nelson et al. Feb 2003 A1
20030059110 Wilt Mar 2003 A1
20030065268 Chen et al. Apr 2003 A1
20030086599 Armato May 2003 A1
20030103666 Edic et al. Jun 2003 A1
20030120743 Coatney et al. Jun 2003 A1
20030123720 Launay et al. Jul 2003 A1
20030149812 Schoenthal et al. Aug 2003 A1
20030158786 Yaron Aug 2003 A1
20030176780 Arnold Sep 2003 A1
20030179197 Sloan et al. Sep 2003 A1
20030194049 Claus et al. Oct 2003 A1
20030220569 Dione Nov 2003 A1
20030220772 Chiang et al. Nov 2003 A1
20030227456 Gritz Dec 2003 A1
20030234791 Boyd et al. Dec 2003 A1
20040010397 Barbour et al. Jan 2004 A1
20040012596 Allen et al. Jan 2004 A1
20040015062 Ntziachristos et al. Jan 2004 A1
20040022348 Heumann Feb 2004 A1
20040059822 Jiang Mar 2004 A1
20040066384 Ohba Apr 2004 A1
20040066385 Kilgard Apr 2004 A1
20040066891 Freytag Apr 2004 A1
20040078238 Thomas et al. Apr 2004 A1
20040101183 Mullick et al. May 2004 A1
20040102688 Walker May 2004 A1
20040125103 Kaufman Jul 2004 A1
20040133652 Miloushev et al. Jul 2004 A1
20040147039 Van Der Mark Jul 2004 A1
20040162677 Bednar Aug 2004 A1
20040170302 Museth et al. Sep 2004 A1
20040193901 Bharara Sep 2004 A1
20040210584 Nir et al. Oct 2004 A1
20040215858 Armstrong et al. Oct 2004 A1
20040215868 Solomon et al. Oct 2004 A1
20040239672 Schmidt Dec 2004 A1
20040240753 Hu Dec 2004 A1
20050012753 Karlov Jan 2005 A1
20050017972 Poole et al. Jan 2005 A1
20050066095 Mullick et al. Mar 2005 A1
20050088440 Sloan et al. Apr 2005 A1
20050128195 Houston et al. Jun 2005 A1
20050152590 Thieret Jul 2005 A1
20050165623 Landi et al. Jul 2005 A1
20050225554 Bastos et al. Oct 2005 A1
20050228272 Yu Oct 2005 A1
20050231503 Heng et al. Oct 2005 A1
20050239182 Berzin Oct 2005 A1
20050240628 Jiang et al. Oct 2005 A1
20050256742 Kohan et al. Nov 2005 A1
20050259103 Kilgard et al. Nov 2005 A1
20050270298 Thieret Dec 2005 A1
20050271302 Khamene et al. Dec 2005 A1
20060010438 Brady et al. Jan 2006 A1
20060010454 Napoli et al. Jan 2006 A1
20060028479 Chun Feb 2006 A1
20060034511 Verstraelen Feb 2006 A1
20060066609 Iodice Mar 2006 A1
20060149850 Bowman Jul 2006 A1
20060197780 Watkins et al. Sep 2006 A1
20060214949 Zhang Sep 2006 A1
20060239540 Serra Oct 2006 A1
20060239589 Omernick Oct 2006 A1
20060282253 Buswell et al. Dec 2006 A1
20070005798 Gropper et al. Jan 2007 A1
20070038939 Challen Feb 2007 A1
20070046966 Mussack Mar 2007 A1
20070067497 Craft et al. Mar 2007 A1
20070092864 Reinhardt Apr 2007 A1
20070097133 Stauffer et al. May 2007 A1
20070116332 Cai et al. May 2007 A1
20070127802 Odry Jun 2007 A1
20070156955 Royer, Jr. Jul 2007 A1
20070165917 Cao et al. Jul 2007 A1
20070185879 Roublev et al. Aug 2007 A1
20070188488 Choi Aug 2007 A1
20070226314 Eick et al. Sep 2007 A1
20070255704 Baek et al. Nov 2007 A1
20070280518 Nowinski Dec 2007 A1
20080009055 Lewnard Jan 2008 A1
20080042923 De Laet Feb 2008 A1
20080052126 Sasai Feb 2008 A1
20080086557 Roach Apr 2008 A1
20080115139 Inglett et al. May 2008 A1
20080137929 Chen et al. Jun 2008 A1
20080147554 Stevens et al. Jun 2008 A1
20080155890 Oyler Jul 2008 A1
20080174593 Ham Jul 2008 A1
20080208961 Kim et al. Aug 2008 A1
20080224700 Sorensen Sep 2008 A1
20080281908 McCanne et al. Nov 2008 A1
20080317317 Shekhar Dec 2008 A1
20090005693 Brauner et al. Jan 2009 A1
20090012382 Dutta et al. Jan 2009 A1
20090043988 Archer et al. Feb 2009 A1
20090077097 Lacapra et al. Mar 2009 A1
20090147793 Hayakawa et al. Jun 2009 A1
20090208082 Westerhoff et al. Aug 2009 A1
20090210487 Westerhoff et al. Aug 2009 A1
20090225076 Vlietinck Sep 2009 A1
20090245610 Can et al. Oct 2009 A1
20090313170 Goldner et al. Dec 2009 A1
20100054556 Novatzky Mar 2010 A1
20100060652 Karlsson Mar 2010 A1
20100123733 Zaharia May 2010 A1
20100174823 Huang Jul 2010 A1
20100272342 Berman et al. Oct 2010 A1
20100278405 Kakadiaris et al. Nov 2010 A1
20110044524 Wang et al. Feb 2011 A1
20110110576 Kreeger et al. May 2011 A1
20110112862 Yu May 2011 A1
20120078088 Whitestone et al. Mar 2012 A1
20120226916 Hahn Sep 2012 A1
20120233153 Roman et al. Sep 2012 A1
20130176319 Westerhoff Jul 2013 A1
20130195329 Canda Aug 2013 A1
20140173287 Mizunuma Jun 2014 A1
20150213288 Bilodeau et al. Jul 2015 A1
20160012181 Massey Jan 2016 A1
20170011514 Westerhoff Jan 2017 A1
20170032546 Westerhoff Feb 2017 A1
20170098329 Westerhoff Apr 2017 A1
20170104811 Westerhoff Apr 2017 A1
20170178593 Westerhoff Jun 2017 A1
20170346883 Westerhoff Nov 2017 A1
20190096062 Westerhoff Mar 2019 A1
20190318512 Westerhoff Oct 2019 A1
20200327669 Westerhoff Oct 2020 A1
20200366614 Stalling Nov 2020 A1
20210133978 Ma May 2021 A1
20210256742 Westerhoff Aug 2021 A1
20210352133 Westerhoff Nov 2021 A1
20220165231 Westerhoff May 2022 A1
20230195937 Stalling Jun 2023 A1
20230260478 Westerhoff Aug 2023 A1
Foreign Referenced Citations (44)
Number Date Country
10317384 Apr 2004 DE
0492897 Jul 1992 EP
0502187 Sep 1992 EP
0611181 Aug 1994 EP
0476070 Aug 1996 EP
0925556 Jun 1999 EP
0953943 Nov 1999 EP
0964 366 Dec 1999 EP
187340 Mar 2001 EP
2098895 Sep 2009 EP
2098994 Sep 2009 EP
2405344 Jan 2012 EP
WO9016072 Dec 1990 WO
WO9102320 Feb 1991 WO
WO9205507 Apr 1992 WO
WO9642022 Dec 1996 WO
WO9810378 Mar 1998 WO
WO9812667 Mar 1998 WO
WO9833057 Jul 1998 WO
WO0120546 Mar 2001 WO
WO0134027 May 2001 WO
WO0163561 Aug 2001 WO
WO0174238 Oct 2001 WO
WO0185022 Nov 2001 WO
WO0241760 May 2002 WO
WO02067201 Aug 2002 WO
WO02082065 Oct 2002 WO
WO03061454 Jul 2003 WO
WO03088133 Oct 2003 WO
WO03090171 Oct 2003 WO
WO03098539 Nov 2003 WO
WO04019782 Mar 2004 WO
WO04020996 Mar 2004 WO
WO04020997 Mar 2004 WO
WO04034087 Apr 2004 WO
WO04044848 May 2004 WO
WO04066215 Aug 2004 WO
WO04072906 Aug 2004 WO
WO05071601 Aug 2005 WO
WO09029636 Mar 2009 WO
WO09067675 May 2009 WO
WO09067680 May 2009 WO
WO11065929 Jun 2011 WO
WO2015063188 May 2015 WO
Non-Patent Literature Citations (99)
Entry
Higgins et al., Distributed System for Processing 3D Medical Images, Compat Biol Med (1997) 27, pp. 97-115.
AU2022200601, Office Action, dated Dec. 8, 2022, 4 pages.
AU2018335370, Office Action, dated May 15, 2023, 4 pages.
CA2991378, Office Action, dated Oct. 24, 2022, 7 pages.
CA2991378, Office Action, dated Oct. 20, 2023, 8 pages.
EP3329405, Summons, dated Jul. 7, 2023, 14 pages.
EP202303856_OA_Aug. 23, 2023, 7 pages.
JP2022-191827, Office Action, dated Oct. 18, 2023, 2 pages (& English translation).
ATI Website Index, http://www.ati.com/developer/index.html, Dec. 20, 2002, 2 pages.
Boone et al., Recognition of Chest Radiograph Orientation for Picture Archiving and Communications Systems Display Using Neural Networks, J. Digital Imaging, 1992, 5(3), 190-193.
Boone et al., Automated Recognition of Lateral from PA Chest Radiographs: Saving Seconds in a PACS Environment, J. Digital Imaging, 2003, 16(4), 345-349.
Luo et al., Automatic Image Hanging Protocol for Chest Radiographs in a PACS, IEEE Transactions on Information Technology in Biomedicine, 2006, 10(2), 302-311.
Cabral et al., Accelerated Volume Rendering and Tomographic Reconstruction Using Texture Mapping Hardware⋅, Silicon Graphics Computer Systems, 1995 IEEE, DD. 91-97.
Carr, Nathan A., Jesse D. Hall, John C. Hart, The ray engine, Proceedings of the ACM SIGGRAPH/EUROGRAPHICS conference on Graphics hardware, Sep. 1-2, 2002, pp. 37-46.
Chidlow, et al, Rapid Emission Tomography Reconstruction, Proceedings of the 2003 Eurographics/IEEE TVCG Workshop on Volume Graphics, Tokyo, Japan, Jul. 7-8, 2003, 13 pages.
Cohen, Michael, et al., A Progressive Refinement Approach to Fast Radiosity Image Generation, Computer Graphics, vol. 22, No. 4, Aug. 1988, pp. 75-84.
Corner, B., University of Nebraska-Lincoln, MatLat.txt, 2003, 1 page.
Dachille, et al., High-Quality Volume Rendering Using Texture Mapping Hardware, Siggraph/Eurographics Hardware Workshop (1998) (8 pages).
Dempster, et al., Maximum Likelihood From Incomplete Data via The EM Algorithm, Harvard University and Educational Testing Service, Dec. 8, 1976, pp. 1-38.
Dennis, C, et al.,, Overview of X-Ray Computed Tomography, http://www.howstuffworks.com/framed.htm?parent=c...tm&url=http://www.ctlab.geo.utexas.edu/overview/, Dec. 26, 2002, 5 pages.
Dobbins, et al., Digital X-Ray Tomosynthesis: Current State of the Art and Clinical Potential, Physics in Medicine and Biology, vol. 48, pp. R65-R106 (2003).
Doggett, Michael, ATI, Programmability Features of Graphics Hardware, (paper) Apr. 23, 2002, pp. C1-C22.
Doggett, Michael, ATI, Programmability Features of Graphics Hardware, (slideshow) slides 1-62 31 pages.
Du, H., Sanchez-Elez, M., Tabrizi, N., Bagherzadeh, N., Anido, M. L., and Fernandez, M. 2003. Interactive ray tracing on reconfigurable SIMD MorphSys. In Proceedings of the 2003 Conference on Asia South Pacific Design Automation (Kitakyushu, Japan, Jan. 21-24, 2003). ASPDAC. ACM, New York, NY, 471-476.
Eldridge Matthew, Homan Igehy, Pat Hanrahan, Pomegranate: a fully scalable graphics architecture, Proceedings of the 27th annual conference on Computer graphics and interactive techniques, p. 443-454, Jul. 2000.
Fang, L., et al., Fast Maximum Intensity Projection Algorithm Using Shear Warp Factorization and Reduced Resampling, Mangetic Resonance in Medicine 47:696-700 (2002).
Filtered Backprojection Reconstruction, http://www.physics.ubd.ca/-mirg/home/tutorial/fbDrecon.html, 216/2003, 5 pages.
Goddard et al., High-speed cone-beam reconstruction: an embedded systems approach, 2002, SPIE vol. 4681, pp. 483-491.
Grass et al., Three-dimensional reconstruction of high contrast objects using C-arm image intensifier projection data, 1999, Computerized Medical Imaging and Graphics, 23, pp. 311-321.
Hadwiger, Markus et al., Hardware-Accelerated High-Quality Reconstruction of Volumetric Data on PC Graphics Hardware, VRVis Research Center, Vienna, Austria, and Institute of Computer Graphics and Algorithms, Vienna University of Technology, Austria, 9 pages.
Hastreiter et al. (Integrated registration and visualization of medical image data, Proc. Computer Graphics International, Jun. 22-26, 1998, pp. 78-85).
Hopf, M. Ertl, T., Accelerating 3d Convolution Using Graphics Hardware, Proc. IEEE Visualization, 1999, 5 pages.
Hudson, et al., Accelerated Image Reconstruction Using Ordered Subsets of Projection Data, IEEE Transactions on Medical Imaging, vol. 13, No. 4, Dec. 1994, pp. 601-609.
Image Registration Slideshow, 105 pages.
Iterative definition, Merriam-Webster on-line dictionary, printed Aug. 26, 2010, 3 pages.
Jain, Anju, A Programmable Graphics Chip, pcquest.com, Jun. 18, 2001.
Jones et al., Positron Emission Tomogrphic Images and Expectation Maximization: A VLSI Architecture for Multiple Iterations Per Second, Computer Technology and Imaging, Inc., 1988 IEEE, pp. 620-624.
Kajiya, J. T., Ray tracing volume densities, Proc. Siggraph, Jul. 1984, Computer Graphics, vol. 18, No. 3, pp. 165-174.
Karlsson, Filip; Ljungstedt, Carl Johan; Ray tracing fully implemented on programmable graphics hardware, Master's Thesis, Chalmers University of Technology, Dept. of Computer Engineering, Goteborg, Sweden, copyright © 2004, 29 pages.
Kruger J. and R. Westermann, Acceleration Techniques for GPU-based Volume Rendering Proceedings of IEEE Visualization, 2003, 6 pages.
Lange et al., EM Reconstruction Algorithms for Emission and Transmission Tomography, J Computer Assisted Tomography 8, DD. 306, et seq. (1984).
Lange et al., Globally Convergent Algorithms for Maximum a Posteriori Transmission Tomography, IEEE Transactions on Image Processing, Vo. 4, No. 10, Oct. 1995, pp. 1430-1438.
Li et al., Tomographic Optical Breast Imaging Guided by Three-Dimensional Mammography, Applied Optics, Sep. 1, 2003, vol. 42, No. 25, pp. 5181-5190.
Li, et al., A Brick Caching Scheme for 30 Medical Imaging, Apr. 15-18, 2004, IEEE International Symposium on Biomedical Imaging: Macro to Nano 2004, vol. 1, pp. 563-566.
Maes, et al. Multimodality Image Registration by Maximization of Mutual Information, IEEE Tran. on Medical Imaging, vol. 16, No. 2, Apr. 1997. pp. 187-198.
Max, N., Optical Models for Direct Volume Rendering, IEEE Transactions on Visualization and Computer Graphics, Jun. 1995, 1(2): pp. 99-108.
McCool, M. et al., Shader Algebra, 2004, pp. 787-795.
McCool, Michael J., Smash: A Next-Generation API for Programmable Graphics Accelerators, Technical Report CS-200-14, Computer Graphics Lab Dept. of Computer Science University of Waterloo, Aug. 1, 2000.
Microsoft, Architectural Overview Direct for 3D, http://msdn.microsoft.com/library/default.asp?url=/library/en-us/dx8_c/directx_cpp/Graphics/ProgrammersGuide/GettingStarted/ Architecture, 12120/2002, 22 pages.
Mitchell, Jason L., RadeonTM 9700 Shading, SIGGRAPH 2002—State of the Art in Hardware Shading Course Notes, DD.3.1-1-3.1-39, 39 pages.
Mitschke et al., Recovering the X-ray projection geometry for three-dimensional tomographic reconstruction with additional sensors: Attached camera versus external navigation system, 2003, Medical Image Analysis, vol. 7, pp. 65-78.
Mueller, K., and R. Yagel, Rapid 3-D Cone Beam Reconstruction With The Simultaneous Algebraic Reconstruction Technique (SART) Using 2-D Texture Mapping Hardware, IEEE Transactions on Medical Imaging, Dec. 2000, 19(12): pp. 1227-1237.
Navab, N. et al., 3D Reconstruction from Projection Matrices in a C-Arm Based 3D-Angiography System, W.M. Wells e al., eds., MICCAI'98 , LNCS 1496, pp. 119-129, 1998.
Parker, S., et al., Interactive Ray Tracing for Isosurface rendering, IEEE, 1998, pp. 233-258.
PCT/US2008/084282, Preliminary and International Search Reports, dated May 11, 2011, 7 pages.
PCT/US2005/000837, Preliminary and International Search Reports, dated May 11, 2005, 7 pages.
PCT/US2008/74397, Preliminary and International Search Reports, dated Dec. 3, 2008, 7 pages.
PCT/US2008/84368, Preliminary and International Search Reports, dated Jan. 13, 2009, 7 pages.
PCT/EP2016/067886, Preliminary and International Search Reports, dated Jan. 17, 2017, 18 pages.
PCT/EP2018/075744, Preliminary and International Search Reports, dated Feb. 1, 2019, 17 pages.
PCT/US2008/84376, Preliminary and International Search Reports, dated Jan. 12, 2009, 6 pages.
JP2018-524544, Office Action, dated, Jun. 2, 2020, 4 pages (& English translation).
JP2018-524544, Office Action, dated Nov. 4, 2020, 4 pages (& English translation).
AU2016299367, Office Action, dated Nov. 12, 2020, 4 pages.
EP3329405, Office Action, dated Apr. 6, 2020, 5 pages.
JP2021-078306, Office Action, dated Apr. 18, 2022, 3 pages (& English translation).
Pfister, H., et al., The VolumePro real-time ray-casting System, Computer Graphics Proceedings of SIGGRAPH), Aug. 1999, No. 251-260.
Phong, B. T. Illumination for Computer Generated Pictures, Communications of the ACM, 18(6), Jun. 1975, pp. 311-317.
Porter, D. H. 2002. Volume Visualization of High Resolution Data using PC-Clusters. Tech. rep., University of Minnesota. Available at http://www.lcse.umn.edu/hvr/pc_vol_rend_L.pdf.
Potmesil, M. and Hoffert, E. M. 1989. The pixel machine: a parallel image computer. In Proceedings of the 16th Annual Conference on Computer Graphics and interactive Techniques SIGGRAPH '89. ACM, New York, NY, 69-78.
Purcell, T., et al., Real-time Ray Tracing on Programmable Graphics Hardware, Department of Computer Science, Stanford University, Stanford, CA, Submitted for review to SIGGRAPH 2002, 2002. http://graphics.stanford.edu/papers/rtongfx/rtongfx_submit.pdf.
Purcell, T., et al. Ray tracings on Programmable Graphics Hardware, Computer Graphics (Proceedings of SIGGRAPH), 1998, pp. 703-712.
Purcell, Timothy J., Craig Donner, Mike Cammarano , Henrik Wann Jensen , Pat Hanrahan, Photon mapping on programmable graphics hardware, Proceedings of the ACM SIGGRAPH/EUROGRAPHICS conference on Graphics hardware, Jul. 26-27, 2003, 11 pages.
Ramirez et al. (Prototypes stability analysis in the design of a binning stategy for mutual information based medical image registration, IEEE Annual Meeting of the Fuzzy Information, Jun. 27-30, 2004, vol. 2, pp. 862-866.
Rib Cage Projection, downloaded from http://www.colorado.edu/physics/2000/tomography/final_rib_cage.html on Dec. 26, 2002, 3 pages.
Roettger, Stefan, et al., Smart Hardware-Accelerated Volume Rendering, Joint EUROGRAPHICS—IEEE TCVG Symposium on Visualization, 2003, pp. 231-238, 301.
Sandborg, Michael, Computed Tomography: Physical principles and biohazards, Department of Radiation Physics, Faculty of Health Sciences, Linkoping University, Sweden, Report 81 ISSN 1102-1799, Sep. 1995 ISRN ULI-RAD-R--81--SE, 18 pages.
Sarrut et al., (Fast 30 Image Transformations for Registration Procedures, Proc. Int. Conf. on Image Analysis and Processing, Sep. 27-29, 1999, pp. 446-451.
Selldin, Hakan, Desing and Implementation of an Application Programming Interface for Volume Rendering, Linkopings Universitet.
Shekhar, R.; Zagrodsky, V., Cine MPR: interactive multiplanar reformatting of four-dimensional cardiac data using hardware-accelerated texture mapping, IEEE Transactions on Information Technolog in Biomedicine, vol. 7, No. 4, pp. 384-393, Dec. 2003.
Silver, et al., Determination and correction of the wobble of C-arm gantry, Medical Imaging 2000: Image Processing, Kenneth M. Hanson, ed., Proceedings of SPIE vol. 3970 (2000).
Stevens, Grant, et al., Alignment of a Volumetric Tomography System, Med. Phys., 28 (7), Jul. 2001.
Tao, W., Tomographic mammography using a limited number of low dose cone beam projection images, Medical Physics, AIP, Melville, NY vol. 30, pp. 365-380, Mar. 2003, ISSN: 0094-2405.
Tasdizen, T. , Ross Whitaker, Paul Burchard , Stanley Osher, Geometric surface processing via normal maps, ACM Transactions on Graphics (TOG), v.22 n.4, p. 1012-1033, Oct. 2003.
Tasdizen, T.; Whitaker, R.; Burchard, P.; Osher, S.; Geometric surface smooting via anisotropic diffusion of normals, IEEE Visualization, VIS 2002, Nov. 2002, pp. 125-132.
Technical Brief: NVIDIA nfiniteFX Engine: Programmable Pixel Shaders, NVIDIA Corporation, 5 pages.
Technical Brief: NVIDIA nfiniteFX Engine: Programmable Vertex Shaders, NVIDIA Corporation, 12 pages.
Viola, I, et al., Hardware Based Nonlinear Filtering and Segmentation Using High Level Shading Languages, Technical Report TR-186-2-03-07, May 2003, 8 pages.
Viola, P., Alignment by Maximization of Mutual Information PhD Thesis MIT (Also Referred to as—A1 Technical report No. 1548), MIT Artificial Intelligence Lab, Jun. 1, 1995, pp. 1-29.
Weiler, M, M. Kraus and T. Ertl, Hardware-Based View-Independent Cell Projection, Proceedings IEEE Symposium on Volume Visualization 2002, pp. 13-22.
Weiler, M. et al., Hardware-based ray casting for tetrahedral meshes, IEEE Visualization, VIS 2003, Oct. 24-24, 2003, pp. 333-340.
Weiler, M. et al., Hardware-Based view-Independent Cell Projection, IEEE, 2002, pp. 13-22.
Weiskopf, D., T. Schafhitzel, T. Ertl, GPU-Based Nonlinear Ray Tracing, EUROGRAPHICS, vol. 23, No. 3, Aug. 2004.
Wen, Junhai; Zigang Wang; Bin Li; Zhengrong Liang; An investigation on the property and fast implementation of a ray-driven method for inversion of the attenuated Radon transform with variable focusing fan-beam collimators, 2003 IEEE Nuclear Science Symposium Conference Record, vol. 3, Oct. 19-25, 2003, pp. 2138-2142.
Wikipedia, Anonymous, ‘Volume Rendering’ May 30, 2015, retrieved Nov. 4, 2016, https://en.wikipedia.org/w/index.php?title=Volume_rendering&oldid=664765767.
Wikipedia, Anonymous, ‘Tomographic Reconstruction’ Dec. 6, 2014, retrieved Nov. 4, 2016, https://en.wikipedia.org/w/index.php?title=Tomographic_Reconstruction&oldid=636925688.
Wu et al., Tomographic Mammography Using a Limited Number of Low-dose Conebeam Projection Images, Med. Phys., pp. 365-380 (2003).
Xu et al., Toward a Unified Framework for Rapid 30 Computed Tomography on Commodity GPUs, Oct. 19-25, 2003, IEEE Nuclear Science Symposium Conference 2003, vol. 4, pp. 2757-2759.
Xu et al., Ultra-fast 30 Filtered Backprojection on Commodity Graphics Hardware, Apr. 1-18, 2004, IEEE International symposium on Biomedical Imaging: Macro to Nano, vol. 1, pp. 571-574 and corresponding power point presentation.
Related Publications (1)
Number Date Country
20230260478 A1 Aug 2023 US
Provisional Applications (2)
Number Date Country
60989881 Nov 2007 US
60989913 Nov 2007 US
Continuations (6)
Number Date Country
Parent 16986020 Aug 2020 US
Child 18139322 US
Parent 16504146 Jul 2019 US
Child 16986020 US
Parent 16036451 Jul 2018 US
Child 16504146 US
Parent 15450888 Mar 2017 US
Child 16036451 US
Parent 14641243 Mar 2015 US
Child 15450888 US
Parent 12275834 Nov 2008 US
Child 14641243 US