Systems and methods for interleaving series of medical images

Information

  • Patent Grant
  • 10096111
  • Patent Number
    10,096,111
  • Date Filed
    Friday, June 23, 2017
    7 years ago
  • Date Issued
    Tuesday, October 9, 2018
    6 years ago
Abstract
After selecting two or more image series for comparison, images of the image series are interleaved so that they are alternatively displayed in a comparison pane on a display device. In one embodiment, after one or more image series are selected for comparison, an interleaved image series is created containing each of the images of the one or more selected image series, or, alternatively, the interleaved image series comprises links to the images arranged in the interleaved pattern. If differences exist in the images of the multiple image series, these differences may be more easily detectable as the display device cycles between the images. Comparison of images in an interleaved image series may be more advantageous if the images of each selected image series are of a common anatomical area, common image size, and the images are in the same order.
Description
BACKGROUND OF THE INVENTION
Field of the Invention

This invention relates to management and viewing of medical images and, more particularly, to systems and methods of comparing related medical images in order to detect differences in the compared images.


Description of the Related Art

Medical imaging is increasingly moving into the digital realm. This includes imaging techniques that were traditionally analog, such as mammography, x-ray imaging, angiography, endoscopy, and pathology, where information can now be acquired directly using digital sensors, or by digitizing information that was acquired in analog form. In addition, many imaging modalities are inherently digital, such as MRI, CT, nuclear medicine, and ultrasound. Increasingly these digital images are viewed, manipulated, and interpreted using computers and related computer equipment. Accordingly, there is a need for improved systems and methods of viewing and manipulating these digital images.


When comparison of related images is required, subtle differences between images may be difficult to detect. For example, if a lung radiograph from two months previous, and a current lung radiograph are to be compared in order to determine if any changes have occurred in the lungs over the previous two months, the viewer or reader typically views the two x-rays side by side. For example, the viewer or reader may have two monitors placed side by side, wherein each of the monitors displays a chest radiographic image. Alternatively, the viewer may view the two images side by side on a single monitor. However, as those of skill in the art will recognize, identifying differences in related images in this manner is often tedious and difficult. Some imaging modalities, such as CT and MRI, produce a large number of images, hundreds to even thousands of images per exam. In many cases, comparison of different series of images within the exam is required. For example, comparison of pre and post contrast images to detect areas of enhancement or comparison of PET and CT images for localization of activity is often necessary. Further, these often large exams may need to be compared to multiple prior exams to detect subtle, progressive changes over time, for example to detect a small, growing tumor. Current imaging software does not provide a satisfactory method for comparing images contained in two or more image series. Accordingly, systems and methods for comparison of images of multiple image series so that differences in the images may be more easily distinguishable are desired.


SUMMARY OF THE INVENTION

In one embodiment, a method of viewing medical images from two or more image series on a display device coupled to a computing device comprises the steps of selecting a first image series comprising two or more medical images, selecting at least one comparison image series, each of the comparison image series comprising two or more medical images, interleaving images of the first image series and the comparison image series in order to form an interleaved image series, and sequentially displaying the images of the interleaved image series at a single location on the display device.


In another embodiment, a method of viewing a series of medical images on a display device coupled to a computing device comprises the steps of (a) selecting a first image series for viewing, the first image series comprising a plurality X of medical images, (b) selecting a second image series for viewing, the second image series comprising a plurality Y of medical images, (c) displaying at a predetermined location on the display device a Nth image of the first image series, (d) replacing the Nth image of the first image series with a Mth image of the second image series at the predetermined location, (e) incrementing N and M, and (f) repeating steps (c) to (f).


In another embodiment, a system for enhancing a viewer's ability to detect differences between medical images in two or more sets of medial images comprises a display device, a graphical user interface displayed on the display device and comprising an image pane configured to display a single medical image at a time, an image selection module to select two or more sets of medical images, each of the sets of medical images comprising two or more medical images, and a user interface to receive commands from a user, wherein in response to receiving a first command from the user, the image pane sequentially displays a first medical image from each of the image sets and, after displaying the first medical image from each image set, the image pane sequentially displays a second medical image from each image set. This process of displaying images from images series alternatively continues through subsequent images in the image series.


In another embodiment, a system of viewing medical images from two or more image series on a display device coupled to a computing device comprises means for selecting a first image series comprising two or more medical images, means for selecting at least one comparison image series, each of the comparison image series comprising two or more medical images, means for interleaving images of the first image series and the comparison image series in order to form an interleaved image series, and means for sequentially displaying the images of the interleaved image series at a single location on the display device.


In another embodiment, a method of forming an interleaved image series comprises selecting N groups of images, each of the groups of images comprising two or more images, determining a starting image of each of the groups of images, creating an interleaved image series comprising images from each of the selected N groups of images, wherein the images of the interleaved image series are ordered so that an image from each of the N groups of images is included in each sequential Nth group of images, and providing the interleaved image series to a user interface for sequential display in a predetermined location of a display device.





BRIEF DESCRIPTION OF THE DRAWINGS


FIG. 1 is a block diagram of an exemplary computing system in communication with a network and various networked devices.



FIG. 2 is a diagram illustrating a display device having images from two image series concurrently displayed in image panes displayed on a display device.



FIG. 3 is a diagram illustrating a display device having images from four image series concurrently displayed in image panes.



FIG. 4 is a diagram illustrating sequential changes to a comparison pane displayed on the display device as images from two image series are compared.



FIG. 5 is a diagram illustrating an exemplary interleaving of four image series.



FIG. 6 is a flowchart illustrating an exemplary method of viewing images from multiple image series.



FIGS. 7A and 7B are images of a first image series.



FIGS. 8A and 8B are images of a second image series.



FIGS. 9A-9D are the images of the first and second image series interleaved for sequential viewing in a predetermined portion of a display.



FIG. 10 is a diagram illustrating a display device displaying two comparison panes.



FIG. 11 illustrates the first images of the first and second image series illustrated in FIGS. 7A and 8A interleaved for alternative viewing in a comparison pane.



FIG. 12 illustrates the second images of the first and second image series illustrated in FIGS. 7B and 8B interleaved for alternative viewing in a comparison pane.





DETAILED DESCRIPTION OF EMBODIMENTS OF THE INVENTION

Embodiments of the invention will now be described with reference to the accompanying figures, wherein like numerals refer to like elements throughout. The terminology used in the description presented herein is not intended to be interpreted in any limited or restrictive manner, simply because it is being utilized in conjunction with a detailed description of certain specific embodiments of the invention. Furthermore, embodiments of the invention may include several novel features, no single one of which is solely responsible for its desirable attributes or which is essential to practicing the inventions herein described.



FIG. 1 is a block diagram of an exemplary computing system 100 in communication with a network 160 and various network devices. The computing system 100 may be used to implement certain systems and methods described herein. The functionality provided for in the components and modules of computing system 100 may be combined into fewer components and modules or further separated into additional components and modules.


The computing system 100 includes, for example, a personal computer that is IBM, Macintosh, or Linux/Unix compatible. In one embodiment, the exemplary computing system 100 includes a central processing unit (“CPU”) 105, which may include a conventional microprocessor, an application module 145 that comprises one or more various applications that may be executed by the CPU 105. The application module 145 may include, by way of example, components, such as software components, object-oriented software components, class components and task components, processes, functions, attributes, procedures, subroutines, segments of program code, drivers, firmware, microcode, circuitry, data, databases, data structures, tables, arrays, and variables.


The computing system 100 further includes a memory 130, such as random access memory (“RAM”) for temporary storage of information and a read only memory (“ROM”) for permanent storage of information, and a mass storage device 120, such as a hard drive, diskette, or optical media storage device. Typically, the modules of the computing system 100 are connected to the computer using a standards-based bus system. In different embodiments of the present invention, the standards based bus system could be Peripheral Component Interconnect (PCI), Microchannel, SCSI, Industrial Standard Architecture (ISA) and Extended ISA (EISA) architectures, for example.


The computing system 100 is generally controlled and coordinated by operating system software, such as the Windows 95, 98, NT, 2000, XP or other compatible operating systems. In Macintosh systems, the operating system may be any available operating system, such as MAC OS X. In other embodiments, the computing system 100 may be controlled by a proprietary operating system. Conventional operating systems control and schedule computer processes for execution, perform memory management, provide file system, networking, and I/O services, and provide a user interface, such as a graphical user interface (“GUI”), among other things.


The exemplary computing system 100 includes one or more of commonly available input/output (I/O) devices and interfaces 110, such as a keyboard, mouse, touchpad, and printer. In one embodiment, the I/O devices and interfaces 110 include one or more display devices, such as a monitor, that allows the visual presentation of data to a user. More particularly, display devices provide for the presentation of GUIs, application software data, and multimedia presentations, for example. In one embodiment, a GUI includes one or more image panes in which medical images may be displayed. As described in further detail below, a GUI may provide a comparison pane on a display device in which images from multiple image series are sequentially displayed. According to the systems and methods described below, medical images may be stored on the computing system 100 or another device that is local or remote, displayed on a display device, and manipulated by the application module 145. The computing system 100 may also include one or more multimedia devices 140, such as speakers, video cards, graphics accelerators, and microphones, for example.


In the embodiment of FIG. 1, the I/O devices and interfaces 110 provide a communication interface to various external devices. In the embodiment of FIG. 1, the computing system 100 is coupled to a network 160, such as a LAN, WAN, or the Internet, for example, via a communication link 115. The network 160 may be coupled to various computing devices and/or other electronic devices. In the exemplary embodiment of FIG. 1, the network 160 is coupled to imaging devices 170, an image server 180, and a medical facility 190. In addition to the devices that are illustrated in FIG. 1, the network 160 may communicate with other computing, imaging, and storage devices.


The imaging devices 170 may be any type of device that is capable of acquiring medical images, such as a MRI, x-ray, mammography, or CT scan systems. The image server 180 includes a data store 182 that is configured to store images and data associated with images. In one embodiment, the imaging devices 170 communicate with the image server via the network 160 and image information is transmitted to the image server 160 and stored in the data store 182. In one embodiment, the image data is stored in Digital Imaging and Communications in Medicine (“DICOM”) format. The complete DICOM specifications may be found on the National Electrical Manufactures Association Website at <medical.nema.org>. Also, NEMA PS 3—Digital Imaging and Communications in Medicine, 2004 ed., Global Engineering Documents, Englewood Colo., 2004, provides an overview of the DICOM standard. Each of the above-cited references is hereby incorporated by reference in their entireties. In one embodiment, the data store 182 also stores the user-defined display parameters associated with one or more of the images stored on the data store 182. As discussed in further detail below, the user-defined display parameters may vary depending of the type of image, area imaged, clinical indication, source of image, display device, user, or other factors. Accordingly, any type of user-defined display parameter is expressly contemplated for use in conjunction with the systems and methods described herein.


The exemplary image server 180 is configured to store images from multiple sources and in multiple formats. For example, the image server 180 may be configured to receive medical images in the DICOM format from multiple sources, store these images in the data store 182, and selectively transmit medical images to requesting computing devices.


The medical facility 190 may be a hospital, clinic, doctor's office, or any other medical facility. The medical facility 190 may include one or more imaging devices and may share medical images with the image server 180 or other authorized computing devices. In one embodiment, multiple computing systems, such as the computing system 100 may be housed at a medical facility, such as medical facility 190.


DEFINITION OF TERMS

“Medical image” is defined to include an image of an organism. It may include but is not limited to a radiograph, computed tomography (CT), magnetic resonance imaging (MRI), Ultrasound (US), mammogram, positron emission tomography scan (PET), nuclear scan (NM), pathology, endoscopy, ophthalmology, or many other types of medical images. While this description is directed to viewing and tracking of medical images, the methods and systems described herein may also be used in conjunction with non-medical images, such as, images of circuit boards, airplane wings, and satellite images, for example.


“Modality” is defined as a medical imaging device (a patient who undergoes an MRI is said to have been scanned with the MRI modality).


“Image series” refers to two or more images that are related. For example, an image series may comprise two or more images of a particular patient that are acquired on a particular date, e.g., different x-ray projections of the chest. A series of contiguous 3 mm axial CT scans of the chest would be another example of an image series. A brain MRI scan might include the following series: sagittal T1 weighted images, axial T1 weighted images, axial FLAIR images, axial T2 weighted images, as well as post contrast axial, sagittal and coronal T1 weighted series. An image series may be limited to images of a certain modality or may comprise images of multiple modalities.


“Patient” refers to an individual who undergoes a medical imaging examination.


“Display parameters” are defined to include methods of display of an image or exam. Display parameters may include, for example, a pixel window level and width (similar to brightness and contrast), a certain color map that renders different pixel intensities as different colors, or opacity map.


“Interleaving,” is defined to include the process of arranging images from multiple image series by regularly alternating between images of the multiple image series in order to create a resultant “interleaved” image series. In one embodiment, an interleaved image series comprises images from multiple image series ordered so that the interleaved image series alternates between the images of the original series. For example, when image series A comprising images A1, A2, . . . An, image series B comprising images B1, B2, . . . Bn, and image series C comprising images C1, C2, . . . Cn are interleaved, the resultant interleaved image series is ordered: A1, B1, C1, A2, B2, C2, . . . An, Bn, Cn. Images from multiple image series may be interleaved in various patterns and multiple interleaved image series may be generated from two or more image series.



FIG. 2 is a diagram illustrating a display device having images from two image series concurrently displayed in image panes 212, 214 displayed on a display device 210. In the discussion that follows, the display device 210 is coupled to a computing device, such as computing device 100, and receives display information from the computing device 100. While the systems and methods described below for interleaving and viewing images of multiple image series may be controlled by any suitable computing device, for ease of explanation herein, reference will be made to a display device coupled to computing device 100.


In the embodiment of FIG. 2, an image of a series A image series is displayed in the image pane 212, while an image of the series B image series is displayed in the image pane 214. As noted above, each image series comprises a group of images that are related in some way, such as having been acquired from a patient on a particular day. Although only a single image of each of the image series is simultaneously displayed on the display device 210, the series A and series B image series each comprise multiple images.


In certain embodiments, such as where the series A and B images are related, e.g., series A comprises mammogram images of a patient taken on a first date and series B comprises the mammogram images of the same patient taking on a later date, it may be advantageous to identify differences between the images of Series A and Series B. However, as described above, it is difficult to distinguish minor or small differences between images using currently available image comparison techniques. Using current image comparison systems, if a lung radiograph from two months previous, and a current lung radiograph are to be compared in order to determine if any changes have occurred in the lungs over the previous two months, the viewer or reader typically views the two x-rays side by side, such as in image panes 212, 214 illustrated in FIG. 2. As those of skill in the art will recognize, identifying differences in related images in this manner is often tedious and difficult. Accordingly, described hereinafter are exemplary systems and methods for comparison of images of multiple image series so that differences between (in contrast to within) the images may be more easily distinguishable. In certain embodiments, related images are displayed sequentially in a single viewing pane on a display device so that difference between the images may be more easily detected. The systems and methods described herein are applicable to any two or more images, including multiple images of multiple image series, for example.



FIG. 3 is a diagram illustrating a display device having images from four image series concurrently displayed in image panes 312, 314, 316, in 318. In the embodiment of FIG. 3, the image pane 312 displays images from series A, the image pane 314 displays images from series B, the image pane 316 displays images from series C, and the image pane 318 displays images from series D. Thus, a single image from each of the four image series A-D is concurrently displayed on the display device 210. In certain prior art systems, comparison of images of multiple series was performed using a graphical user interface such as displayed in FIG. 3, wherein the user distinguishes differences between images that are displayed side-by-side on a display device.



FIG. 4 is a diagram illustrating sequential changes to a comparison pane 410 displayed on the display device 210 as images from two image series are compared. The comparison pane 410 is configured to display a single image. In one embodiment, one of the display panes, e.g. display pane 312 on display device 210, serve as the comparison pane 410. In the embodiment of FIG. 4, images from two image series, images series A and B, have been selected for comparison. After being selected for comparison using any suitable selection method, images from series A and B are interleaved so that they are alternatively displayed in the comparison pane 410.


In embodiments with more than two series, the interleaved images may be ordered according to various schemes. For example images from four image series may be ordered as follows: first image of first image series, first image of second image series, first image of third image series, first image of fourth image series, second image of first image series, and so forth. In other embodiments, however, the interleaved images may be ordered differently. For example, images from four image series may also be ordered as follows: first image of first image series; first image of second image series; first image of first image series; first image of third image series; first image of first image series; and first image of fourth image series. Any other ordering of images from multiple image series falls within the scope of “interleaving” as used herein.



FIG. 4 shows the display device 210 at four steps 402, 404, 406, 408 of the comparison process, where the comparison process describes the process of displaying images in an interleaved image series. More particularly, in step 402, a first image of image series A is displayed in the comparison pane 410. Moving to step 404, the first image of image series A is replaced by a first image of image series B in the comparison pane 410. Assuming the images of series A and B are of the same subject, the image displayed in steps 402 and 404 may be very similar. Accordingly, if differences exist in the first images of series A and B, these differences may be more easily detectable as the display device cycles between the two images. Comparison of images in an interleaved image series may be more advantageous if the images of each selected image series are of a common anatomical area, common image size, common image orientation, and the images are in the same order.


In one embodiment, the computing system 100 that is coupled to the display 210 may store settings for displaying images of particular image series, such as, for example, time for displaying each image, resolution of each image, cropping to be applied to each image, and any other setting that maybe appropriate. In one embodiment, the time for displaying an image may be determined real time by the user. For example, the user may press a designated key on a keyboard or mouse in order to indicate that the current image should be replaced with an adjacent image in the interleaved image series. In another embodiment, the user selects settings for display of the images. For example, the user may select an appropriate zoom level of an image series that should be applied to each image in the image series.


Thus, the images of series A may be magnified more or less than the images of series B. In addition, the user may adjust any other visualization settings for individual images, an entire image series, or two or more image series.


With the first image of series B displayed in the comparison pane 410 (step 404), the user may initiate viewing of an adjacent image in the interleaved image series by pressing a certain key on a keyboard or mouse, for example. In one embodiment, a first input from a mouse indicates that a next image, e.g. image 2 of series A (step 406) should be displayed in the comparison pane 410 and a second input from the mouse indicates that a previous image, e.g. image 1 of series A (step 402) should again be displayed in the comparison pane 410. In one embodiment, the first input is entered by the user moving a scroll button on the mouse in a first direction and the second input is entered by the user moving the scroll button on the mouse in an opposite direction. Thus, the user may change the content of the comparison pane 410 to either a next or a previous image in the interleaved image series. For example, at step 404, if the user wishes to again view the first image of series A, e.g., in order to locate difference in the first images of series A and B, the user may provide an input to the computing device 100 indicating movement to a previous image. Alternatively, at step 404, if the user wishes to view a next image in the interleaved image series, the user may provide an input to the computing device 100 indicating movement to a next image.


At step 406, the second image of series A is displayed in the comparison pane 410, replacing the first image of series B (step 404). At step 406, the user may provide inputs to the computing device 100 indicating that the comparison pane 410 should be updated with a previous image, e.g. step 404, or a subsequent image, e.g., step 408.


At step 408, the second image of series B is displayed in the comparison pane 410, replacing the second image of series B (step 406). At step 406, the user may provide inputs to the computing device 100 indicating that the comparison pane 410 should be updated with a previous image, e.g. step 404, or a subsequent image. In one embodiment, each of the image series A and B include more than two images, such as 3 or more images, and the images of series A and B are displayed in the manner described above with respect to FIG. 4. In one embodiment, more than two images series may be interleaved for display in the comparison pane. For example, if three images series, e.g., series A, B, and C, are selected for comparison, a first image of each of the series will be sequentially displayed in the comparison pane, followed by a second image of each of the series, and so on. As noted above, the user may control the timing of transition between display of images in the interleaved image series and may even control the direction of movement in the interleaved series. Additionally, the user may control alignment and/or positioning of the images of each images series in order to precisely align interleaved images from multiple series.


In one embodiment, the images of each of the image series are automatically modified so that characteristics of the images are similar. For example, images may be adjusted by changing their size, rotation, and location. If the images are of substantially the same anatomical structure, when the images of the interleaved image series are displayed in the comparison pane, differences between adjacent images may be more easily detected. In one embodiment, selected images are morphed in order to achieve a common size of the anatomical structure of interest in each of the images. In one embodiment, photographic filters may be applied to all images of one or more image series, or to selected images of one or more image series, to further enhance the viewer's ability to distinguish differences in the images.


In one embodiment, information regarding the image currently displayed in the comparison pane 410 is displayed on the display device 210 and updated as the images in the comparison pane 410 are changed. For example, information regarding the images series and image number within the series may be displayed for each image. In addition, the exam date and time may also be displayed and updated as the images of the interleaved image series are displayed in the comparison pane 410. In one embodiment, an indicator of whether the current display is of an interleaved image series or a single image series is displayed on the display device. For example, “interleaved” may be displayed at the top of the display device when an interleaved image series is displayed in a comparison pane. In some embodiment, the user chooses what information related to the images of the interleaved image series should be displayed. The user may also be provided the ability to turn the display of information on and off, such as by pressing a particular key or key combination on the keyboard.



FIG. 5 is a diagram illustrating an exemplary interleaving of four image series in creating an interleaved image series. As noted above, the image series that are selected for comparison, and thus, are selected for interleaving, may be selected by a user in one of several manners or may be automatically selected by the computing device based on properties of the image series. In one embodiment, the multiple image series are interleaved so that a first image of each series is displayed in the comparison pane prior to display of a second image of any of the other selected image series. FIG. 5 illustrates an exemplary interleaving of four image series, series A, B, C, and D, each comprising four images. As those of skill in the art will recognize, more or less image series, each comprising more or less images, may be interleaved in a manner similar to that illustrated in FIG. 5.


In FIG. 5, the images of series A are represented by diamonds in a column 510, where a number in each of the diamonds represents a specific image within image series A. Similarly, images of series B are represented by squares in column 520, series C are represented by ovals in a column 530, and series D are represented by parallelograms in a column 540. Each of the images in images series B, C, and D are likewise numbered 1-4, indicating a particular image in each image series. As noted above, the first image selected for comparison in a particular image series may not be the first image in the image series, e.g., the first image of an exam. Thus, although each image series A-D begins with a first image labeled image “1”, this first image may not be the first image in the image series, but may be a user selected, or automatically selected, start image. In addition, each of the images may have user-defined display parameters that are different than other images of the same series and/or other image series. In some embodiments, display parameters, such as zoom level, cropping, and color characteristics, may be simultaneously changed for each image in an image series, such as series A, B, C, D, or an interleaved image series, such as interleaved image series 540.


As illustrated in FIG. 5, an interleaved image series 550 each of the images 1-4 in each of the image series A-D. More particularly, the interleaved image series 550 comprises a first image from each of series A-D, followed by a second image from each of the series A-D, followed by a third image from each of the series A-D, followed by a fourth image from each of the series A-D. Thus, when the interleaved image series 550 is displayed in the comparison pane, a first image of the image series A is displayed, followed by a first image of image series B, a first image of image series C, and a first image of image series D. While the order of the interleaved image series 550 is maintained during viewing of the interleaved images, the direction of movement between adjacent images may be selected by the user or automatically by the computing device 100.



FIG. 6 is a flowchart illustrating an exemplary method of viewing images from multiple image series. Using the method of FIG. 6, multiple image series may be easily compared and differences between images of the multiple image series may be distinguished.


In a block 610, a first image series is selected. As noted above, an image series is a group of two or more images that are in some way related. For example, a first image series may comprise multiple chest x-rays of a patient that are taken on a given date.


In a block 620, one or more comparison image series are selected. These image series also each comprise two or more images that are in some way related, such as having been taken at a common exam. The comparison image series should be related to the first image series so that when the first image series and the comparison image series are compared, meaningful distinctions between the image series may be detected.


In one embodiment, the first image series and the comparison image series are selected by the user clicking on a button indicating that image interleaving is desired. In one embodiment, a user right-clicks with a mouse on an image of a first image series in order to initiate display of an “interleave menu” listing options for selecting image series for interleaving and viewing the interleaved image series. In one embodiment, the interleave menu includes an option, such as “interleave adjacent,” indicating that the user may select one of the other image series displayed on the display for interleaving. In certain embodiments, any number of image panes may be simultaneously displayed on the display device 210. For example, in FIG. 2, two image panes 212,214 are display and in FIG. 3, four image panes 312, 314, 316, 318 are displayed. In other embodiment, six, eight, ten, twelve, or more image panes may be concurrently displayed on the display device.


When the interleave adjacent option is selected, the user may select one or more comparison series by moving the pointer to a border between the adjacent series and clicking the mouse button. In one embodiment, the cursor icon changes when it is positioned in a border indicating that the adjacent image series may be selected for comparison by clicking the mouse button. With reference to FIG. 3, for example, the user may right click on the image pane 314 in order to select series B as the first image series and to initiate display of the interleave menu. From the interleave menu, if the user selects interleave adjacent, the user may then move the pointer to the border between the image panes 312 and 314 and click the mouse button in order to select series A as a comparison image series. In one embodiment, selecting a comparison image series initiates creation of an interleaved image series and displays the first image of the interleaved image series in the comparison pane. In an embodiment when only two image series are represented on the display device, such as FIG. 2, selection of interleave adjacent from the interleave menu may automatically select the two displayed image series for interleaving and initiate creation and viewing of a interleaved image series.


In one embodiment, the interleave menu also includes an option that allows the user to select an adjacent image series for interleaving and, after selection of the first and comparison image series, displays images of the interleaved images series in a comparison pane that covers the entire display area, or substantially all of the display area, of the display device. In this way, the images of the selected image series may be viewed at a higher magnification level and, accordingly, differences in the images may be more easily detectable. Thus, in an embodiment that displays four image panes on the display device (e.g., FIG. 3), after selection of this option from the interleave menu, a single click on a comparison image series may cause the computing device to generate an interleaved image series and display a first image of the interleaved image series in a comparison pane that covers substantially all of the display area of the display device, e.g., the area previously covered by the four image panes 312, 314, 316, 318 or FIG. 3. Advantageously, this “interleave and jump to full screen display” option on the interleave menu provides an efficient transition from display of many image series to the display of a single interleaved series in a comparison pane that covers all, or substantially all, of the display area of a display device.


In one embodiment, the interleave menu includes an option that initiates automatic selection of one or more comparison image series based upon characteristics of the selected first image series. For example, image series with the same or similar names may be selected as comparison image series. In addition, image series may be selected automatically based upon any other criteria, such as one or more information items contained in the DICOM headers of images. In one embodiment, when this option is chosen from the interleave menu, a list of image series that have the same series name, or other criteria that may be user defined, may be displayed. The user may then select one or more of the displayed series as comparison image series.


The interleave menu advantageously allows the user to select image series for interleaving and automatically display the generated interleaved image series with minimal input from the user. For example, after selecting “interleave adjacent” on the interleave menu, a single click of a mouse, for example, on a border between the images to be interleaved causes the computing system 100 to generate an interleaved image series and display a first image of the interleaved image series in a comparison pane on the display device.


Returning to the diagram of FIG. 6, in a block 630, a starting image of each of the image series is selected. In one embodiment, a first image of the first image series and each of the comparison image series are displayed on a display device. A user, using an input device, such as a mouse or keyboard, may cycle through the images in each of the image series in order to determine a first image for comparison. For example, images of certain modalities, such as CT and MRI images, may not have starting images that are each taken at similar physical locations within the patient. Thus, in these embodiments the user may select a different starting image in each of the image series so that adjacent images in the interleaved image series are more closely related. For example, if the images series to be interleaved are series A comprising images A1, A2, . . . An, series B comprising images B1, B2, . . . Bn, and series C comprising images C1, C2, . . . Cn, the user may select images A1, B3, and C5 as the starting images of the respective image series so that the resultant interleaved image series is ordered A1, B3, C5, A2, B4, C6, A3, B5, C7, . . . Ax, Bx+2, Cx+4.


In an advantageous embodiment, the starting image in each of the series should be related so that meaningful differences between the images of the image series are detectable. In one embodiment, the user may adjust display characteristics of each image in an image series by adjusting the display characteristics of the image currently displayed from the desired image series. For example, if the first image series is at a higher zoom level than the comparison image series, the zoom level of each image in the first image series may be adjusted by adjusting the zoom level of the currently display image of the first image series.


After selecting the first image series and the comparison image series, and selecting the display characteristics of one or more of the images which are then applied to the other images in the series to which it belongs, the image series are interleaved so that an interleaved image series is created, as illustrated in FIG. 5, for example.


In a block 640, images from the interleaved image series are displayed in a comparison pane displayed on the display device. In one embodiment, the comparison pane fills substantially the entire display area of a display device. In another embodiment, the comparison pane is smaller than a total display area of the display device. In one embodiment, a user input determines when a current image displayed in the comparison pane is updated with an adjacent image, by moving a scroll wheel on a mouse while pressing a mouse button, for example. In one embodiment, when the user has completed viewing the interleaved image series, the previous layout on the display device may be restored by the user performing a specific action, such as releasing the mouse button that is depressed while viewing interleaved images.



FIGS. 7A and 7B are images of a first image series, FIGS. 8A and 8B are images of a second image series, and FIGS. 9A-9D are the images of an interleaved image series comprising images of the first and second image series. More particularly, FIGS. 7A and 7B are two mammographic images 710, 720 in a first image series. In one embodiment, the images 710, 720 are of a patient on a first date. FIGS. 8A and 8B are two mammographic images 810, 820 in a second image series. In one embodiment, the images 710, 720, 810, 820 are of the same patient, but the images 810, 820 were taken at a later date than the images 710, 720. Thus, differences between the earlier mammographic images 710, 720 and the later mammographic images 810, 820 may be useful in detection and diagnosis of the patient. Accordingly, comparison of the images of the images series illustrated in FIGS. 7 and 8 may allow detection of differences between the earlier and later images.



FIGS. 9A-9D are the images of FIGS. 7 and 8 combined in an interleaved image series for viewing in a comparison pane. For example, image 710 may first be displayed in a comparison pane. When the user indicates that a next image should be displayed, image 810 may replace image 710 in the comparison pane. With image 810 displayed in the comparison pane, when the user indicates that a next image should be displayed, image 720 is replaced with image 810 in the comparison pane. With image 720 displayed in the comparison pane, when the user indicates that a next image should be displayed, image 720 is replaced with image 820 in the comparison pane. As noted above, however, movement between images in an interleaved image series, such as that of FIG. 9, may be bidirectional so that the user may move back and forth between images as desired.


In another embodiment, interleaving of image series produces two or more interleaved image series. In one embodiment, the first image of each image series may be interleaved and alternatively displayed in a first comparison pane of the display device, while the second image of each image series may be interleaved and alternatively displayed in a second comparison pane of the display device, and so on. For example, if the image series to be interleaved are series A comprising images A1, A2, and series B comprising images B1, B2, the system may generate a first interleaved series ordered A1, B1, and a second interleaved image series ordered A2, B2. In one embodiment, images from each from each of the first and second interleaved image series are concurrently displayed on a display device in separate comparison panes. In one embodiment, an option on the interleave menu, discussed above, may be selected in order to initiate generation of multiple interleaved image series and concurrent display of multiple comparison panes on the display device. In one embodiment, more than two images of image series may be interleaved in the above-described manner and more than two comparison panes may be displayed concurrently on a display device for viewing the generated interleaved image series.



FIG. 10 is a diagram illustrating the display device 210 having two comparison panes 1010A, 1010B. As noted above, multiple interleaved image series may be concurrently displayed on the display device 210 in separate comparison panes. In another embodiment, the display device 210 displays three or more comparison panes for concurrently displaying three or more interleaved image series.



FIG. 11 illustrates the first images 710, 810 of the first and second image series illustrated in FIGS. 7A and 8A and FIG. 12 illustrates the second images 720, 820 of the first and second image series illustrated in FIGS. 7B and 8B. The images 710, 810 (FIG. 11) comprise a first interleaved image series, while the images 720, 820 (FIG. 12) comprise a second interleaved image series. In the embodiment illustrated in FIGS. 11 and 12, the images 710, 720 are from a first exam and the images 810, 820 are from a second exam. However, the first selected images from each exam, e.g., images 710, 810, are of a first projection, while the second selected images from each exam, e.g., 720, 820, are of a second projection. Thus, it may be advantageous for an interpreter of the images to view the images of the same projection, from different exams, in separate interleaved image series. Accordingly, in the embodiment of FIGS. 10, 11, and 12, a first interleaved image series comprising images 710, 810 are viewed in a first comparison pane 1010A while a second interleaved image series comprising images 720, 820 are viewed in a second comparison pane 1010B. In this embodiment, the viewer may advantageously move between images of the same projection in a comparison pane in order to identify differences in the images, while also viewing images of one or more additional projections in additional comparison panes. In one embodiment, any number of images may be included in each of the interleaved images series displayed in comparison panes 101A, 1010B, and additional comparison panes may be concurrently displayed on the display device 210.


The foregoing description details certain embodiments of the invention. It will be appreciated, however, that no matter how detailed the foregoing appears in text, the invention can be practiced in many ways. As is also stated above, it should be noted that the use of particular terminology when describing certain features or aspects of the invention should not be taken to imply that the terminology is being re-defined herein to be restricted to including any specific characteristics of the features or aspects of the invention with which that terminology is associated. The scope of the invention should therefore be construed in accordance with the appended claims and any equivalents thereof.

Claims
  • 1. A computer-implemented method comprising: by one or more processors executing program instructions: accessing a set of images comprising at least a first plurality of images of a first image series and a second plurality of images of a second image series;determining values of a first characteristic associated with images of the set of images;ordering the images of the set of images based at least in part on the values of the first characteristic associated with the images of the set of images to generate an ordered set of images;determining a first value of a second characteristic associated with a first image of the set of images;determining a second value of the second characteristic associated with a second image of the set of images, wherein the first image and the second image are adjacent to each other in the ordered set of images;determining a difference between the first value and the second value of the second characteristic;adjusting one or more characteristics of the first image or the second image based on the difference between the first value and the second value of the second characteristic;in response to a first user input, causing display of the first image at a location of an electronic display; andin response to a second user input, causing display of the second image at the location of the electronic display to replace the first image.
  • 2. The computer-implemented method of claim 1, wherein the first characteristic comprises at least one of a time, a modality, an application of contrast agent, an anatomical position associated with the images of the set of images.
  • 3. The computer-implemented method of claim 1, wherein the second characteristic is a display characteristic and adjusting the one or more characteristics of the first image or the second image comprises adjusting the display characteristic of the first image or the second image.
  • 4. The computer-implemented method of claim 3, wherein the display characteristic comprises at least one of brightness, contrast, size, opacity map, rotation, location, zoom level, cropping, morphing, or color.
  • 5. The computer-implemented method of claim 1, wherein ordering the images of the set of images is further based on at least one of a type of image, an area imaged, a clinical indication, a source of image, a display device, or a user.
  • 6. The computer-implemented method of claim 1, wherein the first image series and the second image series are selected based on characteristics of the first or the second image series comprises at least one of: a header information item, a filename, a modality, a name, an anatomical area, an image size, or an image orientation of the first or the second image series.
  • 7. A computing system comprising: an electronic display;an input device;a non-transitory computer-readable storage medium configured to store software instructions; andone or more computer processors in communication with the electronic display, the input device, and the non-transitory computer-readable medium, the one or more computer processors configured to execute the software instructions in order to cause the computing system to: access a set of images comprising at least a first plurality of images of a first image series and a second plurality of images of a second image series;determine values of a first characteristic associated with images of the set of images;order the images of the set of images based at least in part on the values of the first characteristic associated with the images of the set of images to generate an ordered set of images;determine a first value of a second characteristic associated with a first image of the set of images;determine a second value of the second characteristic associated with a second image of the set of images, wherein the first image and the second image are adjacent to each other in the ordered set of images;determine a difference between the first value and the second value of the second characteristic;adjust one or more characteristics of the first image or the second image based on the difference between the first value and the second value of the second characteristic;in response to a first user input, causing display of the first image at a location of the electronic display; andin response to a second user input, causing display of the second image at the location of the electronic display to replace the first image.
  • 8. The computer system of claim 7, wherein the first characteristic comprises at least one of a time, a modality, an application of contrast agent, an anatomical position associated with the images of the set of images.
  • 9. The computer system of claim 7, wherein the second characteristic is a display characteristic and adjusting the one or more characteristics of the first image or the second image comprises adjusting the display characteristic of the first image or the second image.
  • 10. The computer system of claim 9, wherein the display characteristic comprises at least one of brightness, contrast, size, opacity map, rotation, location, zoom level, cropping, morphing, or color.
  • 11. The computer system of claim 7, wherein ordering the images of the set of images is further based on at least one of a type of image, an area imaged, a clinical indication, a source of image, a display device, or a user.
  • 12. The computer system of claim 7, wherein the first image series and the second image series are selected based on characteristics of the first or the second image series comprises at least one of: a header information item, a filename, a modality, a name, an anatomical area, an image size, or an image orientation of the first or the second image series.
  • 13. A non-transitory computer readable storage medium having program instructions embodied therewith, the program instructions executable by one or more processors to cause the one or more processors to: access a set of images comprising at least a first plurality of images of a first image series and a second plurality of images of a second image series;determine values of a first characteristic associated with images of the set of images;order the images of the set of images based at least in part on the values of the first characteristic associated with the images of the set of images to generate an ordered set of images;determine a first value of a second characteristic associated with a first image of the set of images;determine a second value of the second characteristic associated with a second image of the set of images, wherein the first image and the second image are adjacent to each other in the ordered set of images;determine a difference between the first value and the second value of the second characteristic;adjust one or more characteristics of the first image or the second image based on the difference between the first value and the second value of the second characteristic;in response to a first user input, causing display of the first image at a location of an electronic display; andin response to a second user input, causing display of the second image at the location of the electronic display to replace the first image.
  • 14. The non-transitory computer readable medium of claim 13, wherein the first characteristic comprises at least one of a time, a modality, an application of contrast agent, an anatomical position associated with the images of the set of images.
  • 15. The non-transitory computer readable medium of claim 13, wherein the second characteristic is a display characteristic and adjusting the one or more characteristics of the first image or the second image comprises adjusting the display characteristic of the first image or the second image.
  • 16. The non-transitory computer readable medium of claim 15, wherein the display characteristic comprises at least one of brightness, contrast, size, opacity map, rotation, location, zoom level, cropping, morphing, or color.
  • 17. The non-transitory computer readable medium of claim 13, wherein ordering the images of the set of images is further based on at least one of a type of image, an area imaged, a clinical indication, a source of image, a display device, or a user.
  • 18. The non-transitory computer readable medium of claim 13, wherein the first image series and the second image series are selected based on characteristics of the first or the second image series comprises at least one of: a header information item, a filename, a modality, a name, an anatomical area, an image size, or an image orientation of the first or the second image series.
CROSS-REFERENCE TO RELATED APPLICATIONS

This application is a continuation of U.S. patent application Ser. No. 15/254,627, filed on Sep. 1, 2016 and titled “SYSTEMS AND METHODS FOR INTERLEAVING SERIES OF MEDICAL IMAGES,” which is a continuation of U.S. patent application Ser. No. 14/502,055, now U.S. Pat. No. 9,471,210, filed on Sep. 30, 2014 and titled “SYSTEMS AND METHODS FOR INTERLEAVING SERIES OF MEDICAL IMAGES,” which is a continuation of U.S. patent application Ser. No. 12/857,915, now U.S. Pat. No. 8,879,807, filed on Aug. 17, 2010 and titled “SYSTEMS AND METHODS FOR INTERLEAVING SERIES OF MEDICAL IMAGES,” which is a continuation of U.S. patent application Ser. No. 11/268,261, now U.S. Pat. No. 7,885,440, filed on Nov. 3, 2005 and titled “SYSTEMS AND METHODS FOR INTERLEAVING SERIES OF MEDICAL IMAGES,” which claims priority under 35 U.S.C. § 119(e) to U.S. Provisional Application Ser. No. 60/625,690, filed on Nov. 4, 2004, each of which is hereby expressly incorporated by reference in its entirety.

US Referenced Citations (427)
Number Name Date Kind
4672683 Matsueda Jun 1987 A
5123056 Wilson Jun 1992 A
5172419 Manian Dec 1992 A
5179651 Taaffe et al. Jan 1993 A
5431161 Ryals et al. Jul 1995 A
5452416 Hilton et al. Sep 1995 A
5515375 DeClerck May 1996 A
5542003 Wofford Jul 1996 A
5734915 Roewer Mar 1998 A
5740267 Echerer et al. Apr 1998 A
5779634 Ema et al. Jul 1998 A
5807256 Taguchi Sep 1998 A
5835030 Tsutsui et al. Nov 1998 A
5852646 Klotz et al. Dec 1998 A
5857030 Gaborski Jan 1999 A
5926568 Chaney et al. Jul 1999 A
5954650 Saito et al. Sep 1999 A
5976088 Urbano et al. Nov 1999 A
5986662 Argiro et al. Nov 1999 A
5987345 Engelmann et al. Nov 1999 A
5995644 Lai et al. Nov 1999 A
6008813 Lauer et al. Dec 1999 A
6115486 Cantoni Sep 2000 A
6128002 Leiper Oct 2000 A
6130671 Argiro Oct 2000 A
6151581 Kraftson et al. Nov 2000 A
6175643 Lai et al. Jan 2001 B1
6177937 Stockham et al. Jan 2001 B1
6185320 Bick et al. Feb 2001 B1
6211884 Knittel et al. Apr 2001 B1
6219059 Argiro Apr 2001 B1
6219061 Lauer et al. Apr 2001 B1
6243095 Shile et al. Jun 2001 B1
6243098 Lauer et al. Jun 2001 B1
6262740 Lauer et al. Jul 2001 B1
6266733 Knittel et al. Jul 2001 B1
6269379 Hiyama et al. Jul 2001 B1
6297799 Knittel et al. Oct 2001 B1
6304667 Reitano Oct 2001 B1
6310620 Lauer et al. Oct 2001 B1
6313841 Ogata et al. Nov 2001 B1
6342885 Knittel et al. Jan 2002 B1
6347329 Evans Feb 2002 B1
6351547 Johnson et al. Feb 2002 B1
6356265 Knittel et al. Mar 2002 B1
6369816 Knittel et al. Apr 2002 B1
6383135 Chikovani et al. May 2002 B1
6388687 Brackett et al. May 2002 B1
6404429 Knittel Jun 2002 B1
6407737 Zhao et al. Jun 2002 B1
6411296 Knittel et al. Jun 2002 B1
6421057 Lauer et al. Jul 2002 B1
6424346 Correll et al. Jul 2002 B1
6424996 Killcommons et al. Jul 2002 B1
6426749 Knittel et al. Jul 2002 B1
6427022 Craine et al. Jul 2002 B1
6438533 Spackman et al. Aug 2002 B1
6463169 Ino et al. Oct 2002 B1
6476810 Sirnha et al. Nov 2002 B1
6512517 Knittel et al. Jan 2003 B1
6532299 Sachdeva et al. Mar 2003 B1
6532311 Pritt Mar 2003 B1
6556695 Packer et al. Apr 2003 B1
6556724 Chang et al. Apr 2003 B1
6563950 Wiskott et al. May 2003 B1
6574629 Cooke, Jr. et al. Jun 2003 B1
6577753 Ogawa Jun 2003 B2
6603494 Banks et al. Aug 2003 B1
6606171 Renk et al. Aug 2003 B1
6614447 Bhatia et al. Sep 2003 B1
6618060 Brackett Sep 2003 B1
6621918 Hu et al. Sep 2003 B1
6630937 Kallergi et al. Oct 2003 B2
6650766 Rogers Nov 2003 B1
6654012 Lauer et al. Nov 2003 B1
6678764 Parvelescu et al. Jan 2004 B2
6680735 Seiler et al. Jan 2004 B1
6683933 Saito et al. Jan 2004 B2
6697067 Callahan et al. Feb 2004 B1
6697506 Oian et al. Feb 2004 B1
6734880 Chang et al. May 2004 B2
6760755 Brackett Jul 2004 B1
6775402 Bacus et al. Aug 2004 B2
6778689 Aksit et al. Aug 2004 B1
6820093 de la Huerga Nov 2004 B2
6820100 Funahashi Nov 2004 B2
6826297 Saito et al. Nov 2004 B2
6829377 Milioto Dec 2004 B2
6864794 Betz Mar 2005 B2
6886133 Bailey et al. Apr 2005 B2
6891920 Minyard et al. May 2005 B1
6894707 Nemoto May 2005 B2
6909436 Pianykh et al. Jun 2005 B1
6909795 Tecotzky et al. Jun 2005 B2
6917696 Soenksen Jul 2005 B2
6988075 Hacker Jan 2006 B1
6996205 Capolunghi et al. Feb 2006 B2
7016952 Mullen et al. Mar 2006 B2
7022073 Fan et al. Apr 2006 B2
7027633 Foran et al. Apr 2006 B2
7031504 Argiro et al. Apr 2006 B1
7031846 Kaushikkar et al. Apr 2006 B2
7039723 Hu et al. May 2006 B2
7043474 Mojsilovic May 2006 B2
7050620 Heckman May 2006 B2
7054473 Roehrig et al. May 2006 B1
7058901 Hafey et al. Jun 2006 B1
7092572 Huang et al. Aug 2006 B2
7103205 Wang et al. Sep 2006 B2
7106479 Roy et al. Sep 2006 B2
7110616 Ditt et al. Sep 2006 B2
7113186 Kim et al. Sep 2006 B2
7136064 Zuiderveld Nov 2006 B2
7139416 Vuylsteke Nov 2006 B2
7149334 Dehmeshki Dec 2006 B2
7155043 Daw Dec 2006 B2
7162623 Yngvesson Jan 2007 B2
7170532 Sako Jan 2007 B2
7174054 Manber et al. Feb 2007 B2
7209149 Jogo Apr 2007 B2
7209578 Saito et al. Apr 2007 B2
7212661 Samara et al. May 2007 B2
7218763 Belykh et al. May 2007 B2
7224352 Lipton et al. May 2007 B2
7236558 Saito et al. Jun 2007 B2
7260249 Smith Aug 2007 B2
7263710 Hummell et al. Aug 2007 B1
7272610 Torres Sep 2007 B2
7346199 Pfaff Mar 2008 B2
7366992 Thomas, III Apr 2008 B2
7379578 Soussaline et al. May 2008 B2
7412111 Battle et al. Aug 2008 B2
7450747 Jabri et al. Nov 2008 B2
7492970 Saito et al. Feb 2009 B2
7505782 Chu Mar 2009 B2
7516417 Amador et al. Apr 2009 B2
7525554 Morita et al. Apr 2009 B2
7526114 Xia et al. Apr 2009 B2
7526132 Koenig Apr 2009 B2
7545965 Suzuki et al. Jun 2009 B2
7574029 Peterson et al. Aug 2009 B2
7583861 Hanna et al. Sep 2009 B2
7590272 Brejl et al. Sep 2009 B2
7613335 McLennan et al. Nov 2009 B2
7634121 Novatzky et al. Dec 2009 B2
7636413 Toth Dec 2009 B2
7639879 Goto et al. Dec 2009 B2
7656543 Atkins Feb 2010 B2
7660438 Reicher et al. Feb 2010 B2
7660481 Schaap et al. Feb 2010 B2
7668352 Tecotzky et al. Feb 2010 B2
7683909 Takekoshi Mar 2010 B2
7693152 Reid Apr 2010 B2
7716277 Yamatake May 2010 B2
7787672 Reicher et al. Aug 2010 B2
7834891 Yarger et al. Nov 2010 B2
7835560 Vining et al. Nov 2010 B2
7885440 Fram et al. Feb 2011 B2
7885828 Glaser-Seidnitzer et al. Feb 2011 B2
7899514 Kirkland Mar 2011 B1
7920152 Fram et al. Apr 2011 B2
7953614 Reicher May 2011 B1
7970188 Mahesh et al. Jun 2011 B2
7970625 Reicher Jun 2011 B2
7991210 Peterson et al. Aug 2011 B2
7992100 Lundstrom et al. Aug 2011 B2
8019138 Reicher Sep 2011 B2
8046044 Stazzone et al. Oct 2011 B2
8050938 Green, Jr. et al. Nov 2011 B1
8065166 Maresh et al. Nov 2011 B2
8073225 Hagen et al. Dec 2011 B2
8094901 Reicher Jan 2012 B1
8150708 Kotula et al. Apr 2012 B2
8214756 Salazar-Ferrer et al. Jul 2012 B2
8217966 Fram et al. Jul 2012 B2
8244014 Reicher Aug 2012 B2
8249687 Peterson et al. Aug 2012 B2
8262572 Chono Sep 2012 B2
8292811 Relkuntwar et al. Oct 2012 B2
8370293 Iwase et al. Feb 2013 B2
8379051 Brown Feb 2013 B2
8380533 Reicher et al. Feb 2013 B2
8391643 Melbourne et al. Mar 2013 B2
8406491 Gee et al. Mar 2013 B2
8457990 Reicher et al. Jun 2013 B1
8554576 Reicher et al. Oct 2013 B1
8560050 Martin et al. Oct 2013 B2
8610746 Fram et al. Dec 2013 B2
8626527 Reicher Jan 2014 B1
8693757 Gundel Apr 2014 B2
8712120 Reicher et al. Apr 2014 B1
8731259 Reicher May 2014 B2
8751268 Reicher et al. Jun 2014 B1
8797350 Fram Aug 2014 B2
8879807 Fram et al. Nov 2014 B2
8913808 Reicher Dec 2014 B2
9042617 Reicher et al. May 2015 B1
9075899 Reicher Jul 2015 B1
9092551 Reicher Jul 2015 B1
9092727 Reicher Jul 2015 B1
9324188 Fram et al. Apr 2016 B1
9386084 Reicher Jul 2016 B1
9471210 Fram Oct 2016 B1
9495604 Fram Nov 2016 B1
9501617 Reicher Nov 2016 B1
9501627 Reicher Nov 2016 B2
9501863 Fram Nov 2016 B1
9536324 Fram Jan 2017 B1
9542082 Reicher Jan 2017 B1
9672477 Reicher Jun 2017 B1
9684762 Reicher Jun 2017 B2
9727938 Reicher Aug 2017 B1
9734576 Fram et al. Aug 2017 B2
9754074 Reicher Sep 2017 B1
9836202 Reicher Dec 2017 B1
9892341 Reicher Feb 2018 B2
9934568 Reicher Apr 2018 B2
20010016822 Bessette Aug 2001 A1
20010042124 Barron Nov 2001 A1
20020016718 Rothschild et al. Feb 2002 A1
20020021828 Papier et al. Feb 2002 A1
20020039034 Yamaguchi Apr 2002 A1
20020044696 Sirohey et al. Apr 2002 A1
20020070970 Wood et al. Jun 2002 A1
20020073429 Beane et al. Jun 2002 A1
20020090118 Olschewski Jul 2002 A1
20020090119 Saito et al. Jul 2002 A1
20020090124 Soubelet et al. Jul 2002 A1
20020091659 Beaulieu et al. Jul 2002 A1
20020103673 Atwood Aug 2002 A1
20020103827 Sesek Aug 2002 A1
20020110285 Wang et al. Aug 2002 A1
20020144697 Betz Oct 2002 A1
20020172408 Saito et al. Nov 2002 A1
20020172409 Saito et al. Nov 2002 A1
20020180883 Tomizawa et al. Dec 2002 A1
20020186820 Saito et al. Dec 2002 A1
20020190984 Seiler et al. Dec 2002 A1
20030005464 Gropper et al. Jan 2003 A1
20030013951 Stefanescu Jan 2003 A1
20030016850 Kaufman et al. Jan 2003 A1
20030028402 Ulrich et al. Feb 2003 A1
20030034973 Zuiderveld Feb 2003 A1
20030037054 Dutta et al. Feb 2003 A1
20030055896 Hu et al. Mar 2003 A1
20030065613 Smith Apr 2003 A1
20030071829 Bodicker et al. Apr 2003 A1
20030101291 Mussack et al. May 2003 A1
20030115083 Masarie et al. Jun 2003 A1
20030120516 Perednia Jun 2003 A1
20030130973 Sumner, II et al. Jul 2003 A1
20030140141 Mullen et al. Jul 2003 A1
20030156745 Saito et al. Aug 2003 A1
20030160095 Segal Aug 2003 A1
20030164860 Shen et al. Sep 2003 A1
20030184778 Chiba Oct 2003 A1
20030187689 Barnes et al. Oct 2003 A1
20030190062 Noro et al. Oct 2003 A1
20030204420 Wilkes et al. Oct 2003 A1
20030215122 Tanaka Nov 2003 A1
20040015703 Madison et al. Jan 2004 A1
20040024303 Banks et al. Feb 2004 A1
20040068170 Wang et al. Apr 2004 A1
20040086163 Moriyama et al. May 2004 A1
20040088192 Schmidt et al. May 2004 A1
20040105030 Yamane Jun 2004 A1
20040105574 Pfaff Jun 2004 A1
20040114714 Minyard et al. Jun 2004 A1
20040122705 Sabol et al. Jun 2004 A1
20040141661 Hanna et al. Jul 2004 A1
20040143582 Vu Jul 2004 A1
20040161164 Dewaele Aug 2004 A1
20040165791 Kaltanji Aug 2004 A1
20040172306 Wohl et al. Sep 2004 A1
20040174429 Chu Sep 2004 A1
20040190780 Shiibashi et al. Sep 2004 A1
20040202387 Yngvesson Oct 2004 A1
20040243435 Williams Dec 2004 A1
20040252871 Tecotzky et al. Dec 2004 A1
20040254816 Myers Dec 2004 A1
20040255252 Rodriguez et al. Dec 2004 A1
20050010531 Kushalnagar et al. Jan 2005 A1
20050027569 Gollogly et al. Feb 2005 A1
20050027570 Maier et al. Feb 2005 A1
20050043970 Hsieh Feb 2005 A1
20050063575 Ma et al. Mar 2005 A1
20050065424 Shah et al. Mar 2005 A1
20050074150 Bruss Apr 2005 A1
20050074157 Thomas, III Apr 2005 A1
20050075544 Shapiro et al. Apr 2005 A1
20050088534 Shen et al. Apr 2005 A1
20050107689 Sasano May 2005 A1
20050108058 Weidner et al. May 2005 A1
20050110791 Krishnamoorthy et al. May 2005 A1
20050111733 Fors et al. May 2005 A1
20050113681 DeFreitas et al. May 2005 A1
20050114178 Krishnamurthy et al. May 2005 A1
20050114179 Brackett et al. May 2005 A1
20050114283 Pearson et al. May 2005 A1
20050143654 Zuiderveld et al. Jun 2005 A1
20050171818 McLaughlin Aug 2005 A1
20050184988 Yanof et al. Aug 2005 A1
20050197860 Joffe et al. Sep 2005 A1
20050238218 Nakamura Oct 2005 A1
20050244041 Tecotzky et al. Nov 2005 A1
20050251013 Krishnan Nov 2005 A1
20050254729 Saito et al. Nov 2005 A1
20050259118 Mojaver Nov 2005 A1
20050273009 Deischinger et al. Dec 2005 A1
20060008181 Takekoshi Jan 2006 A1
20060031097 Lipscher et al. Feb 2006 A1
20060050152 Rai et al. Mar 2006 A1
20060058603 Dave et al. Mar 2006 A1
20060061570 Cheryauka et al. Mar 2006 A1
20060093207 Reicher May 2006 A1
20060095423 Reicher May 2006 A1
20060095426 Takachio et al. May 2006 A1
20060111941 Blom May 2006 A1
20060122482 Mariotti et al. Jun 2006 A1
20060171574 DelMonego et al. Aug 2006 A1
20060181548 Hafey Aug 2006 A1
20060188134 Quist Aug 2006 A1
20060230072 Partovi et al. Oct 2006 A1
20060241979 Sato et al. Oct 2006 A1
20060267976 Saito et al. Nov 2006 A1
20060274145 Reiner Dec 2006 A1
20060276708 Peterson et al. Dec 2006 A1
20060277075 Salwan Dec 2006 A1
20060282408 Wisely et al. Dec 2006 A1
20070009078 Saito et al. Jan 2007 A1
20070021977 Elsholz Jan 2007 A1
20070050701 El Emam et al. Mar 2007 A1
20070055550 Courtney et al. Mar 2007 A1
20070064984 Vassa et al. Mar 2007 A1
20070067124 Kimpe et al. Mar 2007 A1
20070073556 Lau et al. Mar 2007 A1
20070106535 Matsunaga May 2007 A1
20070106633 Reiner May 2007 A1
20070109299 Peterson May 2007 A1
20070109402 Niwa May 2007 A1
20070110294 Schaap et al. May 2007 A1
20070116345 Peterson et al. May 2007 A1
20070116346 Peterson et al. May 2007 A1
20070122016 Brejl et al. May 2007 A1
20070124541 Lang et al. May 2007 A1
20070140536 Sehnert Jun 2007 A1
20070159962 Mathavu et al. Jul 2007 A1
20070162308 Peters Jul 2007 A1
20070165917 Cao et al. Jul 2007 A1
20070174079 Kraus Jul 2007 A1
20070192138 Saito et al. Aug 2007 A1
20070192140 Gropper Aug 2007 A1
20070237380 Iwase et al. Oct 2007 A1
20070239481 DiSilvestro et al. Oct 2007 A1
20080016111 Keen Jan 2008 A1
20080021877 Saito et al. Jan 2008 A1
20080059245 Sakaida et al. Mar 2008 A1
20080100612 Dastmalchi et al. May 2008 A1
20080103828 Squilla et al. May 2008 A1
20080125846 Battle et al. May 2008 A1
20080126982 Sadikali et al. May 2008 A1
20080136838 Goede et al. Jun 2008 A1
20080275913 van Arragon et al. Nov 2008 A1
20080279439 Minyard et al. Nov 2008 A1
20080300484 Wang et al. Dec 2008 A1
20090005668 West Jan 2009 A1
20090022375 Fidrich Jan 2009 A1
20090028410 Shimazaki Jan 2009 A1
20090080719 Watt Mar 2009 A1
20090091566 Turney Apr 2009 A1
20090123052 Ruth et al. May 2009 A1
20090129643 Natanzon et al. May 2009 A1
20090132586 Napora et al. May 2009 A1
20090150481 Garcia et al. Jun 2009 A1
20090182577 Squilla et al. Jul 2009 A1
20090193514 Rhodes Aug 2009 A1
20090213034 Wu et al. Aug 2009 A1
20090248442 Pacheco et al. Oct 2009 A1
20090268986 Holstein et al. Oct 2009 A1
20090326373 Boese Dec 2009 A1
20100053353 Hunter et al. Mar 2010 A1
20100086182 Luo et al. Apr 2010 A1
20100131887 Salazar-Ferrer et al. May 2010 A1
20100198608 Kaboff et al. Aug 2010 A1
20100201714 Reicher Aug 2010 A1
20100211409 Kotula et al. Aug 2010 A1
20100246981 Hu et al. Sep 2010 A1
20100299157 Fram et al. Nov 2010 A1
20110110572 Guehring et al. May 2011 A1
20110293162 Pajeau Dec 2011 A1
20110316873 Reicher Dec 2011 A1
20120070048 Van Den Brink Mar 2012 A1
20120130729 Raizada et al. May 2012 A1
20120136794 Kushalnagar et al. May 2012 A1
20120163684 Natanzon et al. Jun 2012 A1
20120194540 Reicher Aug 2012 A1
20120284657 Hafey et al. Nov 2012 A1
20130070998 Shibata Mar 2013 A1
20130076681 Sirpal et al. Mar 2013 A1
20130083023 Fram Apr 2013 A1
20130159019 Reicher Jun 2013 A1
20130169661 Reicher Jul 2013 A1
20130297331 Zuehlsdorff et al. Nov 2013 A1
20140022194 Ito Jan 2014 A1
20140142983 Backhaus et al. May 2014 A1
20150101066 Fram Apr 2015 A1
20150160848 Gkanatsios et al. Jun 2015 A1
20160034110 Edwards Feb 2016 A1
20160335395 Wu et al. Nov 2016 A1
20170038951 Reicher Feb 2017 A1
20170039321 Reicher Feb 2017 A1
20170039322 Reicher Feb 2017 A1
20170039350 Reicher et al. Feb 2017 A1
20170039705 Fram Feb 2017 A1
20170046014 Fram Feb 2017 A1
20170046483 Reicher Feb 2017 A1
20170046485 Reicher Feb 2017 A1
20170046495 Fram Feb 2017 A1
20170046870 Fram Feb 2017 A1
20170053404 Reicher Feb 2017 A1
20170200064 Reicher Jul 2017 A1
20170200269 Reicher Jul 2017 A1
20170200270 Reicher Jul 2017 A1
20170206324 Reicher Jul 2017 A1
20170293720 Reicher Oct 2017 A1
20170308647 Reicher et al. Oct 2017 A1
20180059918 Reicher Mar 2018 A1
Foreign Referenced Citations (1)
Number Date Country
WO 2007131157 Nov 2007 WO
Non-Patent Literature Citations (306)
Entry
US 7,801,341, 09/2010, Fram et al. (withdrawn)
US 8,208,705, 06/2012, Reicher et al. (withdrawn)
U.S. Appl. No. 14/540,830, Systems and Methods for Viewing Medical Images, filed Nov. 13, 2014.
U.S. Appl. No. 15/254,627, Systems and Methods for Interleaving Series of Medican Images, filed Sep. 1, 2016.
U.S. Appl. No. 14/095,123, Systems and Methods for Retrieval of Medical Data, filed Dec. 3, 2013.
U.S. Appl. No. 15/292,006, Systems and Methods for Viewing Medical 3D Imaging Volumes, filed Oct. 12, 2016.
U.S. Appl. No. 15/346,560, Systems and Methods for Matching, Naming, and Displaying Medical Images, filed Nov. 8, 2016.
U.S. Appl. No. 14/298,806, Smart Placement Rules, filed Jun. 6, 2014.
U.S. Appl. No. 11/942,687, Smart Forms, filed Nov. 19, 2007.
U.S. Appl. No. 14/043,165, Automated Document Filings, filed Oct. 1, 2013.
U.S. Appl. No. 15/475,930, Exam Scheduling With Customer Configured Notifications, filed Mar. 31, 2017.
U.S. Appl. No. 15/292,014, System and Method of Providing Dynamic and Customizable Medical Examination for, filed Oct. 12, 2016.
U.S. Appl. No. 15/469,342, Rules-Based Rendering of Medical Images, filed Mar. 24, 2017.
U.S. Appl. No. 15/469,281, Rules-Based Processing and Presentation of Medical Images, filed Mar. 24, 2017.
U.S. Appl. No. 15/469,296, filed Computer-Aided Analysis and Rendering of Medical Images, filed Mar. 24, 2017.
U.S. Appl. No. 14/792,210, Dynamic Montage Reconstruction, filed Jul. 6, 2015.
U.S. Appl. No. 15/188,872, Intelligent Management of Computerized Advanced Processing, filed Jun. 21, 2016.
U.S. Appl. No. 15/188,819, Intelligent Management of Computerized Advanced Processing, filed Jun. 21, 2016.
U.S. Appl. No. 15/140,346, Database Systems and Interactive User Interfaces for Dynamic Interaction With, and Sorting of, Digital Medical Image Data, filed Apr. 27, 2016.
U.S. Appl. No. 15/140,363, Database Systems and Interactive User Interfaces for Dynamic Interation With, and Comparison of, Digital Medical Image Data, filed Apr. 27, 2016.
U.S. Appl. No. 15/140,351, Database Systems and Interactive User Interfaces for Dynamic Interaction With, and Review of, Digital Medical Image Data, filed Apr. 27, 2016.
U.S. Appl. No. 15/140,348, Database Systems and Interactive User Interfaces for Dynamic Interaction With, and Indications of, Digital Medical Image Data, filed Apr. 27, 2016
Non-Final Office Action dated Aug. 28, 2007 in U.S. Appl. No. 11/179,384.
Final Office Action dated Jun. 26, 2008 in U.S. Appl. No. 11/179,384.
Non-Final Office Action dated Dec. 29, 2008 in U.S. Appl. No. 11/179,384.
Final Office Action dated Jul. 24, 2009, in U.S. Appl. No. 11/179,384.
Notice of Allowance dated Nov. 3, 2009, in U.S. Appl. No. 11/179,384.
Non-Final Office Action dated Aug. 18, 2010 in U.S. Appl. No. 12/702,976.
Interview Summary dated Dec. 1, 2010, in U.S. Appl. No. 12/702,976.
Final Office Action dated Feb. 17, 2011 in U.S. Appl. No. 12/702,976.
Interview Summary dated May 31, 2011 in U.S. Appl. No. 12/702,976.
Notice of Allowance dated Jul. 20, 2011, in U.S. Appl. No. 12/702,976.
Office Action dated Dec. 1, 2011, in U.S. Appl. No. 13/228,349.
Notice of Allowance dated Feb. 6, 2012, in U.S. Appl. No. 13/228,349.
Notice of Allowance dated Jul. 20, 2012, in U.S. Appl. No. 13/228,349.
Office Action dated Dec. 11, 2013, in U.S. Appl. No. 13/477,853.
Interview Summary dated Mar. 14, 2014, in U.S. Appl. No. 13/477,853.
Final Office Action dated Jun. 13, 2014, in U.S. Appl. No. 13/477,853.
Notice of Allowance dated Aug. 15, 2014, in U.S. Appl. No. 13/477,853.
Office Action dated Jan. 17, 2017, in U.S. Appl. No. 14/540,830.
Interview Summary dated Mar. 24, 2017, in U.S. Appl. No. 14/540,830.
Final Office Action dated May 15, 2017, in U.S. Appl. No. 14/540,830.
Non-Final Office Action dated Oct. 1, 2009, in U.S. Appl. No. 11/268,261.
Notice of Allowance dated Feb. 2, 2010, in U.S. Appl. No. 11/268,261.
Interview Summary dated Jan. 25, 2010, in U.S. Appl. No. 11/268,261.
Interview Summary dated May 14, 2010, in U.S. Appl. No. 11/268,261.
Notice of Allowance dated May 17, 2010, in U.S. Appl. No. 11/268,261.
Supplemental Notice of Allowance dated Aug. 6, 2010, in U.S. Appl. No. 11/268,261.
Notice of Allowance dated Oct. 8, 2010, in U.S. Appl. No. 11/268,261.
Notice of Allowance dated Dec. 3, 2010, in U.S. Appl. No. 11/268,261.
Notice of Allowance dated Jan. 6, 2011, in U.S. Appl. No. 11/268,261.
Office Action dated May 16, 2011, in U.S. Appl. No. 12/857,915.
Interview Summary dated Sep. 6, 2011, in U.S. Appl. No. 12/857,915.
Final Office Action dated Dec. 15, 2011, in U.S. Appl. No. 12/857,915.
Office Action dated Jun. 12, 2012, in U.S. Appl. No. 12/857,915.
Office Action dated Aug. 23, 2013, in U.S. Appl. No. 12/857,915.
Interview Summary dated Feb. 4, 2014, in U.S. Appl. No. 12/857,915.
Notice of Allowance dated Jul. 3, 2014, in U.S. Appl. No. 12/857,915.
“Corrected” Notice of Allowance dated Aug. 15, 2014, in U.S. Appl. No. 12/857,915.
Non-Final Office Action dated Jan. 20, 2016, in U.S. Appl. No. 14/502,055.
Interview Summary dated Apr. 14, 2016, in U.S. Appl. No. 14/502,055.
Notice of Allowance dated Jun. 2, 2016, in U.S. Appl. No. 14/502,055.
Notice of Corrected Allowability dated Jul. 14, 2016, in U.S. Appl. No. 14/502,055.
Notice of Corrected Allowability dated Sep. 19, 2016, in U.S. Appl. No. 14/502,055.
Office Action dated Dec. 12, 2016, in U.S. Appl. No. 15/254,627.
Notice of Allowance dated Apr. 3, 2017 in U.S. Appl. No. 15/254,627.
Non-Final Office Action dated May 13, 2009, in U.S. Appl. No. 11/265,979.
Final Office Action dated Dec. 22, 2009 in U.S. Appl. No. 11/265,979.
Non-Final Office Action dated Jul. 8, 2010 in U.S. Appl. No. 11/265,979.
Interview Summary dated Mar. 4, 2010 in U.S. Appl. No. 11/265,979.
Interview Summary dated Nov. 16, 2010 in U.S. Appl. No. 11/265,979.
Final Office Action dated Dec. 23, 2010 in U.S. Appl. No. 11/265,979.
Interview Summary dated Mar. 17, 2011 in U.S. Appl. No. 11/265,979.
Notice of Allowance dated May 26, 2011 in U.S. Appl. No. 11/265,979.
Office Action dated Jun. 8, 2012 in U.S. Appl. No. 13/171,081.
Interview Summary dated Jul. 31, 2012 in U.S. Appl. No. 13/171,081.
Final Office Action dated Oct. 12, 2012 in U.S. Appl. No. 13/171,081.
Interview Summary dated Nov. 6, 2012 in U.S. Appl. No. 13/171,081.
Notice of Allowance, dated Sep. 4, 2013, in U.S. Appl. No. 13/171,081.
Office Action dated Mar. 3, 2015 in U.S. Appl. No. 14/095,123.
Interview Summary dated May 1, 2015 in U.S. Appl. No. 14/095,123.
Final Office Action dated Jul. 23, 2015 in U.S. Appl. No. 14/095,123.
Interview Summary dated Aug. 27, 2015 in U.S. Appl. No. 14/095,123.
Office Action dated Feb. 23, 2016 in U.S. Appl. No. 14/095,123.
Final Office Action dated Jul. 20, 2016 in U.S. Appl. No. 14/095,123.
Notice of Allowance dated Mar. 30, 2017 in U.S. Appl. No. 14/095,123.
Non-Final Office Action dated Aug. 24, 2009 in U.S. Appl. No. 11/268,262.
Non-Final Office Action dated Apr. 16, 2010 in U.S. Appl. No. 11/268,262.
Interview Summary dated Nov. 24, 2009 in U.S. Appl. No. 11/268,262.
Interview Summary dated May 12, 2010 in U.S. Appl. No. 11/268,262.
Final Office Action dated Oct. 28, 2010 in U.S. Appl. No. 11/268,262.
Interview Summary dated Dec. 1, 2010 in U.S. Appl. No. 11/268,262.
Notice of Allowance dated Dec. 1, 2010 in U.S. Appl. No. 11/268,262.
Notice of Allowance dated Feb. 25, 2011 in U.S. Appl. No. 11/268,262.
Non-Final Office Action dated Jan. 11, 2012 in U.S. Appl. No. 13/079,597.
Notice of Allowance dated Apr. 25, 2012, in U.S. Appl. No. 13/079,597.
Non-Final Office Action dated Apr. 4, 2013 in U.S. Appl. No. 13/535,758.
Notice of Allowance, dated Aug. 23, 2013 in U.S. Appl. No. 13/535,758.
Corrected Notice of Allowance dated Jun. 27, 2016, in U.S. Appl. No. 14/502,055.
Office Action dated Mar. 10, 2016 in U.S. Appl. No. 14/081,225.
Notice of Allowance dated Sep. 2, 2016 in U.S. Appl. No. 14/081,225.
Corrected Notice of Allowance dated Oct. 21, 2016 in U.S. Appl. No. 14/081,225.
Non-Final Office Action dated Jul. 27, 2009 in U.S. Appl. No. 11/265,978.
Notice of Allowance dated Nov. 19, 2009 in U.S. Appl. No. 11/265,978.
Notice of Allowance dated Apr. 19, 2010 in U.S. Appl. No. 11/265,978.
Supplemental Notice of Allowance dated May 3, 2010 in U.S. Appl. No. 11/265,978.
Supplemental Notice of Allowance dated Aug. 3, 2010 in U.S. Appl. No. 11/265,978.
Non-Final Office Action dated May 5, 2011 in U.S. Appl. No. 12/870,645.
Non-Final Office Action dated May 31, 2013, in U.S. Appl. No. 13/345,606.
Interview Summary dated Aug. 15, 2013, in U.S. Appl. No. 13/345,606.
Notice of Allowance, dated Jan. 9, 2014 in U.S. Appl. No. 13/345,606.
Non-Final Office Action dated Mar. 18, 2016 in U.S. Appl. No. 14/244,431.
Interview Summary dated Jun. 17, 2016 in U.S. Appl. No. 14/244,431.
Notice of Allowance dated Aug. 18, 2016 in U.S. Appl. No. 14/244,431.
Corrected Notice of Allowance dated Nov. 16, 2016 in U.S. Appl. No. 14/244,431.
Non-Final Office Action dated May 26, 2010 in U.S. Appl. No. 11/942,674.
Interview Summary dated Jul. 26, 2010 in U.S. Appl. No. 11/942,674.
Final Office Action dated Nov. 26, 2010 in U.S. Appl. No. 11/942,674.
Interview Summary dated Mar. 2, 2011 in U.S. Appl. No. 11/942,674.
Notice of Allowance, dated Apr. 1, 2011 in U.S. Appl. No. 11/942,674.
Non Final Office Action dated Nov. 10, 2011 in U.S. Appl. No. 13/118,085.
Interview Summary, dated Feb. 17, 2012, in U.S. Appl. No. 13/118,085.
Final Office Action, dated Apr. 13, 2012, in U.S. Appl. No. 13/118,085.
Notice of Allowance, dated Feb. 6, 2013 in U.S Appl. No. 13/118,085.
Non Final Office Action dated Aug. 23, 2013 in U.S. Appl. No. 13/907,128.
Final Office Action dated Oct. 9, 2013 in U.S. Appl. No. 13/907,128.
Interview Summary dated Nov. 22, 2013 in U.S. Appl. No. 13/907,128.
Notice of Allowance dated Jan. 31, 2014 in U.S. Appl. No. 13/907,128.
Office Action, dated Dec. 29, 2014 in U.S. Appl. No. 14/298,806.
Interview Summary, dated Mar. 2, 2015 in U.S. Appl. No. 14/298,806.
Final Office Action, dated Jun. 17, 2015 in U.S. Appl. No. 14/298,806.
Office Action, dated Feb. 16, 2016 in U.S. Appl. No. 14/298,806.
Final Office Action, dated Jul. 21, 2016 in U.S. Appl. No. 14/298,806.
Notice of Allowance, dated Apr. 12, 2017 in U.S. Appl. No. 14/298,806.
Non Final Office Action dated Sep. 16, 2010 in U.S. Appl. No. 11/942,687.
Interview Summary dated Dec. 3, 2010 in U.S. Appl. No. 11/942,687.
Final Office Action, dated Apr. 5, 2011 in U.S. Appl. No. 11/942,687.
Office Action, dated Mar. 13, 2014 in U.S. Appl. No, 11/942,687.
Interview Summary, dated Jun. 17, 2014 in U.S. Appl. No. 11/942,687.
Office Action, dated Jul. 18, 2014 in U.S. Appl. No. 11/942,687.
Final Office Action, dated Jan. 5, 2015 in U.S. Appl. No. 11/942,687.
Interview Summary, dated Mar. 4, 2015 in U.S. Appl. No. 11/942,687.
PTAB Examiner's Answer, dated Feb. 25, 2016 in U.S. Appl. No. 11/942,687.
Non-Final Office Action dated Apr. 14, 2010 in U.S. Appl. No. 11/944,027.
Interview Summary dated May 13, 2010 in U.S. Appl. No. 11/944,027.
Final Office Action dated Dec. 23, 2010 in U.S. Appl. No. 11/944,027.
Interview Summary dated Mar. 31, 2011 in U.S. Appl. No. 11/944,027.
Office Action dated Apr. 19, 2012 in U.S. Appl. No. 11/944,027.
Interview Summary dated Jun. 28, 2012 in U.S. Appl. No. 11/944,027.
Final Office Action dated Oct. 22, 2012 in U.S. Appl. No. 11/944,027.
Notice of Allowance dated Jun. 5, 2013 in U.S. Appl. No. 11/944,027.
Office Action dated Oct. 14, 2014 in U.S. Appl. No. 14/043,165.
Final Office Action dated Apr. 1, 2015 in U.S. Appl. No. 14/043,165.
Office Action dated Oct. 2, 2015 in U.S. Appl. No. 14/043,165.
Interview Summary dated Dec. 21, 2015 in U.S. Appl. No. 14/043,165.
Final Office Action dated Feb. 17, 2016 in U.S. Appl. No. 14/043,165.
Appeal Brief dated Jul. 15, 2016 in U.S. Appl. No. 14/043,165.
Examiner's Answer dated Nov. 14, 2016, in U.S. Appl. No. 14/043,165.
Non-Final Office Action dated Sep. 29, 2010 in U.S. Appl. No. 11/944,000.
Final Office Action dated Apr. 20, 2011 in U.S. Appl. No. 11/944,000.
Interview Summary dated Jun. 7, 2011 in U.S. Appl. No. 11/944,000.
Appeal Brief dated Mar. 4, 2013 in U.S. Appl. No. 11/944,000.
Examiner's Answer dated Jun. 26, 2013 in U.S. Appl. No. 11/944,000.
Board Decision dated Mar. 23, 2016 in U.S. Appl. No. 11/944,000.
Office Action, dated Jul. 15, 2016 in U.S. Appl. No. 11/944,000.
Notice of Allowance, dated Jan. 30, 2017, in U.S. Appl. No. 11/944,000.
Office Action dated Feb. 3, 2012 in U.S. Appl. No. 12/622,404.
Interview Summary dated May 8, 2012 in U.S. Appl. No. 12/622,404.
Final Office Action dated Aug. 6, 2012 in U.S. Appl. No. 12/622,404.
Notice of Allowance dated Oct. 15, 2012 in U.S. Appl. No. 12/622,404.
Office Action dated Mar. 17, 2015 in U.S. Appl. No. 13/768,765.
Interview Summary dated Jun. 11, 2015 in U.S. Appl. No. 13/768,765.
Notice of Allowance dated Aug. 28, 2015 in U.S. Appl. No. 13/768,765.
Notice of Allowability dated Nov. 20, 2015 in U.S. Appl. No. 13/768,765.
Notice of Allowability dated Jul. 28, 2016 in U.S. Appl. No. 13/768,765.
Office Action dated Mar. 4, 2013 in U.S. Appl. No. 12/891,543.
Interview Summary dated Apr. 5, 2013 in U.S. Appl. No. 12/891,543.
Notice of Allowance dated Nov. 14, 2013 in U.S. Appl. No. 12/891,543.
Office Action dated Sep. 11, 2014 in U.S. Appl. No. 14/179,328.
Notice of Allowance dated Jan. 14, 2015 in U.S. Appl. No. 14/179,328.
Office Action dated Aug. 13, 2015 in U.S. Appl. No. 14/687,853.
Notice of Allowance dated Feb. 25, 2016 in U.S. Appl. No. 14/687,853.
Supplemental Notice of Allowance dated Jun. 2, 2016 in U.S. Appl. No. 14/687,853.
Notice of Allowance dated Aug. 11, 2016 in U.S. Appl. No. 15/163,600.
Supplemental Notice of Allowance dated Sep. 14, 2016 in U.S. Appl. No. 15/163,600.
Office Action, dated Jan. 12, 2017 in U.S. Appl. No. 15/292,023.
Notice of Allowance, dated Apr. 11, 2017 in U.S. Appl. No. 15/292,023.
Office Action dated Jun. 27, 2014 in U.S. Appl. No. 13/572,397.
Final Office Action dated Jan. 13, 2015 in U.S. Appl. No. 13/572,397.
Notice of Allowance dated Mar. 19, 2015, 2015 in U.S. Appl. No. 13/572,397.
Office Action dated Aug. 6, 2014 in U.S. Appl. No. 13/572,547.
Notice of Allowance, dated Mar. 3, 2015 in U.S. Appl. No. 13/572,547.
Corrected Notice of Allowance, dated Apr. 10, 2015 in U.S. Appl. No. 13/572,547.
Corrected Notice of Allowance, dated May 21, 2015 in U.S. Appl. No. 13/572,547.
Office Action dated Jul. 30, 2014 in U.S. Appl. No. 13/572,552.
Interview Summary dated Sep. 3, 2014 in U.S. Appl. No. 13/572,552.
Final Office Action dated Jan. 28, 2015 in U.S. Appl. No. 13/572,552.
Interview Summary dated Apr. 23, 2015 in U.S. Appl. No. 13/572,552.
Notice of Allowance, dated May 8, 2015 in U.S. Appl. No. 13/572,552.
Restriction Requirement, dated Jul. 28, 2015 in U.S. Appl. No. 14/139,068.
Office Action, dated Mar. 11, 2016 in U.S. Appl. No. 14/139,068.
Notice of Allowance, dated Sep. 21, 2016 in U.S. Appl. No. 14/139,068.
AGFA HealthCare, color brochure “IMPAX 6: Digital Image and Information Management,” © 2012 Agfa HealthCare N.V. Downloaded from http://www.agfahealthcare.com/global/en/he/library/libraryopen?ID=32882925. Accessed on Feb. 9, 2015.
AGFA HealthCare, IMPAX 6.5 Datasheet (US)2012, © 2012 Agfa HealthCare N.V. Downloaded from http://www.agfahealthcare.com/global/en/he/library/libraryopen?ID=37459801. Accessed on Feb. 9, 2015.
AMD Technologies, Inc., Catella PACS 5.0 Viewer User Manual (112 pgs), © 2010, AMD Technologies, Inc. (Doc. 340-3-503 Rev. 01). Downloaded from http://www.amdtechnologies.com/lit/cat5viewer.pdf. Accessed on Feb. 9, 2015.
ASPYRA's Imaging Solutions, 3 page color print out. Accessed at http://www.aspyra.com/imaging-solutions. Accessed on Feb. 9, 2015.
Avreo, interWorks—RIS/PACS package, 2 page color brochure, © 2014, Avreo, Inc. (Document MR-5032 Rev, 4). Downloaded from http://www.avreo.com/ProductBrochures/MR-5032Rev.%204interWORKS%20RISPACSPackage.pdf. Accessed on Feb. 9, 2015.
BRIT Systems, BRIT PACS View Viewer, 2 page color brochure, (BPB-BPV-0001). Downloaded from http://www.brit.com/pdfs/britpacsview.pdf. Accessed on Feb. 9, 2015.
BRIT Systems, Roentgen Works—100% Browers-based VNA (Vendor Neutral Archive/PACS), © 2010 BRIT Systems, 1 page color sheet. Accessed at http://www.roentgenworks.com/PACS. Accessed on Feb. 9, 2015.
BRIT Systems, Vision Multi-modality Viewer—with 3D, 2 page color brochure, (BPB-BVV-0001 REVC). Downloaded from http://wwm.brit.com/pdfs/BPB-BVV-0001REVC_BRIT_Vision_Viewer.pdf. Accessed on Feb. 9, 2015.
CANDELiS, ImageGrid™: Image Management Appliance, 6 page color brochure. (AD-012 Rev. F Nov. 2012), © 2012 Candelis, Inc. Downloaded from http://www.candelis.com/images/pdf/Canidelis_ImageGrid_Appliance_20111121.pdf. Accessed on Feb, 9, 2015.
Carestream, Cardiology PACS, 8 page color brochure. (CAT 866 6075 Jun. 2012). © Carestream Health, Inc., 2012. Downloaded from http://www.carestream.com/cardioPACS_brochure_M1-877.pdf. Accessed on Feb. 9, 2015.
Carestream, Vue PACS, 8 page color brochure. (CAT 300 1035 May 2014). © Carestream Health, Inc., 2014. Downloaded from http://www.carestream.com/csPACS_brochure_M1-876,pdf. Accessed on Feb. 9, 2015.
Cemer, Radiology—Streamline image management, 2 page color brochure, (fl03_332_10_v3). Downloaded from http://www.cerner.com/uploadedFiles/Clinical_Imaging.pdf. Accessed on Feb. 9, 2015.
CoActiv, Exam-PACS, 2 page color brochure, © 2014 CoActiv, LLC. Downloaded from http://coactiv.com/wp-content/uploads/2013/08/EXAM-PACS-BROCHURE-final-web.pdf. Accessed on Feb. 9, 2015.
Crowley, Rebecca et al., Development of Visual Diagnostic Expertise in Pathology: an Information-processing Study, Jan. 2003, Journal of the American medical informatics Association, vol. 10, No. 1, pp. 39-51.
DR Systems, Dominator™ Guide for Reading Physicians, Release 8.2, 546 pages, (TCP-000260-A), © 1997-2009, DR Systems, Inc. Downloaded from https://resources.dominator.com/004/6999.pdf. Document accessed Feb. 9, 2015.
DR Systems, DR Scheduler User Guide, Release 8.2, 410 pages, (TCP-000115-A), © 1997-2009, DR Systems, Inc, Downloaded from https://resources.dominator.com/assets/003/6850,pdf, Document accessed Feb. 9, 2015.
Erickson, et al.: “Effect of Automated Image Registration on Radiologist Interpretation,” Journal of Digital Imaging, vol. 20, No. 2 Jun. 2007; pp. 105-113.
Erickson, et al.: “Image Registration Improves Confidence and Accuracy of Image Interpretation,” Special Issue-Imaging Informatics, Cancer Informatics 2007: 1 19-24.
FUJIFilmMedical Systems, SYNAPSE® Product Data, Synapse Release Version 3.2.1, Foundation Technologies, 4 page color brochure, (XBUSSY084) Aug. 2008. Downloaded from http://www.fujifilmusa.com/shared/bin/foundation.pdf. Accessed on Feb. 9, 2015.
FUJIFilm Medical Systems, SYNAPSE® Product Data, Synapse Release Version 3.2.1, Server Modules and Interfaces, 4 page color brochure, (XBUSSY085) Aug. 2008, Downloaded from http://www.fujifilmusa.com/shared/bin/server-interface.pdf. Accessed on Feb. 9, 2015.
FUJIFilm Medical Systems, Synapse® Product Data, Synapse Release Version 3.2.1, Workstation Software, 4 page color brochure, (XBUSSY082) Aug. 2008. Downloaded from http://www.fujifilmusa.com/shared/bin/workstation.pdf. Accessed on Feb. 9, 2015.
GE Healthcare, Centricity PACS, in 8 page printout. Accessed at http://www3.gehealthcare.com/en/products/categories/healthcare_it/medical_imaging_informatics_-_ris-pacs-cvis/centricity_pacs. Accessed on Feb. 9, 2015.
Handylife.com—Overview of Handy Patients Enterprise, in 2 page printout. Accessed from http://www.handylife.com/en/software/overview.html. Accessed on Feb. 18, 2015.
Handylife.com—Features of Handy Patients Enterprise, in 4 page printout. Accessed from http://www.handylife.com/en/software/features.html. Accessed on Feb. 18, 2015.
Handylife.com—Screenshots of Handy Patients Enterprise, in 2 page printout. Accessed from http://www.handylife.com/en/software/screenshots.html. Accessed on Feb. 18, 2015.
ICRco, I See the Future, in 2 pages, color brochure, (BR080809AUS), © iCRco.ClarityPACS. Downloaded from http://www.claritypacs.com/pdf/ISeeFuture_26_Web.pdf. Accessed on Feb. 9, 2015.
Imageanalysis, dynamika, 2 page color brochure. Downloaded from http://www.imageanalysis.org.uk/what-we-do. Accessed on Feb. 9, 2015.
Imageanalysis, MRI Software, in 5 page printout. Accessed at http://www.imageanalysis.org.uk/mri-software. Accessed on Feb. 9, 2015.
IMIS, Integrated Modular Systems, Inc., Hosted / Cloud PACS in one page printout. Accessed at http://www.imsimed.com/#lproducts-services/ctnu. Accessed on Feb. 9, 2015.
Infinitt, PACS, RIS, Mammo PACS, Cardiology suite and 3D/Advanced Visualization | Infinittna, 2 page printout. Accessed at http://www.infinittna.com/products/radiology-pacs. Accessed on Feb. 9, 2015.
Intelerad, IntelePACS, 2 page color brochure, © 2014 Intelerad Medical Systems Incorporated. Downloaded http://www.intelerad.com/wp-content/uploads/sites/2/2014/08/IntelePACS-brochure.pdf. Accessed on Feb. 9, 2015.
Intelerad, InteleViewer, 2 page color brochure, © Intelerad Medical Systems Incoprorated. Downloaded from http://www.intelerad.com/wp-content/uploads/sites/2/2014/09/InteleViewer-brochure.pdf. Accessed on Feb. 9, 2015.
Intuitive Imaging Informatics, ImageQube, 1 page in color. Downloaded from http://www.intuitiveimaging.com/2013/pdf/ImageQube%20one-sheet.pdf. Accessed on Feb. 9, 2015.
Imaging Technology News, itnonline.com, Comparison Chart PACS, May 2012, pp. 24-27. Downloaded from http://www.merge.corn/MergeHealthcare/media/company/In%20The%20News/merge-pacs-comparison.pdf. Accessed on Feb. 9, 2015.
Kuhl, Helen: Comparison Chart/PACS, Customers Are Happy, But Looking for More, (color) Imaging Techology News, itnonline.com, May 2012, pp. 24-27. Downloaded from http://www.merge.com/MergeHealthcare/media/company/In%20The%20News/merge-pacs-comparison.pdf. Accessed on Feb. 9, 2015.
LUMEDX CardioPACS 5.0 Web Viewer, Cardiopacs Module, 2 page color brochue, (506-10011 Rev A). Downloaded from http://cdn.medicexchange.com/images/whitepaper/cardiopacs_web_viewer.pdf?1295436926. Accessed on Feb. 9, 2015.
LUMEDX Cardiovascular Information System, CardioPACS, one page in color printout. Accessed at http://www.lumedx..com/pacs.aspx. Accessed on Feb. 9, 2015.
McKesson Enterprise Medical imagining and PACS | McKesson, 1 page (color) printout. Accessed at http://www.mckesson.com/providers/health-systems/diagnostic-imaging/enterprise-medical-imaging. Accessed on Feb. 9, 2015.
Medweb Radiology Workflow Solutions, Radiology Workflow Solutions, Complete Workflow & Flexible Turnkey Solutions, Web RIS/PACS with Advanced Viewer, 3 page color brochure, © 2006-2014 Medweb. Downloaded from http://www.medweb.com/docs/rispacs_brochure_2014.pdf. Accessed on Feb. 9, 2015.
Mendelson, et al., “Informatics in Radiology—Image Exchange: IHE and the Evolution of Image Sharing,” RadioGraphics, Nov.-Dec. 2008, vol. 28, No. 7.
Merge Radiology Solutions, Merge PACS, A real-time picture archiving communication system, (PAX-21990 rev 2.0), 2 page color brochure. Downloaded from http://www.merge.com/MergeHealthcare/media/documents/brochures/Merge_PACS-web.pdf, Accessed on Feb. 9, 2015.
NOVARAD Enterprise Imaging Solutions, NOVAPACS, 2 page (color) printout. Accessed at http://ww1.novarad.net/novapacs. Accessed on Feb. 9, 2015.
PACSPLUS, PACSPLUS Server, 1 page (color) printout. Accessed at http://www.pacsplus.com/01_products/products_01.html. Accessed on Feb. 9, 2015.
PACSPLUS, PACSPLUS Workstation, 3 page (color) printout. Accessed at http://www.pacsplus.com/01_products/products_01.html. Accessed on Feb. 9, 2015.
Philips, IntelliSpace PACS, in 2 color page printout. Accessed at https://www.healthcare.phillips.com/main/products/healthcare_informatics/products/enterprise_imaging_informatics/isite_pacs. Accessed on Feb. 9, 2015.
Philips, IntelliSpace: Multi-modality tumor tracking application versus manual PACS methods, a time study for Response Evaluation Criteria in Solid Tumors (RECIST). 2012, Koninklijke Philips Electronics N.V., in four pages.
Radcliffe, et al., “Comparison of Stereo Disc Photographs and Alternation Flicker Using a Novel Matching Technology for Detecting Glaucoma Progression”, Ophthalmic Surgery, Lasers & Imaging, Jun. 9, 2010.
RamSoft, RIS PACS Teleradiology, PowerServer PACS, Lite PACS, XU PACS Compare RamSoft PACS Products, 2 color page printout, Accessed at http://www.ramsoft.com/products/powerserver-pacs-overview. Accessed on Feb. 9, 2015.
Rosset et al.: “OsiriX: An Open-Source Software for Navigating in Multidimensional DICOM Images,” Journal of digital Imaging, Sep. 2004, pp. 205-216.
Sage Intergy PACS | Product Summary. Enhancing Your Workflow by Delivering Web-based Diagnostic Images When and Where You Need Them, in 2 color pages. (IRV-SS-INTPACS-PSS-031309). © 2009 Sage Software Healcare, Inc. Downloaded from http://www.greenwayhealth.com/solutions/intergy/, Accessed on Feb. 9, 2015.
Sandberg, et al., “Automatic detection and notification of “wrong paitent-wrong location” errors in the operating room” Surgical Innovation, vol. 12, No. 3, Sep. 2005, pp. 253-260.
Schellingerhout, Dawid, MD, et al.: “Coregistration of Head CT Comparison Studies: Assessment of Clinical Utility,” Acad Radiol 2003; 10:242-248.
ScImage, Cardiology PACS, in 8 color page printout. Accessed at http://www.scimage.com/solutions/clinical-solutions/cardiology, Accessed on Feb. 9, 2015.
Sectra RIS PACS, in 2 color page printout. Accessed at https://www.sectra.com/medical/diagnostic_imaging/solutions/ris-pacs/, Accessed on Feb. 9, 2015.
Siemens syngo.plaza, Features and Benefits, in 2 color page printout. Accessed at http://www.healthcare.siemens.com/medical-imaging-it/imaging-it-radiology-image-management-pacs/syngoplaza/features, Accessed on Feb. 9, 2015.
Simms | RIS and PACS Medical Imaging Software, in 2 color page printout. http://www.mysimms.com/ris-pacs.php. Accessed on Feb. 9, 2015.
Sprawls, “Image Characteristics and Quality,” Physical Principles of Medical Imaging, http://www.sprawls.org/resources pp. 1-14.
Stryker, Imaging—OfficePACS Power Digital Imaging, in one color page printout. Accessed from http://www.stryker.com/emea/Solutions/Imaging/OfficePACSPowerDigitalImaging/index.htm. Accessed on Feb. 9, 2015.
Stryker, OfficePACS Power—Digital Imaging, 8 page color brochure, (MPP-022 Rev 4 BC/MP 300 Jan. 2007). © 2007 Stryker. Downloaded from http://www.stryker.com/emea/Solutions/Imaging/OfficePACSPowerDigitalImaging/ssLINK/emea/1557/022268. Accessed on Feb. 9, 2015.
Syed, et al., “Detection of Progressive Glaucomatous Optic Neuropathy Using Automated Alternation Flicker With Stereophotography,” Research Letter, Arch Ophthalmol., published online Dec. 13, 2010. 2010 American Medical Association.
Syed, et al.. “Automated alternation flicker for the detection of optic disc haemorrhages”, ACTA Ophthalmologica 2011, accepted for publication on Nov. 26, 2010.
Tay, et al., “Assessing Signal Intensity Change on Well-registered Images: Comparing Subtraction, Color-encoded Subtraction, and Parallel Display Formats”, Original Research:Computer Applications. Radiology, vol. 260: No. 2—Aug. 2011.
TeraRecon iNtuition pamphlet in 20 page, retrieved on Nov. 8, 2013, available at http://int.terarecon.com/wp-content/uploads/2013/11/brochure_english2013.pdf.
TeraRecon iNtuition—Workflow. <www.terarecon.com/wordpress/our-solutions/intuition-workflow> Last accessed Nov. 8, 2013. 2 pages.
UltraRAD—ultra Vision, 1 page (color), Downloaded from http://www.ultraradcorp.com/pdf/UltraVISION.pdf. Accessed on Feb. 9, 2015.
VanderBeek, et al., “Comparing the detection and agreement of parapapillary atrophy progression using digital optic disk photographs and alternation flicker”, Glaucoma, Graefes Arch Clin Exp Ophthalmol (2010) 248:1313-1317, Apr. 15, 2010.
VioStream for VitreaView, 2 color pages printout. Accessed at http://www.vitalimages.com/solutions/universal-viewing/viostream-for-vitreaview. Accessed on Feb. 9, 2015.
Visage Imaging Visage 7, 3 color page printout. Accessed at http://www.visageimaging.com/visage-7. Accessed on Feb. 9, 2015.
VIZTEK Radiology PACS Software Vixtek Opal-RAD, 4 color page printout. Accessed at http://viztek.net/products/opal-rad. Accessed on Feb. 9, 2015.
Voyager Imaging—Voyager PACS Radiologist Workstation, 2 page color brochure. Downloaded from http://www.intellirad.com.au/assets/Uploads/Voyager-PacsWorkstations.pdf?. Accessed on Feb. 9, 2015.
Voyager Imaging—Voyager PACS, 3 page color brochure. Downloaded from http://www.intellirad.com.au/index.php/assets/Uploads/Voyager-Pacs3.pdf. Accessed on Feb. 9, 2015.
Ivetic, D., and Dragan, D., Medical Image on the Go!, 2009, J Med Syst, vol. 35. pp. 499-516.
Tahmoush, D. and Samet, H., A New Database for Medical Images and Information, 2007, Medical Imaging 2007; PACS and Imaging Informatics, vol. 6516. pp. 1-9.
Office Action from the U.S. Patent and Trademark Office for U.S. Appl. No. 15/346,530 dated Mar. 26, 2018 (40 pages).
Office Action from the U.S. Patent and Trademark Office for U.S. Appl. No. 15/799,657 dated Mar. 8, 2018 (25 pages).
Notice of Allowance from the U.S. Patent and Trademark Office for U.S. Appl. No. 15/254,627 dated Jul. 13, 2017 (4 pages).
Notice of Allowance from the U.S. Patent and Trademark Office for U.S. Appl. No. 14/540,830 dated Aug. 15, 2017 (9 pages).
Non-Final Office Action from the U.S. Patent and Trademark Office for U.S. Appl. No. 15/346,530 dated Mar. 26, 2018 (12 pages).
Examiner Interview Summary from the U.S. Patent and Trademark Office for U.S. Appl. No. 11/179,384 dated Sep. 24, 2008 (4 pages).
Examiner Interview Summary from the U.S. Patent and Trademark Office for U.S. Appl. No. 11/179,384 dated Feb. 18, 2009 (2 pages).
Examiner Interview Summary from the U.S. Patent and Trademark Office for U.S. Appl. No. 12/870,645 dated Jun. 10, 2011 (2 pages).
Non-Final Office Action from the U.S. Patent and Trademark Office for U.S. Appl. No. 15/799,657 dated Mar. 8, 2018 (25 pages).
Non-Final Office Action from the U.S. Patent and Trademark Office for U.S. Appl. No. 11/944,000 dated Oct. 5, 2012 (11 pages).
Examiner Interview Summary from the U.S. Patent and Trademark Office for U.S. Appl. No. 11/944,000 dated Feb. 4, 2011 (3 pages).
Applicant Initiated Interview Summary from the U.S. Patent and Trademark Office for U.S. Appl. No. 14/179,328 dated Dec. 11, 2014 (3 pages).
Non-Final Office Action from the U.S. Patent and Trademark Office for U.S. Appl. No. 15/469,296 dated Jun. 27, 2017 (58 pages).
Non-Final Office Action from the U.S. Patent and Trademark Office for U.S. Appl. No. 15/469,281 dated Jun. 26, 2017 (51 pages).
Final Office Action from the U.S. Patent and Trademark Office for U.S. Appl. No. 15/469,281 dated Jan. 11, 2018 (60 pages).
Non-Final Office Action from the U.S. Patent and Trademark Office for U.S. Appl. No. 15/469,281 dated Apr. 2, 2018 (59 pages).
Notice of Allowance from the U.S. Patent and Trademark Office for U.S. Appl. No. 12/857,915 dated Jul. 3, 2014 (20 pages).
Notice of Allowability from the U.S. Patent and Trademark Office for U.S. Appl. No. 15/254,627 dated Jul. 13, 2017 (4 pages).
Notice of Allowance from the U.S. Patent and Trademark Office for U.S. Appl. No. 11/265,979 dated May 13, 2011 (14 pages).
Notice of Allowance from the U.S. Patent and Trademark Office for U.S. Appl. No. 13/171,081 dated Sep. 4, 2013 (12 pages).
Notice of Allowance from the U.S. Patent and Trademark Office for U.S. Appl. No. 12/870,645 dated Sep. 13, 2011 (8 pages).
Notice of Allowability from the U.S. Patent and Trademark Office for U.S. Appl. No. 12/870,645 dated Dec. 7, 2011 (4 pages).
Notice of Allowance from the U.S. Patent and Trademark Office for U.S. Appl. No. 12/891,543 dated Nov. 14, 2013 (14 pages).
Notice of Allowability from the U.S. Patent and Trademark Office for U.S. Appl. No. 15/163,600 dated Sep. 14, 2016 (6 pages).
Notice of Allowability from the U.S. Patent and Trademark Office for U.S. Appl. No. 13/572,397 dated Sep. 29, 2015 (2 pages).
Notice of Allowance from the U.S. Patent and Trademark Office for U.S. Appl. No. 15/469,296 dated Jan. 22, 2018 (11 pages).
Non-Final Office Action from the U.S. Patent and Trademark Office for U.S. Appl. No. 15/475,930 dated Jan. 10, 2018 (11 pages).
Non-Final Office Action from the U.S. Patent and Trademark Office for U.S. Appl. No. 15/469,342 dated Jun. 27, 2017 (62 pages).
Notice of Allowance from the U.S. Patent and Trademark Office for U.S. Appl. No. 15/469,342 dated Nov. 30, 2017 (12 pages).
Non-Final Office Action from the U.S. Patent and Trademark Office for U.S. Appl. No. 15/292,006 dated May 9, 2018 (17 pages).
Final Office Action from the U.S. Patent and Trademark Office for U.S. Appl. No. 15/475,930 dated Jun. 1, 2018 (17 pages).
Related Publications (1)
Number Date Country
20170301090 A1 Oct 2017 US
Provisional Applications (1)
Number Date Country
60625690 Nov 2004 US
Continuations (4)
Number Date Country
Parent 15254627 Sep 2016 US
Child 15631313 US
Parent 14502055 Sep 2014 US
Child 15254627 US
Parent 12857915 Aug 2010 US
Child 14502055 US
Parent 11268261 Nov 2005 US
Child 12857915 US