Method and apparatus for providing threat image projection (TIP) in a luggage screening system, and luggage screening system implementing same

Information

  • Patent Grant
  • 7899232
  • Patent Number
    7,899,232
  • Date Filed
    Friday, May 11, 2007
    17 years ago
  • Date Issued
    Tuesday, March 1, 2011
    13 years ago
Abstract
A method and apparatus for testing luggage screening equipment operators is provides. A sequence of images of contents of luggage items derived from a device that scans the luggage items with penetrating radiation are received. The image generation device is characterized by introducing a certain distortion into these images of contents of luggage items. A display device is caused to display images derived from this sequence of images. Occasionally, the display device is caused to show a simulated threat in a luggage item by displaying a combined image derived based at least in part on an image in the sequence of images and a distorted threat image. The distorted threat image was derived by applying a distortion insertion process to a reference threat image wherein the distortion insertion process tends to approximate the certain distortion introduced in the images of contents of luggage items by the device that scans the luggage items with penetrating radiation.
Description
FIELD OF THE INVENTION

The present invention relates generally to security systems and, more particularly, to methods and systems for screening luggage items having functionality for testing luggage screening equipment operators by displaying simulated threats in images conveying contents of luggage items.


BACKGROUND

Typically, security screening systems in airports make use of devices generating penetrating radiation, such as x-ray devices, to scan individual pieces of luggage to generate an image conveying the contents of the luggage. The image is displayed on a screen and is examined by a human operator whose task it is to detect and possibly identify, on the basis of the image, potentially threatening objects located in the luggage.


A deficiency with conventional security screening systems is that the job of luggage screening can become repetitious, leading to lapses in attention during which potentially dangerous items may be inadvertently allowed on board an aircraft for example.


In order to alleviate at least in part the above described deficiency, luggage screening systems are sometimes equipped with threat image projection (TIP) systems adapted for displaying an image of a piece of luggage with a simulated threat appearing thereon.


These threat image projection (TIP) systems allow merging two images: one image of a threat (e.g.: handgun, gun part, knife, improvised explosive device, etc.) based on stored images with one “live” image of a piece of luggage passing on the belt of the screening systems (such as an x-ray machine). The combined image is then displayed on the usual security screening monitor to the luggage screener who is then expected to process the luggage in accordance with the usual procedure when an actual threat is detected in the luggage. The functionality of such threat image projection systems is used during normal operation for benchmarking purposes, to keep luggage screeners alert, for maintaining the proficiency of experienced users and for training new users. An example of such a system is described in U.S. Pat. No. 6,899,540 issued to Neiderman, et al. on May 31, 2005 and entitled “Threat Image Projection System”. The contents of the above referenced document are incorporated herein by reference. In Neiderman et al., a computer system which causes a luggage screening monitor to show computer-generated (i.e. simulated) contraband items in a piece of luggage is described. This is done by either overlaying a computer-generated image of a contraband or other item onto the image of the luggage being screened, or by substituting the image of an entire piece of luggage which contains contraband for the image of the luggage being screened. The system also checks whether the contraband item (threat) fits into the current luggage being screened in order to ensure that a large threat is not projected into a small bag (e.g., a purse), thus cueing the screener that the threat is simulated. If the system determines that the selected contraband item (threat) does not fit in the luggage, the simulated contraband item is not displayed.


A deficiency in systems of the type described above is that the images shown to the user containing the computer-generated (i.e. simulated) contraband item (e.g. threat) do not convincingly convey that what the user is looking at is a real contraband item in the piece of luggage being screened. Frequently, the manner in which the images are merged together leads screeners to conclude that the threat displayed in the image of the piece of luggage was computer-generated as the result of the threat image projection (TIP) image-merging feature rather than a real threat.


Consequently, there is a need in the industry for providing a method and system for performing threat image projection (TIP) that alleviate at least in part the deficiencies of the prior art.


SUMMARY OF THE INVENTION

In accordance with a broad aspect, the present application seeks to provide a method for testing luggage screening equipment operators. The method comprises receiving a sequence of images of contents of luggage items derived from a device that scans the luggage items with penetrating radiation, the device being characterized by introducing a certain distortion into the images of contents of luggage items. The method also comprises causing a display device to display images derived from the sequence of images of contents of luggage items. The method also comprises occasionally causing the display device to show a simulated threat in a luggage item by displaying a combined image. The combined image is derived based at least in part on an image in the sequence of images and a distorted threat image, the distorted threat image having been derived by applying a distortion insertion process to a reference threat image to introduce a distortion in the reference threat image. The distortion insertion process tends to approximate the certain distortion introduced in the images of contents of luggage items by the device that scans the luggage items with penetrating radiation.


For the purpose of this specification, the expression “luggage item” is used to broadly describe luggage, suitcases, handbags, backpacks, briefcases, boxes, parcels or any other similar type of item suitable for containing objects therein.


Advantageously, the above-described method allows taking into account distortions introduced by the device that scans the luggage items with penetrating radiation when displaying a combined image conveying a representation of a luggage item with a simulated threat contained therein.


In accordance with a specific example of implementation, the method comprises occasionally causing the display device to show a same simulated threat positioned at different perceived heights in luggage items. This is effected by displaying combined images derived based at least in part on images in the sequence of images and a group of distorted threat images. The distorted threat images in the group are associated to respective heights and are derived by applying distinct distortion insertion processes to a reference threat image to introduce distortions in a reference threat image, wherein the distinct distortion insertion processes are each associated to respective heights.


In accordance with another broad aspect, the invention provides an apparatus for testing luggage screening equipment operators in accordance with the above-described method.


In accordance with another broad aspect, the invention provides a computer readable storage medium including a program element suitable for execution by a CPU for implementing an apparatus for testing luggage screening equipment operators in accordance with the above described method.


In accordance with another broad aspect, the present application seeks to provide a method for testing luggage screening equipment operators. The method comprises receiving a sequence of images of contents of luggage items derived from a device that scans the luggage items with penetrating radiation, the device being characterized by introducing a certain distortion into the images of contents of luggage items. The method also comprises causing a display device to display images derived from the sequence of images of contents of luggage items. The method also comprises occasionally causing the display device to show a simulated threat in a luggage item by:

    • applying a distortion insertion process to a reference threat image to derive a distorted threat image, the distortion insertion process tending to approximate the certain distortion introduced in the images of contents of luggage items by the device that scans the luggage items with penetrating radiation;
    • deriving a combined image conveying a simulated threat in the luggage item based at least in part on an image in the sequence of images and the distorted threat image; and
    • causing the display device to display the combined image.


Advantageously, by applying a distortion insertion process to a reference threat image associated to a given threat and then using this distorted threat image in the generation of a combined image for showing a simulation of that given threat, a more realistic representation of a luggage item that would have truly contained that given threat can be obtained.


In accordance with a specific example of implementation, the method comprises processing the image of contents of the luggage item to derive spatial information associated to the luggage item and conditioning the distortion insertion process at least in part based on the spatial information.


In accordance with a specific example of implementation, the method comprises providing a set of original threat images associated to respective potential threats and selecting at least one image from the set of original threat images as the reference threat image. The step of selecting may be effected using any suitable rule. In a specific example of implementation, selecting at least one image from the plurality of original threat images as the reference threat image is conditioned at least in part based on the spatial information associated to the luggage item. In a non-limiting example, wherein the spatial information associated to the luggage item includes size information, the method comprises selecting at least one image from the plurality of threat images as the original threat image based at least in part on the size information associated to the luggage item.


In accordance with a specific example of implementation, the method comprises providing an original threat image associated to a threat and assigning positioning information to the original threat image to derive the reference threat image. The positioning information assigned to the original threat image to derive the reference threat image is conditioned at least in part based on the spatial information associated to the luggage item.


In accordance with a specific example of implementation, the reference threat image includes intensity data for a set of coordinates, and the distorted threat image is defined by modified intensity data for a set of new coordinates. Applying a distortion insertion process comprises applying an image transformation to the intensity data for the set of coordinates to derive the modified intensity data for the new coordinates. The image transformation involves processing of a data structure representative of an inferred spatial transformation applied by the device that scans the luggage item with penetrating radiation. The inferred spatial transformation may be two-dimensional or three-dimensional. In a non-limiting example of implementation, the data structure is characterized by a set of parameters derived from registration of observed coordinates with respect to a set of reference coordinates.


In accordance with another specific example of implementation, the method comprises occasionally causing the display device to show a same simulated threat positioned at different perceived heights in luggage items. The method comprises applying distortion insertion processes to a reference threat image to derive a plurality of distorted threat images, the distortion insertion processes being associated to respective heights such that the distorted threat images in the plurality of distorted threat images are associated to respective perceived heights. The method comprises deriving combined images conveying the same simulated threat positioned at different perceived heights in luggage items based at least in part on images in the sequence of images and the plurality of distorted threat images and causing a display device to display at least some of the combined images.


In accordance with another broad aspect, the invention provides an apparatus for testing luggage screening equipment operators in accordance with the above-described method.


In accordance with another broad aspect, the invention provides a computer readable storage medium including a program element suitable for execution by a CPU, the program element implementing an apparatus for testing luggage screening equipment operators in accordance with the above described method.


In accordance with another aspect, the invention provides a system for screening luggage items. The system comprises an image generation device, a display device and an apparatus including functionality for testing luggage screening equipment operators in communication with the image generation device and the display device. The image generation device is for scanning luggage items with penetrating radiation to generate a sequence of images of contents of the luggage items. The image generation device is characterized by introducing a certain distortion into the images of contents of luggage items. The apparatus includes functionality for testing luggage screening equipment operators and comprises an input, and output and a processing unit. The input is for receiving the sequence of images of contents of luggage items. The output is for releasing an output signal suitable for causing the display device to display images derived from the sequence of images of contents of luggage items. The processing unit is operative for occasionally causing the output signal released at the output to convey a simulated threat in a luggage item by conveying a combined image. The combined image is derived based at least in part on an image in the sequence of images and a distorted threat image, wherein the distorted threat image was derived by applying a distortion insertion process to a reference threat image to introduce a distortion in the reference threat image. The distortion insertion process applied is such that it tends to approximate the certain distortion introduced in the images of contents of luggage items by the image generation device.


In accordance with another broad aspect, the invention provides a client-server system for implementing a process for testing luggage screening equipment operators. The client-server system comprising a client system and a server system operative to exchange messages there between over a data network. The server system stores a program element for execution by a CPU. The program element comprises a first program element component executed on the server system for receiving a sequence of images of contents of luggage items derived from a device that scans the luggage items with penetrating radiation. The device that scans the luggage items with penetrating radiation is characterized by introducing a certain distortion into the images of contents of luggage items. The program element also comprises a second program element component executed on the server system for generating messages for transmission to the client system for causing a display device in communication with the client system to display images derived from the sequence of images of contents of luggage items. The second program element component is operative for occasionally generating messages for causing the display device in communication with the client system to show a simulated threat in a luggage item by conveying a combined image. The combined image is derived based at least in part on an image in the sequence of images and a distorted threat image. The distorted threat image was derived by applying a distortion insertion process to a reference threat image to introduce a distortion in the reference threat image, wherein the distortion insertion process approximates the certain distortion introduced in the images of contents of luggage items by the device that scans the luggage items with penetrating radiation.


In accordance with another broad aspect, the invention provides an apparatus for testing luggage screening equipment operators. The apparatus comprises means for receiving a sequence of images of contents of luggage items derived from a device that scans the luggage items with penetrating radiation. The apparatus also comprises means for causing a display device to display images derived from the sequence of images of contents of luggage items and means for occasionally causing the display device to show a simulated threat in a luggage item by conveying a combined image. The combined image is derived based at least in part on an image in the sequence of images and a distorted threat image, the distorted threat image having been derived by applying a distortion insertion process to a reference threat image to introduce a distortion in the reference threat image. The distortion insertion process approximates the certain distortion introduced in the images of contents of luggage items by the device that scans the luggage items with penetrating radiation.


Other aspects and features of the present invention will become apparent to those ordinarily skilled in the art upon review of the following description of specific embodiments of the invention in conjunction with the accompanying Figures.





BRIEF DESCRIPTION OF THE DRAWINGS

A detailed description of embodiments of the present invention is provided herein below, by way of example only, with reference to the accompanying drawings, in which:



FIG. 1 is a high-level block diagram of a system for screening a luggage item in accordance with a specific example of implementation of the present invention;



FIG. 2 shows a process for testing luggage screening equipment operators in accordance with a specific example of implementation of the present invention;



FIG. 3 is a functional block diagram of a apparatus providing threat image projection (TIP) functionality suitable for use in the system shown in FIG. 1 in accordance with a specific example of implementation of the present invention;



FIG. 4
a shows a process implemented by the apparatus shows in FIG. 3 for testing luggage screening equipment operators in accordance with a specific example of implementation of the present invention;



FIG. 4
b shows a process for deriving a reference threat image in accordance with a specific example of implementation of the present invention;



FIG. 5 is a functional block diagram of a apparatus providing threat image projection (TIP) functionality suitable for use in the system shown in FIG. 1 in accordance with an alternative specific example of implementation of the present invention;



FIG. 6 diagrammatically illustrates the effect of applying distortion insertion processes to a original threat image to generate distorted threat images;



FIG. 7 diagrammatically illustrates a template for use in a registration process in order to model the distortion introduced by an image generation device in accordance with a specific example of implementation of the present invention;



FIG. 8 is a block diagram of an apparatus suitable for implementing a process for testing luggage screening equipment operators in accordance with a specific example of implementation of the present invention;



FIG. 9 shows a functional block diagram of a client-server system suitable for implementing a process for testing luggage screening equipment operators accordance with an alternative specific example of implementation of the present invention.





In the drawings, the embodiments of the invention are illustrated by way of examples. It is to be expressly understood that the description and drawings are only for the purpose of illustration and are an aid for understanding. They are not intended to be a definition of the limits of the invention.


DETAILED DESCRIPTION

Shown in FIG. 1 is a system 100 for screening a luggage item 104 providing threat image projection (TIP) functionality in accordance with a specific example of implementation of the present invention. The system 100 includes an image generation device 102, a display device 108 and an apparatus 106.


The image generation device 102 scans luggage items with penetrating radiation to generate a sequence of images of contents of the luggage items, the image generation device being characterized by introducing a certain distortion into the images of contents of luggage items. Specific examples of image generation devices that can be used include, without being limited to, x-ray, gamma ray, computed tomography (CT scans) and millimeter wave devices. Such devices are known in the art and as such will not be described further here. The images generated by the image generation device 102 may be two-dimensional (2D) or three-dimensional (3D) without detracting from the spirit of the invention. In a non-limiting example of implementation, the image generation device 102 is a conventional x-ray machine and the images in the sequence of images of the contents of the luggage items 104 are x-ray images of the luggage items 104 generated by the x-ray machine. In another non-limiting example of implementation, the image generation device 102 is a multi-view x-ray machine. The sequence of images of contents of luggage items 104 generated by the image generation device 102 may be in any suitable format such as, without limitation, VGA, SVGA, XGA, JPEG, GIF, TIFF and bitmap amongst others. Alternatively, the images may be in a proprietary image format. It will be appreciate that the actual format is not critical to the invention. Preferably, the sequence of images of contents of luggage items 104 is in a format that can be displayed on a display screen.


The image generation device 102 introduces a certain distortion into the images it generates, herein referred to as the distortion signature of the image generation device 102. More specifically, different objects in the luggage items 104 may be distorted to different degrees when captured in an image by the image generation device 102. The amplitude and nature of the distortion introduced by the image generation device 102 vary in space, both in the XY-plane (location of the luggage item 104 on the conveyor belt 120) and in z-height (position of an object in the luggage item 104 with regards to the z-axis which sets the distance between the object in question and the image generation device 102). As such, a given image generation device 102 will be characterized by introducing a certain distortion into the images of contents of luggage items it generates. In examples of implementation where the image generation device 102 is described as an x-ray machine the distortion introduced in the images is caused by the x-ray source's and detectors arrays' locations and the angle between them.


The apparatus 106 receives the sequence of images of contents of luggage items generated by the image generation device 102 and releases output signals for causing a display device, such as display device 108, to display images derived from the sequence of images of contents of luggage items. The apparatus 106 also includes functionality for testing luggage screening equipment operators by occasionally causing the output signal released to convey a simulated threat in a luggage item. The simulated threat in a luggage item is conveyed by a combined image derived based at least in part on an image in the sequence of images generated by the image generation device 102 and on a distorted threat image. The distorted threat image is derived by applying to a reference threat image a distortion insertion process that tends to approximate the certain distortion introduced by the image generation device 102. The manner in which the distorted image and the combined image can be generated will be described later on in the specification.


The display device 108 displays information to a luggage screening equipment operator in visual format. The display device 108 may be embodied as any suitable device including, without being limited to, a display screen of a computer terminal and a display screen of a portable display device. The display device 108 receives from the apparatus 106 signals suitable for causing the display device 108 to display images derived from the sequence of images of contents of luggage items generated by the image generation device 102. As will be described in detail later on in the specification, certain signals received from the apparatus 106 convey combined images for showing a simulated threat in a given luggage item.


A high level representation of a process for testing luggage screening equipment operators implemented by the system 100 is depicted in FIG. 2 of the drawings.


At step 200, an image of contents of a luggage item 104 is obtained by the image generation device 102 (shown in FIG. 1) and provided to the apparatus 106.


At step 201, a decision is made by the apparatus 106 as to whether a simulated threat should be added to the image received at step 200. The decision as to whether a threat should be added may be based on any suitable decision rule. For example, the decision rules may be based on some desired frequency of simulated presence of a threat, may be based on a suitable random function, may be periodic, and/or may be conditioned based on external factors. External factors may include for example, an identity associated to the luggage screening equipment operator making use of the system and a security alert level amongst others.


If step 201 is answered in the negative in that no simulated threat should be added to the image received at step 200, the system proceeds to step 202. At step 202, the apparatus 106 (shown in FIG. 1) releases an output signal for causing the display device 108 (shown in FIG. 1) to convey the image of contents of luggage item received at step 200 without the addition of a simulated threat. The system then returns to step 200 once again where a new image of contents of a luggage item is obtained from the image generation device 102.


If step 201 is answered in the affirmative in that a simulated threat should be added to the image received at step 200, the system proceeds to step 204. At step 204 the apparatus 106 releases an output signal for causing the display device 108 (both shown in FIG. 1) to show a simulated threat in a luggage item by displaying a combined image. The apparatus 106 derives the combined image based at least in part on the image received at step 200 and a distorted threat image. The distorted threat image is derived by applying a distortion insertion process to a reference threat image to introduce a distortion in that reference threat image, wherein the distortion insertion process tends to approximate the certain distortion introduced by the image generation device 102 (shown in FIG. 1). The distorted threat image used in the generation of the combined image may either be pre-computed and stored on a computer readable storage medium accessible by the apparatus 106 or, alternatively, may be derived by the apparatus 106 as part of the process for deriving the combined image. The system then returns to step 200 once again where a new image of contents of a luggage item is received obtained by the image generation device 102.


Advantageously, the system 100 allows a more realistic simulation of a threat in a luggage item to be presented to a luggage screening equipment operator by attempting to compensate for the distortion introduced by the image generation device 102 during the generation of the combined image.


An embodiment of apparatus 106 will now be described in greater detail with reference to FIG. 3.


Apparatus 106


As depicted, the apparatus 106 includes an input 302, an output 304 and a processing unit 330 in communication with the input 302 and the output 304. The first input 302 is for receiving images of contents of luggage items from the image generation device 102 (shown in FIG. 1). The output 304 is for releasing an output signal suitable for causing the display device 108 (shown in FIG. 1) to display images derived based on images of contents of luggage items received at input 302.


As shown, the processing unit 106 comprises a plurality of functional elements including an image display control module 306 and a distorted threat image generation module 308.


The distorted threat image generation module 308 is adapted for releasing to the image control module 306 a distorted threat image 326. In the embodiment shown in FIG. 3, the distorted threat image generation module 308 makes use of information conveyed by the image of contents of a luggage item received at input 302 to generate the distorted threat image 326. Optionally, the distorted threat image generation module 308 is adapted for releasing a threat image selection failure signal in cases where a suitable a distorted threat image could not be generated.


The image display control module 306 is adapted for causing the output signal released at output 304 to convey either one of the image of contents of a luggage item received at input 302 or a combined image conveying a simulated threat in the luggage item. In the latter case, the control module 306 derives the combined image at least in part based on the image received at input 302 and a distorted threat image received from the distorted threat image generation module 308. The image display control module 306 is responsive to receipt of a threat image selection failure signal for causing the output signal released at output 304 to convey the image of contents of a luggage item received at input 302 without the addition of a simulated threat.


In a specific example of implementation, the display control module 306 is adapted for selectively issuing a control signal 342 to the distorted threat image generation module 308 for causing the latter to provide a distorted threat image 326. In such an implementation, the distorted threat image generation module 308 is responsive to the control signal 342 from the image display control module 306 for releasing a distorted threat image 326 to the image display control module 306. Alternatively, the distorted threat image generation module 308 is adapted for automatically generating a distorted threat image 326 upon receipt of a new image of contents of a luggage item received at input 302. In such an alternative, the control signal 342 may be omitted.


Any suitable method for generating a combined image based on the image received at input 302 and the distorted threat image 326 may be used such that the resulting combined image conveying the presence of a simulated threat in the luggage item. In a non-limiting example, the combined image is generated by overlaying the distorted threat image 326 onto the image received at input 302.


In an alternative example of implementation, where the image received at input 302 and the distorted threat image are x-ray images, the display control module 306 is adapted to derive the combined image by performing a combination of the attenuation information conveyed in the image received at input 302 and the attenuation information conveyed in the distorted threat image 326. In this specific example, the resulting combined image is derived based on the true energy levels in the image received at input 302 and in the distorted threat image (high-energy level E1 and low-energy level E2). Advantageously, this alternative method allows generating a combined image that takes into account compound material effects. In other words, the combined image when displayed on display device 108 (shown in FIG. 1) to the luggage screening equipment operators will show that the simulated threat is impacted by the presence of other objects that are located above or underneath it in the luggage item.


It will be appreciated, in light of the above description, that other suitable methods for deriving the combined image at least on part based in the image received at input 302 and the distorted threat image 326 may be used without detracting from the spirit of the invention.


In the specific example of implementation depicted in FIG. 3, the distorted threat image generation module 308 includes a spatial coordinate estimator 310, a threat selection unit 312, a threat position selection unit 316, a threat positioning unit 320, a distortion function generator 318 and a distortion processor 322. The distorted threat image generation module 308 may also include a database of threats 340 and a distortion database 314.


Database of Threats 340


The database of threats 340 includes a memory unit for storing a set of original threat images associated to respective potential threats whose presence in luggage items the system 100 (shown in FIG. 1) is designed to simulate. Examples of potential threats to which original threat images in the database of threats may be associated include, without being limited to, weapons (e.g. guns, knives and ammunition) and forbidden objects (e.g. containers of liquid, aerosol cans, lighters, detonators).


It will be appreciated certain embodiments of the database of threats 340 may include a single original threat image, which would allow the system 100 (shown in FIG. 1) to simulate a single potential threat in luggage items. However, practical implementations of the system 100 (shown in FIG. 1) will most likely find it beneficial to include in the database of threats 340 original threat images associated to multiple potential threats in order to provide a system capable of simulating different types of potential threats in luggage items.


In a specific example of implementation, the images in the set of original threat images in the database 340 may be derived by an image generation device using a similar type of penetrating radiation as the image generation device 102 depicted in FIG. 1. For example, where the image generation device 102 depicted in FIG. 1 is an x-ray machine, the images in the database 340 are x-ray images of the respective potential threat to which they are associated and are preferably generated by a device similar to image generation device 102.


It will be appreciated that the image generation device used for the purpose of generating the set of original threat images in the database 340 need not be the same device as the image generation device 102 used in the system 100 and may even be a device manufactured by a different entity. For example, the image generation device used for the purpose of generating the set of original threat images in the database 340 will most likely be characterized by introducing a certain distortion into the original threat images it generated. For the purpose of simplicity, let us refer to this certain distortion as a first distortion signature. The image generation device 102 used in the system 100 is also characterized by introducing a certain distortion into the images of contents of luggage items it generates. For the purpose of simplicity, let us refer to this other certain distortion as a second distortion signature. The first distortion signature and the second distortion signature may be substantially the same or may be different.


In a specific example of implementation, the original threat images in the database of threats 340 have been pre-processed to remove therefrom at least in part the distortion introduced by the image generation device used for the purpose of generating the set of original threat images in the database 340. As such, the original threat images in the database 340 can be considered to be “normalized” images that have been rendered substantially independent from the distortion signature characterizing the image generation device that generated them. Any suitable method for correction distortion in a scanned image may be used. An example of such a method suitable for correcting distortion introduced by an image generation device is described in co-pending U.S. patent application Ser. No. 11/431,627 entitled “APPARATUS, METHOD AND SYSTEM FOR SCREENING RECEPTACLES AND PERSONS, HAVING IMAGE DISTORTION CORRECTION FUNCTIONALITY”, filed on May 11, 2006 by L. Perron et al. and published under publication no. 20070041612A1 on Feb. 22, 2007. The contents of the above referenced document are incorporated herein by reference. Advantageously, by storing “normalized” images in database 340, the latter may be used in connection with systems including image generation devices characterized by distinct distortion signatures without requiring the systems to be adapted to the distortion signature of the image generation device that generated the database 340. In addition, by making the set of original threat images in the database 340 substantially independent from the distortion signature of the image generation device that generated them, the update of the database 340 may be made easier. For instance, new threat images to be added to the database 340 need not be generated by an image generation device characterized by the same the distortion signature as the image generation device that generated the threat images already in the database.


For the purpose of the present description, we will describe an implementation where the threat images in the database of threats 340 are “normalized” image, i.e. substantially rendered independent from the distortion signature of the image generation device that generated the image. It will become apparent to the person skilled in the art in light of the present description how the distortion transformation database 314 can be modified in implementations where the images in the database 240 are not “normalized” images.


In addition, although the database 340 has been shown in FIG. 3 to be a component of apparatus 106, it will be appreciated that in certain embodiments the database 340 may be a component in communication with but separate from apparatus 106 and that such implementations do not detract from the spirit of the invention. In addition, it will also be appreciated that in certain implementations where the database 340 is a component separate from apparatus 106, the database 340 may be shared between multiple apparatuses 106 without detracting from the spirit of the invention.


Distortion Transformation Database 314


One recalls that the image generation device 102 (shown in FIG. 1) introduces distortion into the images of contents of the receptacle 104.


The distortion transformation database 314 stores information conveying characteristics of the distortion introduced by the image generation device 102 into the images of contents of luggage items. For the purpose of this description, we will refer to the distortion introduced by the image generation device 102 into the images of contents of luggage items as the distortion signature of the image generation device 102. Typically, different image generation device 102 will have different distortion signatures. In other words, different image generation devices will typically introduce different amounts and types of distortions into the images they generate. As such, the distortion transformation database 314 stores information associated to image generation devices having a certain distortion signature, such as for example a certain specific manufacturer model (or make) of image generation device. Where the apparatus 106 is adapted to be used with image generation devices having different distortion signatures, respective distortion transformation databases are provided for each distortion signature.


In a specific example of implementation, the distortion transformation database 314 stores information characterizing a distortion insertion process that tends to approximate the certain distortion introduced in the images of contents of luggage items by the image generation device 102. When this distortion insertion process is applied to an image of an object, the distortion insertion process mimics the effect of the image generation device 102 by introducing therein a distortion based on the distortion signature of the image generation device 102.


In a specific practical example of implementation, the distortion transformation database 314 includes a memory unit storing a transformation data structure denoted as T*, where T* is an inferred spatial transformation. In a specific implementation, T* includes a plurality of entries, denoted TH*, wherein each TH* is an inferred spatial transformation associated to a respective height (z-height) H. It will therefore be appreciated that, where T* includes TH* for different values of H (i.e. different heights), the inferred spatial transformation T* is considered a three-dimensional spatial transformation comprised of multiple inferred two-dimensional spatial transformations TH*. In cases where an inferred spatial transformation is provided for only one height, T* is a two-dimensional spatial transformation.


An example of the manner in which the transformation data structure denoted T* can be derived will now be described.


One could model the distortion introduced by the image generation device 102 as a spatial transformation T on a “true” (undistorted) image to arrive at a corresponding image generated by the image generation device 102. Thus, T represents a spatial transformation that models the distortion affecting a certain object having a given shape and location in a “true” (undistorted) image, resulting in that certain object's “distorted” shape and location in an image of that object that would be generated by the image generation device 102. Thus, to derive a simulation of an image that would have been generated by the image generation device 102 based in a “true” (undistorted) image, it is reasonable to want to make the distortion insertion process to be applied to that “true” (undistorted) image to resemble T as closely as possible.


It will be appreciated that the amplitude and nature of the distortion introduced by the image generation device 102 vary in space, both in the XY-plane (location of the luggage item 104 on the conveyor belt 120) and in z-height (position of an object in the luggage item 104 with regards to the z-axis which sets the distance between the object in question and the image generation device 102). As such it will be appreciated that characteristics of the spatial transformation T will vary not only according the position of the object in the XY-plane but will also vary according to different heights (z-heights).


Thus, one can denote the spatial transformation for a given candidate height H by TH, which therefore models the distortion affecting objects located at the candidate height H when scanned by the image generation device 102.


Now, although the exact TH is not known, it may be inferred. The inferred version of TH is denoted TH* and is hereinafter referred to as an “inferred spatial transformation” for a given candidate height H. Basically, TH* can be defined as a data structure that represents an estimate of TH. Although the number of possible heights that an object may occupy is a continuous variable, it may be possible to granularize this number to a limited set of “candidate heights” (e.g., such as 5-10) without introducing a noticeable error. Moreover, since in a Threat Image Projection (TIP) context it is possible to control the selection of the height at which a threat image will be inserted, that selection may be made to correspond to a “candidate height” for which a spatial transformation TH* has been determined. Of course, the number of candidate heights in a given embodiment may be as low as one, while the upper bound on the number of candidate heights is not particularly limited.


The data structure that represents the inferred spatial transformation TH* for a given candidate height H may be characterized by a set of parameters. The set of parameters may be derived from registration of observed coordinates with respect to a set of reference coordinates. More specifically, the set of parameters can be derived from the coordinates of a set of “control points” in both an image generated by the image generation device 102 and an “true” image for that candidate height. A “control point” corresponds to a picture element that occurs at a known location in the “true” image (defining the reference coordinates) for a given candidate height H, and whose “distorted” position can be located in the image generated by the image generation device 102 (defining the observed coordinates).


In one specific non-limiting embodiment, to obtain control points specific to a given image generation device 102, and with reference to FIG. 7, one can use a template 700 having a set of spaced apart holes 710 at known locations in the horizontal and vertical directions. The template is placed at a given candidate height H1420. One then acquires an input image 714, from which control points 712 (i.e., the holes 710 present at known locations in the template 700) are identified in the input image 714. This may also be referred to as “a registration process”. Having performed the registration process on the input image 714 that was derived from the template 700, one obtains TH1420*, the inferred spatial transformation for the height H1420.


To obtain the inferred spatial transformation TH* for a given candidate height H, one may utilize a “transformation model”. The transformation model that is used may fall into one or more of the following non-limiting categories, depending on the type of distortion that is sought to be corrected:

    • linear conformal;
    • affine;
    • projective
    • polynomial warping (first order, second order, etc.);
    • piecewise linear;
    • local weighted mean;
    • etc.


The use of the function cp2tform in the Image Processing Toolbox of Matlab® (available from Mathworks Inc.) is particularly suitable for the computation of inferred spatial transformations such as TH* based on coordinates for a set of control points. Other techniques will now be apparent to persons skilled in the art to which the present invention pertains. The above process can be repeated several times, for different candidate heights, thus obtaining TH* for various candidate heights.


The TH* for various candidate heights derived in the manner described above are combined to form transformation data structure T* which is then stored on a memory unit as part of distortion transformation database 314.


To better illustrate the manner in which transformation data structure T* can be used to derive a simulation of an image that would have been generated by the image generation device 102 based in a “true” (undistorted) image, an example will be described with reference to FIG. 6.


Block 602 is a graphical representation of an original threat image. The original threat image may be derived for example from a database of threat images.


Blocks 604a and 604b are graphical representations of the original threat image repositioned in a three-dimensional space (X, Y, Z-height) at locations at which a threat image is to be inserted in the combined images. The repositioned original threat images 604a and 604b are the reference images to which a distortion insertion process will be applied. Block 604a depicts a first reference image positioned at Z-height H890 while block 604b depicts a second reference image positioned at Z-height H470.


Based on the Z-height information, H890 and H470, blocks 606a and 606b select from the distortion transformation database 314 the inferred spatial transformations TH* corresponding to heights H890 and H470. Assuming that TH* for H890 and H470 can be retrieved from distortion transformation database 314, one applies the image transformation TH* to the reference threat images 604a and 604b in order to obtain the distorted threat images 606a and 606b. Distorted threat images 606a is associated to z-height H890 and distorted threat images 606b is associated to z-height H470. Hence an image of a simulated threat can be positioned at different heights in luggage items. Different distortion insertion processes associated to respective heights are applied to a reference threat image to derive a plurality of distorted threat images.


The use of the function imtransform in the Image Processing Toolbox of Matlab® (available from Mathworks Inc.) is particularly suitable for the computation of an output image (such as the distorted images 606a and 606b) based on an input image (such as the reference threat images 604a and 604b) and an inferred spatial transformation such as TH*. Other techniques will now be apparent to persons skilled in the art to which the present invention pertains.


As was the case for database 340, although the database 314 has been shown in FIG. 3 to be a component of apparatus 106, it will be appreciated that in certain embodiments the database 314 may be a component in communication with but separate from apparatus 106 and that such implementations do not detract from the spirit of the invention. In addition, it will also be appreciated that in certain implementations where the database 314 is a component separate from apparatus 106, the database 314 may be shared between multiple apparatuses 106 without detracting from the spirit of the invention.


Spatial Coordinate Estimator 310


The spatial coordinate estimator 310 is adapted for processing the image of contents of a luggage item received from input 302 to derive spatial information associated to the luggage item. The spatial information generated provides information related to the boundaries of the luggage item. Such information related to the boundaries of the luggage item may include, without being limited to, position information and size information associated to the luggage item. In a specific example of implementation, the position information includes boundaries of a luggage item in the XY-plane (location of the luggage item 104 on the conveyor belt 120) and in z-height (height of the boundaries of the luggage item 104 with regards to the z-axis which sets the distance between the boundaries of the luggage item 104 and the image generation device 102). Any suitable image processing method for deriving the boundaries of the luggage items in an image may be used. It will be readily appreciated that where the image of contents of a luggage item received at input 302 includes a representation of multiple luggage items, spatial information associated to each of the respective luggage items in the image may be generated by the spatial coordinate estimator 310. For the purpose of simplicity, we will describe the case where the image of contents of a luggage item received at input 302 include a representation of contents of a single luggage item.


The spatial coordinate estimator 310 releases the spatial information conveying boundary information associated to the luggage item in the image received from input 302.


The Threat Selection Unit 312


The threat selection unit 312 selects one or more original threat images from the set of original threat images in the database of threats 340 and releases the selected one or more original threat images as signal 328.


The threat selection unit 312 may make use of any suitable set of threat image selection rules to effect the selection of a given threat image from the database of threats 340. Such threat image selection rules may for example, cycle through the original threat images in the database of threats 340 so that the various original threat image are presented to the user in some pre-determined order. Alternatively, the threat image selection rules may provide that original threat images in the database of threats 340 are selected according to respective frequencies such that a given original threat image is selected more (or less) frequently than another original threat image. Alternatively still, the threat image selection rules may provide for a pseudo random selection of an original threat image from the set of original threat images in the database of threats 340. Advantageously, by providing for a pseudo random selection, the selected pattern of threat images varies over time and should not easily be discernable by a luggage screening equipment operator using the system.


Preferably, the threat selection unit 312 conditions the selection of the original threat image at least in part based on the spatial information associated to the luggage item received from the spatial coordinate estimator 310. As described previously, the spatial information associated to the luggage item received from the spatial coordinate estimator 310 provides information related to the boundaries of the luggage item depicted in the image received at input 302.


In a specific example of implementation, the threat selection unit 312 effects an initial selection of an original threat image from the database of threats 340. The threat object selection unit 312 then verifies whether the selected original threat image can be made to fit within the luggage item depicted in the image received at input 302 based on the spatial information received from the spatial coordinate estimator 310. Any suitable image processing method for verifying whether an object fits within a defined boundary may be used to effect this verification. Such methods are known in the art of image processing and as such will not be described further here.


If the threat object selection unit 312 determines that the selected original threat image can be made to fit within the luggage item, the threat object selection unit 312 releases the selected original threat image 328 to the threat position selection unit 316.


If the threat object selection unit 312 determines that the selected original threat image cannot be made to fit within the luggage item, the threat object selection unit 312 selects another original threat image from the database 340 and repeats the verification of whether the newly selected original threat image can be made to fit within the luggage item. The threat object selection unit 312 repeats the process of selecting an original threat image from the database 340 and verifying whether the selected object can be made to fit within the luggage item until:

    • either a selected original threat image that can be made to fit within the luggage item is found; or
    • a termination condition is reached.


The termination condition may be reached for any number of suitable reasons. As an example, a termination condition may be reached when a maximum number of selections of original threat images has been reached and has failed to result in a selected original threat image that will fit into the luggage item. The maximum number of selections of original threat images should be selected such as not to delay the display of an image conveying contents of luggage items on the display device 108 by an unreasonable amount of time. If a termination condition is reached, the threat selection unit 312 issues a threat image selection failure signal conveying that the selection process has failed to result in the selection of an original threat image. The threat image selection failure signal is conveyed to the image display control module 306. In response to receipt of a threat image selection failure signal, the image display control module 306 causes the output signal released at output 304 to convey the image of contents of a luggage item received at input 302 without the addition of a simulated threat.


Optionally, the threat selection unit 312 uses size information conveyed by the spatial information associated to the luggage item and received from the spatial coordinate estimator to effect a pre-selection of objects from the database of threats 340. In a first example, prior to initiating the process described above, the threat selection unit 312 uses size information conveyed by the spatial information associated to the luggage to effect a pre-selection of objects from the database of threats 340. In this manner, threat objects in the database of threats 340 that have a size (such as a length, width, height and/or volume) that is larger than that of the luggage item, may be eliminated by the pre-selection step. The remaining threat objects in the database of threats 340 can then be processed to select one or more original threat images on the basis of selection rules of the type described above. In a second example, prior to initiating the selection process described above, the threat selection unit 312 uses size information conveyed by the spatial information associated to the luggage to determine whether the luggage item has a size that is at least as large as a pre-determined minimum size. If the luggage item has a size that is smaller that the pre-determined minimum size, the threat selection unit issues a threat image selection failure signal conveying that the selection process has failed to result in the selection of an original threat image. If the threat selection unit 312 determines that the luggage item has a size that is at least as large as the pre-determined minimum size, the threat selection unit 312 proceeds to select one or more original threat images on the basis of selection rules of the type described above.


It will be readily apparent to the person skilled in the art that the above described embodiment have been described for the purpose of illustration and that other suitable manners of effecting a selection of a threat image from a database of threats may be envisaged without detracting from the spirit of the invention.


The Threat Position Selection Unit 316


The threat position selection unit 316 is adapted to derive positioning information to be assigned to the selected original threat image 328 received from the threat selection unit 316. This positioning information allows a same threat image stored in the database of threats 340 to be positioned in a plurality of different locations in the images conveyed by the output signal released at out put 304. The derived position information is released to the threat positioning unit 320 and to the distortion function generator 318.


In a specific example of implementation, the position information includes positioning data in the XY-plane and in z-height.


The threat position selection unit 316 may make use of any suitable set of threat position selection rules to effect the selection of a given position for the selected threat image 328 received from the threat selection unit 312.


Such threat position selection rules may for example, cycle through a set of pre-determined positions so that the threat images are presented to the user in some pre-determined set of positions. Alternatively, the threat position selection rules may provide that positions are selected according to respective frequencies such that threat images are placed in certain positions more (or less) frequently than in other positions. Alternatively still, the threat position selection rules may provide for a pseudo random selection of position. Advantageously, by providing for a pseudo random selection of the original threat image, the selected pattern of positions for the threat images varies over time and should not easily be discernable by a luggage screening equipment operator using the system.


Most preferably, the threat position selection rules are such that they make use of the spatial information conveying boundary information associated to the luggage item in the image received from input 302 and released by the spatial coordinate estimator 310. Any suitable image processing method for positioning an object within a defined boundary may be used to effect the positioning selection. Such methods are known in the art of image processing and as such will not be described further here. By providing spatial information conveying boundary information associated to the luggage item in the XY-plane and in z-height and by generating positioning information for the threat image in the XY-plane and in z-height, the selected threat image can be positioned such as convey a simulated threat located within the luggage item. This is particularly advantageous where the image of contents of a luggage item received at input 302 and the threat image selected from the database 340 are three-dimensional (3D) images.


The Threat Positioning Unit 320


The threat positioning unit 320 receives the positioning information from the threat position selection unit 316 and assigns the positioning information to the original threat image to derive the reference threat image 324. The reference threat image 324 is released to the distortion processor 322.


The assigning of the positioning information to the original threat image includes modifying the original threat image according to the positioning information so that the image of the threat appears at a different location in the reference image. Said otherwise, the reference threat image 324 is a repositioned version of the original threat image 328. Any suitable image processing method for positioning an image based on desired positioning information may be used. Such methods are known in the art of image processing and as such will not be described further here.


In a non-limiting specific example of implementation, the assigning of the positioning information to the original threat image includes modifying the original threat image according to the XY-plane positioning information.


In another alternative embodiment, not shown in the figures, the database of threats 340 stores sets of original threat images, where each original threat image in a given set is associated to the same threat in a respective position in the XY-plane. In such an alternative embodiment, the threat positioning unit 320 may be omitted and threat position selection unit 316 may be limited to selecting a z-height position. In such a case, the threat selection unit 312 is modified to release a reference threat image 324 to the distortion processor 322 and the threat position selection unit 316 may be limited to selecting a z-height position and to provide the latter to the distortion function generator 318.


The Distortion Function Generator 318


The distortion function generator 318 is adapted for accessing the distortion transformation database 314 to select therefrom a distortion insertion process to be applied to the reference threat image 324. The selected distortion insertion process is then released and provided to the distortion processor 322.


As described above, the amplitude and nature of the distortion introduced by the image generation device 102 vary in space, both in the XY-plane and in z-height. The distortion function generator 318 receives the position information derived by the threat position selection unit 316 and conditions the distortion insertion process based at least in part on the positioning information.


In the specific example of implementation described, the distortion transformation database 314 includes a data structures storing a transformation data structure denoted as T*, where T* includes a plurality of entries, denoted TH*, wherein each TH* is an inferred spatial transformation associated to a respective height (z-height) H. The distortion function generator 318 is adapted for selecting from the distortion transformation database 314 a certain inferred spatial transformation TH′* corresponding to z-height H′. Z-height H′ is the height for which a spatial transformation is present in the database 314 that is in proximity to the height conveyed by the positioning information derived by the threat position selection unit 316.


The Distortion Processor 322


The distortion processor 322 receives the reference threat image 324 from the threat positioning unit 320 and the selected distortion insertion process from the distortion function generator 318. The distortion processor 322 processes the reference threat image 324 by applying the selected distortion insertion process to introduce in reference threat image 324 a distortion tending to approximate the certain distortion introduced in the images of contents of luggage items by the image generation device 102.


In a specific example of implementation, the certain spatial transformation TH′* received from the distortion function generator 318 is applied to the reference threat image 324 received from the threat positioning unit 320 to derive a distorted threat image 326. The distorted threat image 326 is then released and provided to the image display control module 306.


In a specific example of implementation, the reference threat image 324 includes intensity data for a set of coordinates. Applying the certain spatial transformation TH′* to the intensity data for the set of coordinates in the reference threat image 324 results in modified intensity data for a set of new coordinates. The modified intensity data is released as part of the distorted threat image 326.


Description of Typical Interaction


A specific example of a process implemented by the apparatus 106 shown in FIG. 3 for testing luggage screening equipment operators will now be described with reference to FIGS. 4a and 4b. For the purpose of increasing the clarity of the description, FIGS. 4a and 4b show which components of apparatus 106 perform the different steps.


At step 400, an image of contents of a luggage item is obtained by the image generation device 104 (shown in FIG. 1) and is provided to the apparatus 106.


At step 402, the image display control module 306 of apparatus 106 makes a decision as to whether a simulated threat should be added to the image received at step 400. The decision as to whether a threat should be added may be based on any suitable decision rule.


If step 402 is answered in the negative in that no simulated threat should be added to the image received at step 400, the image display control module 306 proceeds to step 412. At step 412, the image display control module 306 releases an output signal for causing the display device 108 (shown in FIG. 1) to convey the image of contents of luggage item received at step 400 without the addition of a simulated threat. The system then returns to step 400 once again where a new image of contents of a luggage item is obtained from the image generation device 104 (shown in FIG. 1).


If step 402 is answered in the affirmative in that a simulated threat should be added to the image received at step 400, the image display control module 306 issues a control signal 342 (shown in FIG. 3) to the distorted threat image generation module 308 for causing the latter to initiate step 404.


At step 404, the distorted threat image generation module 308 derives a reference threat image conveying a threat whose presence it to be simulated in the image received at step 400. If step 404 fails and a reference threat image cannot be derived, the distorted threat image generation module 308 issues a threat image selection failure signal to the image display control module 306 for causing the latter to initiate step 412. If a reference threat image is successfully derived, the distorted threat image generation module 308 proceeds to step 406.


In a specific example of implementation, step 404 can be subdivided into steps 450452 and 454 depicted in FIG. 4b.


More specifically, at step 450, an original threat image is selected from the database of threats 340 (shown in FIG. 3). If step 450 fails, due for example to the inability to select a threat image that will fit within the boundaries of a luggage item, a threat image selection failure signal is generated. The threat image selection failure signal is then transmitted to the image display control module 306 for causing the latter to initiate step 412 (shown in FIG. 4a). If step 450 is successful, it will result in a selected original threat image, which is then provided to step 452.


At step 452, the position in the image received at step 400 where the threat image selected at step 450 is to be inserted is selected. Completion of step 452 results in positioning information, which is then forwarded to step 454.


At step 454, the positioning information is applied to the threat image selected at step 450 to derive a reference threat image. Completion of step 454 results in a reference threat image that corresponds to a repositioned version of the threat image selected at step 450. The reference threat images is then forwarded to step 406 (FIG. 4a).


Returning to FIG. 4a, at step 406, a distorted threat image is derived by applying a distortion insertion process to the reference threat image derived at step 404 to introduce therein a distortion. As described previously, the distortion insertion process tends to approximate the certain distortion introduced by the image generation device 102 (shown in FIG. 1). The distorted threat image is then released by the distorted threat image generation module 308 to the image display control module 306, which is caused to initiate step 408.


At step 408, image display control module 306 derives a combined image based at least in part on the image received at step 400 and the distorted threat image derived at step 406 by the distorted threat image generation module 308.


At step 410, image display control module 306 releases an output signal for causing the display device 108 (shown in FIG. 1) to show a simulated threat in a luggage item by displaying the combined image derived at step 408. The image display control module 306 then returns to step 400 once again where a new image of contents of a luggage item is received.


Advantageously, the above described process allows a more realistic simulation of a threat in a luggage item to be presented to a luggage screening equipment operator by attempting to mimic the distortion introduced by the image generation device 102 in the simulated threat image.


Variant—Pre-Computed Distorted Threat Images


It will be appreciated that, although the above described embodiment of the apparatus 106 has been described as providing distorted threat images by applying a distortion insertion process during the operation of the system, distorted threat images for different (X, Y, Z) locations for a same original threat image may be pre-computed and stored on a computer readable medium.


An example of such a variant is depicted in FIG. 5 of the drawings as apparatus 106′.


As depicted, the apparatus 106′ includes an input 514 and an output 516 analogous to input 302 and output 304 described with reference to FIG. 3. The apparatus 106′ also includes a processing unit 510 in communication with the input 514 and the output 516 and releases an output signal for transmission to the display device 108.


As shown, the processing unit 510 comprises a plurality of functional elements including an image display control module 512, analogous to image display control module 306 described with reference to FIG. 3, and a distorted threat image generation module 500.


The distorted threat image generation module 500 is adapted for releasing to the image control module 512 a distorted threat image 518 analogous to distorted threat image 326 described with reference to FIG. 3. In the embodiment shown in FIG. 5, the distorted threat image generation module 500 makes use of information conveyed by the image of contents of a luggage item received at input 514 to generate the distorted threat image 518. As a variant, the distorted threat image generation module 500 is adapted for releasing a threat image selection failure signal (not shown) in cases where a suitable a distorted threat image could not be generated.


In the variant depicted in FIG. 5, the distorted threat image generation module 500 includes a spatial coordinate estimator 520, a threat selection unit 504 and a database of distorted threat images 502.


The spatial coordinate estimator 520 is analogous to spatial coordinate estimator 310 described with reference to FIG. 3.


The database of distorted threat images 502 includes a memory unit for storing a set of distorted threat images associated to potential threats whose presence in luggage items the system 100 (shown in FIG. 1) is designed to simulate. The distorted threat images in the database 502 were derived by applying a distortion insertion process to reference threat images to introduce therein distortions tending to approximate the distortions introduced by the image generation device 102 (shown in FIG. 1). In a specific example of implementation, the set of distorted threat images in the database 502 includes a plurality of subsets, each subset being associated to a respective potential threat. The distorted threat images in a given subset correspond to different positions at which a potential threat is to be simulated. The distorted threat images in the database 502 may be generated for example by the methods described previously in the present specification.


It will be appreciated certain embodiments of the database of threats 502 may include a single subset of distorted threat images associated to a single potential threat, which would allow the system 100 (shown in FIG. 1) to simulate a single potential threat in luggage items. However, practical implementations of the system 100 (shown in FIG. 1) will most likely find it beneficial to include in the database of threats 502 distorted threat images associated to multiple potential threats in order to provide a system capable of simulating different potential threats in luggage items.


The threat selection unit 504 includes functionality similar to that of threat selection unit 312 except that threat selection unit 504 performs the selection process on the database of distorted threat images 502. The threat selection unit 504 selects one or more distorted threat images from the database of distorted threats 502 and releases the selected one or more distorted threat images as signal 518.


Advantageously, by using pre-computed distorted threat images rather than computing the distorted threat images during the operation of the system 100, a reduction in delay and in computational requirements of for the apparatus 106 can be achieved.


Specific Practical Implementation


Those skilled in the art will appreciate that certain portions of the apparatus 106 (shown in FIG. 1) can be implemented on a general purpose digital computer 800, of the type depicted in FIG. 8, including a processing unit 802 and a memory 804 connected by a communication bus. The memory includes data 808 and program instructions 806. The processing unit 802 is adapted to process the data 808 and the program instructions 806 in order to implement the functional blocks described in the specification and depicted in the drawings. The digital computer 800 may also comprise an I/O interface 810 for receiving or sending data elements to external devices.


Alternatively, the above-described apparatus 106 can be implemented on a dedicated hardware platform where electrical components implement the functional blocks described in the specification and depicted in the drawings. Specific implementations may be realized using ICs, ASICs, DSPs, FPGAs or other suitable hardware platform.


It will be appreciated that the screening system 100 (depicted in FIG. 1) may also be of a distributed nature where the images of contents of luggage items are obtained at one location or more locations and transmitted over a network to a server unit implementing the method described above. The server unit may then transmit a signal for causing a display device to display information to the user. The display device may be located in the same location where the images of contents of luggage items were obtained or in the same location as the server unit or in yet another location. In specific examples of implementations, the display device may be a hand-held portable display device or may be part of a user console.



FIG. 9 illustrates a network-based client-server system 900 for system for screening receptacles. The client-server system 900 includes a plurality of client systems 902, 904, 906 and 908 connected to a server system 910 through network 912. The communication links 914 between the client systems 902, 904, 906 and 908 and the server system 910 can be metallic conductors, optical fibres or wireless, without departing from the spirit of the invention. The network 912 may be any suitable network including but not limited to a global public network such as the Internet, a private network and a wireless network. The server 910 may be adapted to process and issue signals concurrently using suitable methods known in the computer related arts.


The server system 910 includes a program element 916 for execution by a CPU. Program element 916 includes functionality to implement the methods described above, including a method for testing luggage screening equipment operators, and includes the necessary networking functionality to allow the server system 910 to communicate with the client systems 902, 904, 906 and 908 over network 912. In a specific implementation, the client systems 902, 904, 906 and 908 include display units responsive to signals received from the server system 910 for displaying information to a user on the basis of messages received from the server system 1110. Certain ones of the clients systems (not shown) may be embodied as image generation devices and may be adapted for sending signal conveying images of contents of luggage items to the server system 910 through network 912.


In a variant, the server system 910 is adapted for receiving images conveying contents of luggage items from a set of image generation devices, the image generation devices is the set having distinct distortion signatures. In such a case, the server system 910 stores respective distortion transformation databases (of the type described with reference to component 314) for each distinct distortion signature in association with the corresponding image generation device in the set. In a specific implementation, the server system 910 includes a memory unit storing a database of threats (of the type described with reference to component 340) whose presence the system is designed to simulate. In this manner it is possible to use a joint database of threats stored by the server system 910 for providing threat image projection functionality for image generation devices having distinct distortion signatures.


Although the present invention has been described in considerable detail with reference to certain preferred embodiments thereof, variations and refinements are possible without departing from the spirit of the invention. Therefore, the scope of the invention should be limited only by the appended claims and their equivalents.

Claims
  • 1. A method for testing luggage screening equipment operators, said method comprising: a. receiving a sequence of images of contents of luggage items derived from a device that scans the luggage items with penetrating radiation, the device being characterized by introducing a certain distortion into the images of contents of luggage items;b. causing a display device to display images derived from the sequence of images of contents of luggage items;c. occasionally causing the display device to show a simulated threat in a luggage item by displaying a combined image derived based at least in part on: i. an image in the sequence of images; andii. a distorted threat image, the distorted threat image having been derived by applying a distortion insertion process to a reference threat image to introduce a distortion in the reference threat image, wherein the distortion insertion process tends to approximate the certain distortion introduced in the images of contents of luggage items by the device that scans the luggage items with penetrating radiation.
  • 2. A method as defined in claim 1, wherein said method comprises: a. occasionally causing the display device to show a same simulated threat positioned at different perceived heights in luggage items by displaying combined images derived based at least in part on: i. images in the sequence of images; andii. a group of distorted threat images, the distorted threat images in said group being associated to respective heights and being derived by applying distinct distortion insertion processes to a reference threat image to introduce distortions in the reference threat image, wherein the distinct distortion insertion processes are associated to respective heights.
  • 3. A method for testing luggage screening equipment operators, said method comprising: a. receiving a sequence of images of contents of luggage items derived from a device that scans the luggage items with penetrating radiation, the device being characterized by introducing a certain distortion into the images of contents of luggage items;b. causing a display device to display images derived from the sequence of images of contents of luggage items;c. occasionally causing the display device to show a simulated threat in a luggage item by: i. applying a distortion insertion process to a reference threat image to derive a distorted threat image, the distortion insertion process tending to approximate the certain distortion introduced in the images of contents of luggage items by the device that scans the luggage items with penetrating radiation;ii. deriving a combined image conveying a simulated threat in the luggage item based at least in part on an image in the sequence of images and the distorted threat image;iii. causing the display device to display the combined image.
  • 4. A method as defined in claim 3, said method comprising: a. processing the image of contents of the luggage item to derive spatial information associated to the luggage item;b. conditioning the distortion insertion process at least in part based on said spatial information.
  • 5. A method as defined in claim 4, wherein the spatial information includes position information associated to the luggage item.
  • 6. A method as defined in claim 5, wherein the position information includes height information associated to the luggage item.
  • 7. A method as defined in claim 4, said method comprising: a. providing a set of original threat images wherein the images in said set are associated to respective potential threats;b. selecting at least one original threat image from the set of original threat images as the reference threat image.
  • 8. A method as defined in claim 7, wherein selecting at least one image from the set of original threat images as the reference threat image is conditioned at least in part based on the spatial information associated to the luggage item.
  • 9. A method as defined in claim 7, wherein the spatial information associated to the luggage item includes size information, said method comprising selecting at least one image from the set of original threat images as the reference threat image based at least in part on the size information associated to the luggage item.
  • 10. A method as defined in claim 4, said method comprising: a. providing an original threat image associated to a threat;b. assigning positioning information to the original threat image to derive the reference threat image.
  • 11. A method as defined in claim 10, said method comprising conditioning the distortion insertion process at least in part based on the positioning information assigned to the original threat image to derive the reference threat image.
  • 12. A method as defined in claim 10, wherein the positioning information assigned to the original threat image to derive the reference threat image includes perceived height information, said method comprising conditioning the distortion insertion process at least in part based on the perceived height information.
  • 13. A method as defined in claim 10, wherein assigning positioning information to the original threat image to derive the reference threat image is conditioned at least in part based on the spatial information associated to the luggage item.
  • 14. A method defined in claim 3, wherein the reference threat image includes intensity data for a set of coordinates, wherein the distorted threat image is defined by modified intensity data for a set of new coordinates, and wherein said applying a distortion insertion process comprises applying an image transformation to the intensity data for the set of coordinates to derive said modified intensity data for the new coordinates.
  • 15. A method defined in claim 14, wherein said image transformation involves processing of a data structure representative of an inferred spatial transformation applied by the device that scans the luggage item with penetrating radiation.
  • 16. A method defined in claim 15, wherein said inferred spatial transformation is two-dimensional.
  • 17. A method defined in claim 15, wherein said data structure is characterized by a set of parameters derived from registration of observed coordinates with respect to a set of reference coordinates.
  • 18. A method as defined in claim 3, wherein said method comprises occasionally causing the display device to show a same simulated threat positioned at different perceived heights in luggage items by: a. applying distortion insertion processes to a reference threat image to derive a plurality of distorted threat images, the distortion insertion processes being associated to respective heights such that the distorted threat images in the plurality of distorted threat images are associated to respective perceived heights;b. deriving combined images conveying the same simulated threat positioned at different perceived heights in luggage items based at least in part on images in the sequence of images and the plurality of distorted threat images;c. causing a display device to display at least some of the combined images.
  • 19. A method as defined in claim 3, where said the image of contents of the luggage item derived is an x-ray image.
  • 20. A method as defined in claim 19, wherein deriving the combined image conveying the simulated threat in the luggage includes performing a combination of attenuation information conveyed in the image in the sequence of images and attenuation information conveyed in the distorted threat image.
  • 21. An apparatus for testing luggage screening equipment operators, said apparatus comprising: a. an input for receiving a sequence of images of contents of luggage items derived from a device that scans the luggage items with penetrating radiation, the device being characterized by introducing a certain distortion into the images of contents of luggage items;b. an output for releasing an output signal suitable for causing a display device to display images derived from the sequence of images of contents of luggage items;c. a processing unit in communication with said input and said output, said processing unit being operative for occasionally causing the output signal released at said output to convey a simulated threat in a luggage item by conveying a combined image derived based at least in part on: i. an image in the sequence of images; andii. a distorted threat image, the distorted threat image having been derived by applying a distortion insertion process to a reference threat image to introduce a distortion in the reference threat image, wherein the distortion insertion process tends to approximate the certain distortion introduced in the images of contents of luggage items by the device that scans the luggage items with penetrating radiation.
  • 22. An apparatus as defined in claim 21, wherein said processing unit is operative for: a. occasionally causing the display device to show a same simulated threat positioned at different perceived heights in luggage items by displaying combined images derived based at least in part on: i. images in the sequence of images; andii. a group of distorted threat images, the distorted threat images in said group being associated to respective heights and being derived by applying distinct distortion insertion processes to a reference threat image to introduce distortions in the reference threat image, wherein the distinct distortion insertion processes are associated to respective heights.
  • 23. An apparatus for testing luggage screening equipment operators, said apparatus comprising: a. an input for receiving a sequence of images of contents of luggage items derived from a device that scans the luggage items with penetrating radiation, the device being characterized by introducing a certain distortion into the images of contents of luggage items;b. an output for releasing an output signal suitable for causing a display device to display images derived from the sequence of images of contents of luggage items;c. a processing unit in communication with said input and said output, said processing unit being operative for occasionally causing the output signal released at said output to convey a simulated threat in a luggage item by: i. applying a distortion insertion process to a reference threat image to derive a distorted threat image, the distortion insertion process tending to approximate the certain distortion introduced in the images of contents of luggage items by the device that scans the luggage items with penetrating radiation;ii. deriving a combined image conveying a simulated threat in the luggage item based at least in part on an image in the sequence of images and the distorted threat image;iii. causing the display device to display the combined image.
  • 24. An apparatus as defined in claim 23, said processing unit being operative for: a. processing the image of contents of the luggage item to derive spatial information associated to the luggage item;b. conditioning the distortion insertion process at least in part based on said spatial information.
  • 25. An apparatus as defined in claim 24, wherein the spatial information includes position information associated to the luggage item.
  • 26. An apparatus as defined in claim 25, wherein the position information includes height information associated to the luggage item.
  • 27. An apparatus as defined in claim 24, wherein said apparatus comprises a memory unit for storing a set of original threat images wherein the images in said set are associated to respective potential threats, said processing unit being operative for selecting at least one image from the set of original threat images in said memory unit as the reference threat image.
  • 28. An apparatus as defined in claim 27, wherein processing unit is operative for selecting at least one image from the set of original threat images as the reference threat image is conditioned at least in part based on the spatial information associated to the luggage item.
  • 29. An apparatus as defined in claim 27, wherein the spatial information associated to the luggage item includes size information, said processing unit being operative for selecting at least one image from the set of original threat images as the reference threat image based at least in part on the size information associated to the luggage item.
  • 30. An apparatus as defined in claim 24, said apparatus comprising a memory unit for storing an original threat image associated to a threat, said processing unit being operative for assigning positioning information to the original threat image to derive the reference threat image.
  • 31. An apparatus as defined in claim 30, said processing unit is operative for conditioning the distortion insertion process at least in part based on the positioning information assigned to the original threat image to derive the reference threat image.
  • 32. An apparatus as defined in claim 30, wherein the positioning information assigned to the original threat image to derive the reference threat image includes desired perceived height information, said processing unit being operative for conditioning the distortion insertion process at least in part based on the perceived height information.
  • 33. An apparatus as defined in claim 30, wherein assigning positioning information to the original threat image to derive the reference threat image is conditioned at least in part based on the spatial information associated to the luggage item.
  • 34. An apparatus defined in claim 23, wherein the reference threat image includes intensity data for a set of coordinates, wherein the distorted threat image is defined by modified intensity data for a set of new coordinates, and wherein said applying a distortion insertion process comprises applying an image transformation to the intensity data for the set of coordinates to derive said modified intensity data for the new coordinates.
  • 35. An apparatus defined in claim 34, wherein said image transformation involves processing of a data structure representative of an inferred spatial transformation applied by the device that scans the luggage item with penetrating radiation.
  • 36. An apparatus defined in claim 35, wherein said inferred spatial transformation is two-dimensional.
  • 37. An apparatus defined in claim 35, wherein said data structure is characterized by a set of parameters derived from registration of observed coordinates with respect to a set of reference coordinates.
  • 38. An apparatus as defined in claim 23, wherein said processing unit is operative for occasionally causing the display device to show a same simulated threat positioned at different perceived heights in luggage items by: a. applying distortion insertion processes to a reference threat image to derive a plurality of distorted threat images, the distortion insertion processes being associated to respective heights such that the distorted threat images in the plurality of distorted threat images are associated to respective perceived heights;b. deriving combined images conveying the same simulated threat positioned at different perceived heights in luggage items based at least in part on images in the sequence of images and the plurality of distorted threat images;c. causing a display device to display at least some of the combined images.
  • 39. An apparatus as defined in claim 23, where said the image of contents of the luggage item derived is an x-ray image.
  • 40. An apparatus as defined in claim 39, wherein deriving the combined image conveying the simulated threat in the luggage includes performing a combination of attenuation information conveyed in the image in the sequence of images and attenuation information conveyed in the distorted threat image.
  • 41. A non-transitory computer readable storage medium storing a program element suitable for execution by a CPU, said program element implementing an apparatus for testing luggage screening equipment operators, said computing apparatus comprising: a. a memory unit;b. a processor operatively connected to said memory unit, said program element when executing on said processor being operative for: i. receiving a sequence of images of contents of luggage items derived from a device that scans the luggage items with penetrating radiation, the device being characterized by introducing a certain distortion into the images of contents of luggage items;ii. causing a display device to display images derived from the sequence of images of contents of luggage items;iii. occasionally causing the display device to show a simulated threat in a luggage item by displaying a combined image derived based at least in part on: 1. an image in the sequence of images; and2. a distorted threat image, the distorted threat image having been derived by applying a distortion insertion process to a reference threat image to introduce a distortion in the reference threat image, wherein the distortion insertion process tends to approximate the certain distortion introduced in the images of contents of luggage items by the device that scans the luggage items with penetrating radiation.
  • 42. A computer readable storage medium as defined in claim 41, said program element when executing on said processor is operative for occasionally causing the display device to show a same simulated threat positioned at different perceived heights in luggage items by displaying combined images derived based at least in part on: i. images in the sequence of images; andii. a group of distorted threat images, the distorted threat images in said group being associated to respective heights and being derived by applying distinct distortion insertion processes to a reference threat image to introduce distortions in the reference threat image, wherein the distinct distortion insertion processes are associated to respective heights.
  • 43. A non-transitory computer readable storage medium storing a program element suitable for execution by a CPU, said program element implementing an apparatus for testing luggage screening equipment operators, said computing apparatus comprising: a. a memory unit;b. a processor operatively connected to said memory unit, said program element when executing on said processor being operative for: i. receiving a sequence of images of contents of luggage items derived from a device that scans the luggage items with penetrating radiation, the device being characterized by introducing a certain distortion into the images of contents of luggage items;ii. causing a display device to display images derived from the sequence of images of contents of luggage items;iii. occasionally causing the display device to show a simulated threat in a luggage item by: 1. applying a distortion insertion process to a reference threat image to derive a distorted threat image, the distortion insertion process tending to approximate the certain distortion introduced in the images of contents of luggage items by the device that scans the luggage items with penetrating radiation;2. deriving a combined image conveying a simulated threat in the luggage item based at least in part on an image in the sequence of images and the distorted threat image;3. causing the display device to display the combined image.
  • 44. A computer readable storage medium as defined in claim 43, wherein said program element when executing on said processor is operative for: a. processing the image of contents of the luggage item to derive spatial information associated to the luggage item;b. conditioning the distortion insertion process at least in part based on said spatial information.
  • 45. A computer readable storage medium as defined in claim 44, wherein the spatial information includes position information associated to the luggage item.
  • 46. A computer readable storage medium as defined in claim 45, wherein the position information includes height information associated to the luggage item.
  • 47. A computer readable storage medium as defined in claim 44, wherein said program element when executing on said processor being operative for: a. providing a set of original threat images wherein the images in said set are associated to respective potential threats;b. selecting at least one image from set of original threat images as the reference threat image.
  • 48. A computer readable storage medium as defined in claim 47, wherein selecting at least one image from the set of original threat images as the reference threat image is conditioned at least in part based on the spatial information associated to the luggage item.
  • 49. A computer readable storage medium as defined in claim 47, wherein the spatial information associated to the luggage item includes size information, said program element when executing on said processor being operative for selecting at least one image from the set of original threat images as the reference threat image based at least in part on the size information associated to the luggage item.
  • 50. A computer readable storage medium as defined in claim 44, said program element when executing on said processor being operative for: a. providing an original threat image associated to a threat;b. assigning positioning information to the original threat image to derive the reference threat image.
  • 51. A computer readable storage medium as defined in claim 50, said program element when executing on said processor being operative for conditioning the distortion insertion process at least in part based on the positioning information assigned to the original threat image to derive the reference threat image.
  • 52. A computer readable storage medium as defined in claim 50, wherein the positioning information assigned to the original threat image to derive the reference threat image includes perceived height information, said program element when executing on said processor being operative for conditioning the distortion insertion process at least in part based on the perceived height information.
  • 53. A computer readable storage medium as defined in claim 50, wherein assigning positioning information to the original threat image to derive the reference threat image is conditioned at least in part based on the spatial information associated to the luggage item.
  • 54. A computer readable storage medium defined in claim 43, wherein the reference threat image includes intensity data for a set of coordinates, wherein the distorted threat image is defined by modified intensity data for a set of new coordinates, and wherein said applying a distortion insertion process comprises applying an image transformation to the intensity data for the set of coordinates to derive said modified intensity data for the new coordinates.
  • 55. A computer readable storage medium defined in claim 54, wherein said image transformation involves processing of a data structure representative of an inferred spatial transformation applied by the device that scans the luggage item with penetrating radiation.
  • 56. A computer readable storage medium defined in claim 55, wherein said inferred spatial transformation is two-dimensional.
  • 57. A computer readable storage medium defined in claim 55, wherein said data structure is characterized by a set of parameters derived from registration of observed coordinates with respect to a set of reference coordinates.
  • 58. A computer readable storage medium as defined in claim 43, wherein said program element when executing on said processor being operative for occasionally causing the display device to show a same simulated threat positioned at different perceived heights in luggage items by: a. applying distortion insertion processes to a reference threat image to derive a plurality of distorted threat images, the distortion insertion process approximating a distortion introduced in the images of contents of luggage items by the device that scans the luggage items with penetrating radiation, the distortion insertion processes being associated to respective heights such that the distorted threat images in the plurality of distorted threat images are associated to respective perceived heights;b. deriving combined images conveying the same simulated threat positioned at different perceived heights in luggage items based at least in part on images in the sequence of images and the plurality of distorted threat images;c. causing a display device to display at least some of the combined images.
  • 59. A computer readable storage medium as defined in claim 43, where said the image of contents of the luggage item derived is an x-ray image.
  • 60. A computer readable storage medium as defined in claim 59, wherein deriving the combined image conveying the simulated threat in the luggage includes performing a combination of attenuation information conveyed in the image in the sequence of images and attenuation information conveyed in the distorted threat image.
  • 61. A system for screening luggage items, comprising: a. an image generation device suitable for scanning luggage items with penetrating radiation to generate a sequence of images of contents of the luggage items, the image generation device being characterized by introducing a certain distortion into the images of contents of luggage items;b. a display device;c. an apparatus in communication with said image generation device and said display device, said apparatus including functionality for testing luggage screening equipment operators, said apparatus comprising: i. an input for receiving the sequence of images of contents of luggage items;ii. an output for releasing an output signal suitable for causing the display device to display images derived from the sequence of images of contents of luggage items;iii. a processing unit in communication with said input and said output, said processing unit being operative for occasionally causing the output signal released at said output to convey a simulated threat in a luggage item by conveying a combined image derived based at least in part on: 1. an image in the sequence of images; and2. a distorted threat image, the distorted threat image having been derived by applying a distortion insertion process to a reference threat image to introduce a distortion in the reference threat image, wherein the distortion insertion process tends to approximates the certain distortion introduced in the images of contents of luggage items by the image generation device.
  • 62. A client-server system for implementing a process for testing luggage screening equipment operators, said client-server system comprising a client system and a server system, said client system and said server system operative to exchange messages over a data network, said server system storing a program element for execution by a CPU, said program element comprising: a. first program element component executed on said server system for receiving a sequence of images of contents of luggage items derived from a device that scans the luggage items with penetrating radiation, the device being characterized by introducing a certain distortion into the images of contents of luggage items;b. second program element component executed on said server system for generating messages for transmission to said client system for causing a display device in communication with said client system to display images derived from the sequence of images of contents of luggage items;c. wherein said second program element component is operative for occasionally generating messages for causing the display device in communication with said client system to show a simulated threat in a luggage item by conveying a combined image derived based at least in part on: i. an image in the sequence of images; andii. a distorted threat image, the distorted threat image having been derived by applying a distortion insertion process to a reference threat image to introduce a distortion in the reference threat image, wherein the distortion insertion process approximates the certain distortion introduced in the images of contents of luggage items by the device that scans the luggage items with penetrating radiation.
  • 63. An apparatus for testing luggage screening equipment operators, said apparatus comprising: a. means for receiving a sequence of images of contents of luggage items derived from a device that scans the luggage items with penetrating radiation;b. means for causing a display device to display images derived from the sequence of images of contents of luggage items;c. a means for occasionally causing the display device to show a simulated threat in a luggage item by conveying a combined image derived based at least in part on: i. an image in the sequence of images; andii. a distorted threat image, the distorted threat image having been derived by applying a distortion insertion process to a reference threat image to introduce a distortion in the reference threat image, wherein the distortion insertion process approximates the certain distortion introduced in the images of contents of luggage items by the device that scans the luggage items with penetrating radiation.
CROSS-REFERENCE TO RELATED APPLICATIONS

This application is a continuation-in-part claiming the benefit under 35 USC §120 of U.S. patent application Ser. No. 11/431,627 entitled “APPARATUS, METHOD AND SYSTEM FOR SCREENING RECEPTACLES AND PERSONS, HAVING IMAGE DISTORTION CORRECTION FUNCTIONALITY”, filed on May 11, 2006 by L. Perron et al., and presently pending, the contents of which are incorporated herein by reference.

US Referenced Citations (625)
Number Name Date Kind
4338626 Lemelson Jul 1982 A
4379348 Haas et al. Apr 1983 A
4383327 Kruger May 1983 A
4470303 O'Donnell Sep 1984 A
4480899 Sprague Nov 1984 A
4481575 Bazlen et al. Nov 1984 A
4482958 Nakayama et al. Nov 1984 A
4509075 Simms et al. Apr 1985 A
4573198 Anderson Feb 1986 A
4612666 King Sep 1986 A
4637056 Sherman et al. Jan 1987 A
4651957 Minnich, II Mar 1987 A
4653109 Lemelson et al. Mar 1987 A
4722096 Dietrich et al. Jan 1988 A
4724543 Klevecz et al. Feb 1988 A
4725733 Horman et al. Feb 1988 A
4736399 Okazaki Apr 1988 A
4736401 Donges et al. Apr 1988 A
4737650 West Apr 1988 A
4756015 Doenges et al. Jul 1988 A
4759047 Donges et al. Jul 1988 A
4775895 Traupe et al. Oct 1988 A
4783794 Dietrich Nov 1988 A
4788704 Donges et al. Nov 1988 A
4795253 Sandridge et al. Jan 1989 A
4819188 Matsubara et al. Apr 1989 A
4832447 Javidi May 1989 A
4837733 Shiraishi et al. Jun 1989 A
4838644 Ochoa et al. Jun 1989 A
4841554 Doenges et al. Jun 1989 A
4849912 Leberl et al. Jul 1989 A
4862358 Kimura et al. Aug 1989 A
4869574 Hartman Sep 1989 A
4870670 Geus Sep 1989 A
4884289 Glockmann et al. Nov 1989 A
4887899 Hung Dec 1989 A
4916722 Ema Apr 1990 A
4955060 Katsuki et al. Sep 1990 A
5003616 Orita et al. Mar 1991 A
5018178 Katsumata May 1991 A
5020111 Weber May 1991 A
5022062 Annis Jun 1991 A
5034812 Rawlings Jul 1991 A
5041993 Rawlings Aug 1991 A
5056130 Engel Oct 1991 A
5060249 Eisen et al. Oct 1991 A
5063602 Peppers et al. Nov 1991 A
5065418 Bermbach et al. Nov 1991 A
5073782 Huguenin et al. Dec 1991 A
5079698 Grenier et al. Jan 1992 A
5091924 Bermbach et al. Feb 1992 A
5107351 Leib et al. Apr 1992 A
5109276 Nudelman et al. Apr 1992 A
5132811 Iwaki et al. Jul 1992 A
5132842 Yeh Jul 1992 A
5132998 Tsutsui et al. Jul 1992 A
5138167 Barnes Aug 1992 A
5150229 Takesue et al. Sep 1992 A
5179581 Annis Jan 1993 A
5181234 Smith Jan 1993 A
5198669 Namiki et al. Mar 1993 A
5216541 Takesue et al. Jun 1993 A
5239595 Takemura et al. Aug 1993 A
5257085 Ulich et al. Oct 1993 A
5257322 Matsuoka et al. Oct 1993 A
5268967 Jang et al. Dec 1993 A
5283641 Lemelson Feb 1994 A
5297222 Mori et al. Mar 1994 A
5309244 Katagiri et al. May 1994 A
5309523 Iwaki et al. May 1994 A
5311359 Lucas et al. May 1994 A
5319547 Krug et al. Jun 1994 A
5323472 Falk Jun 1994 A
5327286 Sampsell et al. Jul 1994 A
5345081 Rogers Sep 1994 A
5345173 Bito et al. Sep 1994 A
5365560 Tam Nov 1994 A
5365564 Yashida et al. Nov 1994 A
5367552 Peschmann Nov 1994 A
5371542 Pauli et al. Dec 1994 A
5375156 Kuo-Petravic et al. Dec 1994 A
5376796 Chan et al. Dec 1994 A
5379334 Zimmer et al. Jan 1995 A
5379336 Kramer et al. Jan 1995 A
5418380 Simon et al. May 1995 A
5420788 Vissers May 1995 A
5425113 Ito Jun 1995 A
5428657 Papanicolopoulos et al. Jun 1995 A
5430787 Norton Jul 1995 A
5481584 Tang et al. Jan 1996 A
5481622 Gerhardt et al. Jan 1996 A
5483569 Annis Jan 1996 A
5485312 Horner et al. Jan 1996 A
5490218 Krug et al. Feb 1996 A
5493444 Khoury et al. Feb 1996 A
5506880 Scardino et al. Apr 1996 A
5519225 Mohr et al. May 1996 A
5524133 Neale et al. Jun 1996 A
5528702 Mitsuoka et al. Jun 1996 A
5528703 Lee Jun 1996 A
5546189 Svetkoff et al. Aug 1996 A
5568256 Korner et al. Oct 1996 A
5580471 Fukumoto et al. Dec 1996 A
5595767 Cinquin et al. Jan 1997 A
5600303 Husseiny et al. Feb 1997 A
5600485 Iwaki et al. Feb 1997 A
5600700 Krug et al. Feb 1997 A
5604634 Khoury et al. Feb 1997 A
5619596 Iwaki et al. Apr 1997 A
5625192 Oda et al. Apr 1997 A
5625717 Hashimoto et al. Apr 1997 A
5638420 Armistead Jun 1997 A
5642393 Krug et al. Jun 1997 A
5642394 Rothschild Jun 1997 A
5647018 Benjamin Jul 1997 A
5664574 Chance Sep 1997 A
5668846 Fox et al. Sep 1997 A
5680525 Sakai et al. Oct 1997 A
5684565 Oshida et al. Nov 1997 A
5692028 Geus et al. Nov 1997 A
5692029 Husseiny et al. Nov 1997 A
5692446 Becker et al. Dec 1997 A
5699400 Lee et al. Dec 1997 A
5703921 Fujita et al. Dec 1997 A
5706816 Mochizuki et al. Jan 1998 A
5726449 Yoshiike et al. Mar 1998 A
5739539 Wang et al. Apr 1998 A
5745542 Gordon et al. Apr 1998 A
5748305 Shimono et al. May 1998 A
5748697 Tam May 1998 A
5754621 Suzuki et al. May 1998 A
5756875 Parker et al. May 1998 A
5757981 Kawakubo May 1998 A
5761334 Nakajima et al. Jun 1998 A
5764683 Swift et al. Jun 1998 A
5764719 Noettling Jun 1998 A
5768334 Maitrejean et al. Jun 1998 A
5777742 Marron Jul 1998 A
5778046 Molloi et al. Jul 1998 A
5779641 Hatfield et al. Jul 1998 A
5784429 Arai Jul 1998 A
5786597 Lingren et al. Jul 1998 A
5787145 Geus Jul 1998 A
5794788 Massen Aug 1998 A
5796802 Gordon Aug 1998 A
5796868 Dutta-Choudhury Aug 1998 A
5799100 Clarke et al. Aug 1998 A
5800355 Hasegawa Sep 1998 A
5802133 Kawai et al. Sep 1998 A
5809171 Neff et al. Sep 1998 A
5815198 Vachtsevanos et al. Sep 1998 A
5815264 Reed et al. Sep 1998 A
5828722 Ploetz et al. Oct 1998 A
5828774 Wang Oct 1998 A
5834153 Hasegawa et al. Nov 1998 A
5838758 Krug et al. Nov 1998 A
5838759 Armistead Nov 1998 A
5841828 Gordon et al. Nov 1998 A
5841907 Javidi et al. Nov 1998 A
5850465 Shimura et al. Dec 1998 A
5862198 Samarasekera et al. Jan 1999 A
5862258 Taylor Jan 1999 A
5864598 Hsieh et al. Jan 1999 A
5866907 Drukier et al. Feb 1999 A
5877849 Ramer et al. Mar 1999 A
5881123 Tam Mar 1999 A
5893095 Jain et al. Apr 1999 A
5894345 Takamoto et al. Apr 1999 A
5901196 Sauer et al. May 1999 A
5901198 Crawford et al. May 1999 A
5903623 Swift et al. May 1999 A
5909285 Beaty et al. Jun 1999 A
5909477 Crawford et al. Jun 1999 A
5910765 Slemon et al. Jun 1999 A
5910973 Grodzins Jun 1999 A
5911139 Jain et al. Jun 1999 A
5917190 Yodh et al. Jun 1999 A
5926568 Chaney et al. Jul 1999 A
5940468 Huang et al. Aug 1999 A
5943388 Tumer Aug 1999 A
5951474 Matsunaga et al. Sep 1999 A
5953452 Boone et al. Sep 1999 A
5960104 Conners et al. Sep 1999 A
5974111 Krug et al. Oct 1999 A
5978440 Kang et al. Nov 1999 A
5981949 Leahy et al. Nov 1999 A
5987095 Chapman et al. Nov 1999 A
6005916 Johnson et al. Dec 1999 A
6008496 Winefordner et al. Dec 1999 A
6009142 Sauer et al. Dec 1999 A
6011620 Sites et al. Jan 2000 A
6018561 Tam Jan 2000 A
6018562 Willson Jan 2000 A
6031890 Bermbach et al. Feb 2000 A
6035014 Hiraoglu et al. Mar 2000 A
6043870 Chen Mar 2000 A
6049381 Reintjes et al. Apr 2000 A
6057761 Yukl May 2000 A
6057909 Yahav et al. May 2000 A
6058159 Conway et al. May 2000 A
6060677 Ulrichsen et al. May 2000 A
6070583 Perelman et al. Jun 2000 A
6075591 Vokhmin Jun 2000 A
6075880 Kollhof et al. Jun 2000 A
6078638 Sauer et al. Jun 2000 A
6080994 Carrott et al. Jun 2000 A
6081580 Grodzins et al. Jun 2000 A
6084939 Tamura Jul 2000 A
6088423 Krug et al. Jul 2000 A
6094472 Smith Jul 2000 A
6097427 Dey et al. Aug 2000 A
6097483 Komatsu Aug 2000 A
6149300 Greenway et al. Nov 2000 A
6153873 Wolf Nov 2000 A
6155179 Aust et al. Dec 2000 A
6157730 Roever et al. Dec 2000 A
6163403 Carrott et al. Dec 2000 A
6175417 Do et al. Jan 2001 B1
6175613 Boutenko et al. Jan 2001 B1
6188747 Geus et al. Feb 2001 B1
6195413 Geus et al. Feb 2001 B1
6195444 Simanovsky et al. Feb 2001 B1
6198795 Naumann et al. Mar 2001 B1
6205195 Lanza Mar 2001 B1
6205243 Migdal et al. Mar 2001 B1
6218943 Ellenbogen Apr 2001 B1
6222902 Lin et al. Apr 2001 B1
6229872 Amos May 2001 B1
6233303 Tam May 2001 B1
6236704 Navab et al. May 2001 B1
6236708 Lin et al. May 2001 B1
6249341 Basiji et al. Jun 2001 B1
6252929 Swift et al. Jun 2001 B1
6256370 Yavuz Jul 2001 B1
6256404 Gordon et al. Jul 2001 B1
6263044 Joosten Jul 2001 B1
6263231 Reitter Jul 2001 B1
6272204 Amtower et al. Aug 2001 B1
6272233 Takeo Aug 2001 B1
6278760 Ogawa et al. Aug 2001 B1
6288974 Nelson Sep 2001 B1
6289235 Webber et al. Sep 2001 B1
6292260 Lin et al. Sep 2001 B1
6292530 Yavus et al. Sep 2001 B1
6292533 Swift et al. Sep 2001 B1
6324245 Tam Nov 2001 B1
6353673 Shnitser et al. Mar 2002 B1
6366638 Hsieh et al. Apr 2002 B1
6370222 Cornick, Jr. Apr 2002 B1
6373916 Inoue et al. Apr 2002 B1
6373970 Dong et al. Apr 2002 B1
6381297 Hsieh Apr 2002 B1
6388788 Harris et al. May 2002 B1
6403960 Wellnitz et al. Jun 2002 B1
6404841 Pforr et al. Jun 2002 B1
6408042 Hsieh Jun 2002 B1
6415012 Taguchi et al. Jul 2002 B1
6418184 Wang et al. Jul 2002 B1
6418189 Schafer Jul 2002 B1
6424692 Suzuki Jul 2002 B1
6442288 Haerer et al. Aug 2002 B1
6445765 Frank et al. Sep 2002 B1
6448545 Chen Sep 2002 B1
6453003 Springer et al. Sep 2002 B1
6459755 Li Oct 2002 B1
6463181 Duarte Oct 2002 B2
6473489 Bani-Hashemi et al. Oct 2002 B2
6477221 Ning Nov 2002 B1
6480285 Hill Nov 2002 B1
6480564 Kim et al. Nov 2002 B1
6483894 Hartick et al. Nov 2002 B2
6487307 Hennessey et al. Nov 2002 B1
6502984 Ogura et al. Jan 2003 B2
6507025 Verbinski et al. Jan 2003 B1
6507278 Brunetti et al. Jan 2003 B1
6525331 Ngoi et al. Feb 2003 B1
6526120 Gray et al. Feb 2003 B1
6532276 Hartick et al. Mar 2003 B1
6542574 Grodzins Apr 2003 B2
6542578 Ries et al. Apr 2003 B2
6542579 Takasawa Apr 2003 B1
6542580 Carver et al. Apr 2003 B1
6542628 Muller et al. Apr 2003 B1
6549683 Bergeron et al. Apr 2003 B1
6552809 Bergeron et al. Apr 2003 B1
6559769 Anthony et al. May 2003 B2
6570177 Struckhoff et al. May 2003 B1
6570708 Bergeron et al. May 2003 B1
6570951 Hsieh May 2003 B1
6570956 Rhee et al. May 2003 B1
6574296 Stierstorfer Jun 2003 B2
6574297 Tam Jun 2003 B2
6580777 Ueki et al. Jun 2003 B1
6580778 Meder Jun 2003 B2
6583895 Kuwahara et al. Jun 2003 B1
6584170 Aust et al. Jun 2003 B2
6586193 Yguerabide et al. Jul 2003 B2
6587575 Windham et al. Jul 2003 B1
6587595 Henkel et al. Jul 2003 B1
6597760 Beneke et al. Jul 2003 B2
6603536 Hasson et al. Aug 2003 B1
6608921 Inoue et al. Aug 2003 B1
6611575 Alyassin et al. Aug 2003 B1
6618466 Ning Sep 2003 B1
6621887 Albagli et al. Sep 2003 B2
6621888 Grodzins et al. Sep 2003 B2
6621925 Ohmori et al. Sep 2003 B1
6628982 Thomas et al. Sep 2003 B1
6628983 Gagnon Sep 2003 B1
6654443 Hoffman Nov 2003 B1
6663280 Doenges Dec 2003 B2
6665373 Kotowski et al. Dec 2003 B1
6707879 McClelland et al. Mar 2004 B2
6714623 Sako et al. Mar 2004 B2
6721387 Naidu et al. Apr 2004 B1
6721391 McClelland et al. Apr 2004 B2
6724922 Vilsmeier Apr 2004 B1
6731819 Fukushima et al. May 2004 B1
6735274 Zahavi et al. May 2004 B1
6735279 Jacobs et al. May 2004 B1
6738450 Barford May 2004 B1
6744909 Kostrzewski et al. Jun 2004 B1
6746864 McNeil et al. Jun 2004 B1
6751349 Matama Jun 2004 B2
6754374 Miller et al. Jun 2004 B1
6763148 Sternberg et al. Jul 2004 B1
6785410 Vining et al. Aug 2004 B2
H2110 Newman Oct 2004 H
6801647 Arakawa Oct 2004 B1
6803997 Stanek Oct 2004 B2
6804412 Wilkinson Oct 2004 B1
6813395 Kinjo Nov 2004 B1
6825854 Beneke et al. Nov 2004 B1
6837422 Meder Jan 2005 B1
6839403 Kotowski et al. Jan 2005 B1
6839406 Ries et al. Jan 2005 B2
6843599 Le et al. Jan 2005 B2
6856272 Levitan et al. Feb 2005 B2
6865287 Beneke Mar 2005 B1
6865509 Hsiung et al. Mar 2005 B1
6868138 Clinthorne et al. Mar 2005 B2
6873261 Anthony et al. Mar 2005 B2
6876322 Keller Apr 2005 B2
6895072 Schrock et al. May 2005 B2
6895338 Hsiung et al. May 2005 B2
6899540 Neiderman et al. May 2005 B1
6918541 Knowles et al. Jul 2005 B2
6928141 Carver et al. Aug 2005 B2
6936828 Saccomanno Aug 2005 B2
6938488 Diaz et al. Sep 2005 B2
6940943 Claus et al. Sep 2005 B2
6950492 Besson Sep 2005 B2
6952163 Huey et al. Oct 2005 B2
6970531 Eberhard et al. Nov 2005 B2
6980681 Hsieh Dec 2005 B1
6982643 Garfinkle Jan 2006 B2
6990171 Toth et al. Jan 2006 B2
7000827 Meder Feb 2006 B2
7017256 Joly et al. Mar 2006 B2
7020241 Beneke et al. Mar 2006 B2
7043474 Mojsilovic et al. May 2006 B2
7045787 Verbinski et al. May 2006 B1
7046761 Ellenbogen et al. May 2006 B2
7050616 Hsieh et al. May 2006 B2
7062074 Beneke Jun 2006 B1
7065175 Green Jun 2006 B2
7068751 Toth et al. Jun 2006 B2
7092485 Kravis Aug 2006 B2
7098461 Endo Aug 2006 B2
7099004 Masten Aug 2006 B2
7099432 Ichihara et al. Aug 2006 B2
7100165 Eldridge et al. Aug 2006 B2
7103137 Seppi et al. Sep 2006 B2
7105828 Unger et al. Sep 2006 B2
7116749 Besson Oct 2006 B2
7130456 Hillmann Oct 2006 B2
7136716 Hsiung et al. Nov 2006 B2
7139406 McClelland et al. Nov 2006 B2
7142633 Eberhard et al. Nov 2006 B2
7154650 Lettington Dec 2006 B2
7164750 Nabors et al. Jan 2007 B2
7183906 Zanovitch et al. Feb 2007 B2
7193515 Roberts et al. Mar 2007 B1
7212113 Zanovitch May 2007 B2
7212661 Samara et al. May 2007 B2
7233682 Levine Jun 2007 B2
7244941 Roos et al. Jul 2007 B2
7257189 Modica et al. Aug 2007 B2
20010016030 Nicolas et al. Aug 2001 A1
20010021013 Hecht et al. Sep 2001 A1
20010021244 Suzuki et al. Sep 2001 A1
20010028696 Yamada et al. Oct 2001 A1
20010033636 Hartick et al. Oct 2001 A1
20010038681 Stanton et al. Nov 2001 A1
20010038705 Rubbert et al. Nov 2001 A1
20010038707 Ohara Nov 2001 A1
20010048734 Uppaluri et al. Dec 2001 A1
20010053197 Murayama et al. Dec 2001 A1
20020001366 Tamura et al. Jan 2002 A1
20020015475 Matsumoto et al. Feb 2002 A1
20020016546 Cerofolini Feb 2002 A1
20020017620 Oomori et al. Feb 2002 A1
20020018199 Blumenfeld et al. Feb 2002 A1
20020024016 Endo Feb 2002 A1
20020027970 Chapman et al. Mar 2002 A1
20020028994 Kamiyama Mar 2002 A1
20020031246 Kawano Mar 2002 A1
20020037068 Oikawa Mar 2002 A1
20020044691 Matsugu Apr 2002 A1
20020054694 Vachtsevanos et al. May 2002 A1
20020067259 Fufidio et al. Jun 2002 A1
20020067793 Stierstorfer Jun 2002 A1
20020085046 Furuta et al. Jul 2002 A1
20020088952 Rao et al. Jul 2002 A1
20020094062 Dolazza et al. Jul 2002 A1
20020094119 Sahadevan Jul 2002 A1
20020098518 Levinson Jul 2002 A1
20020106052 Menhardt Aug 2002 A1
20020122528 Besson Sep 2002 A1
20020124664 Call et al. Sep 2002 A1
20020126800 Matsumoto et al. Sep 2002 A1
20020127586 Mortensen Sep 2002 A1
20020141625 Nelson Oct 2002 A1
20020150200 Zonneveld Oct 2002 A1
20020161534 Adler et al. Oct 2002 A1
20020168083 Garms et al. Nov 2002 A1
20020168657 Chen et al. Nov 2002 A1
20020172324 Ellengogen Nov 2002 A1
20020172409 Saito et al. Nov 2002 A1
20020175921 Xu et al. Nov 2002 A1
20020176534 Meder Nov 2002 A1
20020186862 McClelland et al. Dec 2002 A1
20020188197 Bishop et al. Dec 2002 A1
20020191209 Yasumaru Dec 2002 A1
20030012420 Verwoerd et al. Jan 2003 A1
20030023592 Modica et al. Jan 2003 A1
20030024315 Merkel et al. Feb 2003 A1
20030031289 Hsieh Feb 2003 A1
20030031291 Yamamoto et al. Feb 2003 A1
20030036006 Feke et al. Feb 2003 A1
20030038945 Mahner Feb 2003 A1
20030072414 Sakaida Apr 2003 A1
20030072418 Albagli et al. Apr 2003 A1
20030072484 Kokko et al. Apr 2003 A1
20030076924 Mario et al. Apr 2003 A1
20030081720 Swift et al. May 2003 A1
20030081859 Kasutani May 2003 A1
20030082516 Straus May 2003 A1
20030085348 Megerle May 2003 A1
20030085353 Almogy et al. May 2003 A1
20030091145 Mohr et al. May 2003 A1
20030095633 Van Woezik May 2003 A1
20030095692 Mundy et al. May 2003 A1
20030128812 Appleby et al. Jul 2003 A1
20030138147 Ongkojoyo Jul 2003 A1
20030148393 Woodbury et al. Aug 2003 A1
20030149346 Arnone et al. Aug 2003 A1
20030165213 Maglich Sep 2003 A1
20030179853 Amemiya et al. Sep 2003 A1
20030194121 Eberhard et al. Oct 2003 A1
20030205676 Nelson et al. Nov 2003 A1
20030206649 Moshe Nov 2003 A1
20030210139 Brooks et al. Nov 2003 A1
20030215051 Suzuki Nov 2003 A1
20030215143 Zakrzewski et al. Nov 2003 A1
20030231788 Yukhin et al. Dec 2003 A1
20030231791 Torre-Bueno et al. Dec 2003 A1
20040012853 Garcia et al. Jan 2004 A1
20040013239 Gregerson et al. Jan 2004 A1
20040016271 Shah et al. Jan 2004 A1
20040017882 Misawa et al. Jan 2004 A1
20040017883 Takagi et al. Jan 2004 A1
20040017888 Seppi et al. Jan 2004 A1
20040017935 Avinash et al. Jan 2004 A1
20040022425 Avinash et al. Feb 2004 A1
20040027127 Mills Feb 2004 A1
20040037462 Lewis et al. Feb 2004 A1
20040041082 Harmon Mar 2004 A1
20040051030 Olszak et al. Mar 2004 A1
20040062342 Cahill Apr 2004 A1
20040062349 Schuster Apr 2004 A1
20040062351 Yoshioka Apr 2004 A1
20040066882 Eberhard et al. Apr 2004 A1
20040066884 Hermann Claus et al. Apr 2004 A1
20040066890 Dalmijn et al. Apr 2004 A1
20040075058 Blevis et al. Apr 2004 A1
20040080315 Beevor et al. Apr 2004 A1
20040082846 Johnson et al. Apr 2004 A1
20040083958 Saidman et al. May 2004 A1
20040086075 Hein et al. May 2004 A1
20040086160 Zimmermann May 2004 A1
20040087844 Yen May 2004 A1
20040102700 Asafusa May 2004 A1
20040109231 Haisch et al. Jun 2004 A1
20040120857 Smith et al. Jun 2004 A1
20040141056 Izumi et al. Jul 2004 A1
20040142386 Rigler et al. Jul 2004 A1
20040160599 Hamamatsu et al. Aug 2004 A1
20040161073 Nokita Aug 2004 A1
20040175041 Miller Sep 2004 A1
20040176677 Hwu et al. Sep 2004 A1
20040212492 Boesch et al. Oct 2004 A1
20040213377 Endo Oct 2004 A1
20040213600 Watanabe et al. Oct 2004 A1
20040218729 Xue et al. Nov 2004 A1
20040225222 Zeng et al. Nov 2004 A1
20040236520 Williams et al. Nov 2004 A1
20040240612 Suzuki Dec 2004 A1
20040247071 Dafni Dec 2004 A1
20040247171 Hashimoto et al. Dec 2004 A1
20040252024 Huey et al. Dec 2004 A1
20040252870 Reeves et al. Dec 2004 A1
20040253660 Gibbs et al. Dec 2004 A1
20040258198 Carver et al. Dec 2004 A1
20040258202 Wernick et al. Dec 2004 A1
20040263379 Keller Dec 2004 A1
20040264624 Tanaka et al. Dec 2004 A1
20040264648 Claus et al. Dec 2004 A1
20040265175 Witty et al. Dec 2004 A1
20050008119 McClelland et al. Jan 2005 A1
20050008203 Dixon Jan 2005 A1
20050017181 Kearfott et al. Jan 2005 A1
20050018812 Wolfs Jan 2005 A1
20050025280 Schulte Feb 2005 A1
20050025350 Engelbart et al. Feb 2005 A1
20050031069 Kaucic et al. Feb 2005 A1
20050053307 Nose et al. Mar 2005 A1
20050057354 Jenkins et al. Mar 2005 A1
20050058242 Peschmann Mar 2005 A1
20050058350 Dugan et al. Mar 2005 A1
20050061955 Endo Mar 2005 A1
20050069085 Lewis Mar 2005 A1
20050074088 Ichihara et al. Apr 2005 A1
20050085721 Fauver et al. Apr 2005 A1
20050094856 Warren May 2005 A1
20050098728 Alfano et al. May 2005 A1
20050105680 Nabors et al. May 2005 A1
20050110672 Cardiasmenos et al. May 2005 A1
20050111618 Sommer, Jr. et al. May 2005 A1
20050113961 Sabol et al. May 2005 A1
20050117693 Miyano Jun 2005 A1
20050117700 Peschmann Jun 2005 A1
20050123093 Lawaczeck et al. Jun 2005 A1
20050123174 Gorsky et al. Jun 2005 A1
20050128069 Skatter Jun 2005 A1
20050133708 Eberhard et al. Jun 2005 A1
20050147199 Dunham et al. Jul 2005 A1
20050153356 Okawa et al. Jul 2005 A1
20050163354 Ziegler Jul 2005 A1
20050173284 Ambrefe, Jr. Aug 2005 A1
20050189412 Hudnut et al. Sep 2005 A1
20050190882 McGuire Sep 2005 A1
20050206514 Zanovitch et al. Sep 2005 A1
20050207655 Chopra et al. Sep 2005 A1
20050212913 Richter Sep 2005 A1
20050219523 Onuma et al. Oct 2005 A1
20050220264 Homegger Oct 2005 A1
20050226375 Eberhard et al. Oct 2005 A1
20050240858 Croft et al. Oct 2005 A1
20050248450 Zanovitch Nov 2005 A1
20050249416 Leue et al. Nov 2005 A1
20050251397 Zanovitch et al. Nov 2005 A1
20050251398 Zanovitch et al. Nov 2005 A1
20050259868 Sones Nov 2005 A1
20050265517 Gary Dec 2005 A1
20050271184 Ovadia Dec 2005 A1
20050275831 Silver Dec 2005 A1
20050276443 Slamani et al. Dec 2005 A1
20050279936 Litman et al. Dec 2005 A1
20050283079 Steen et al. Dec 2005 A1
20060000911 Stekel Jan 2006 A1
20060002504 De Man et al. Jan 2006 A1
20060008054 Ohara Jan 2006 A1
20060009269 Hoskinson et al. Jan 2006 A1
20060013455 Watson et al. Jan 2006 A1
20060013464 Ramsay et al. Jan 2006 A1
20060017605 Lovberg et al. Jan 2006 A1
20060018434 Jacobs et al. Jan 2006 A1
20060018517 Chen et al. Jan 2006 A1
20060019409 Nelson et al. Jan 2006 A1
20060034503 Shimayama Feb 2006 A1
20060036167 Shina Feb 2006 A1
20060045235 Bruder et al. Mar 2006 A1
20060045323 Ateya Mar 2006 A1
20060064246 Medberry et al. Mar 2006 A1
20060065844 Zelakiewicz et al. Mar 2006 A1
20060072702 Chapman Apr 2006 A1
20060083418 Watson et al. Apr 2006 A1
20060084872 Ichikawa et al. Apr 2006 A1
20060086794 Knowles et al. Apr 2006 A1
20060093088 Sowerby et al. May 2006 A1
20060098773 Peschmann May 2006 A1
20060098866 Whitson et al. May 2006 A1
20060115109 Whitson et al. Jun 2006 A1
20060116566 Bruijns Jun 2006 A1
20060119837 Raguin et al. Jun 2006 A1
20060133650 Xie et al. Jun 2006 A1
20060133659 Hammond Jun 2006 A1
20060142662 Van Beek Jun 2006 A1
20060142984 Weese et al. Jun 2006 A1
20060173268 Mullick et al. Aug 2006 A1
20060176062 Yang et al. Aug 2006 A1
20060203960 Schlomka et al. Sep 2006 A1
20060204080 Sones et al. Sep 2006 A1
20060215811 Modica et al. Sep 2006 A1
20060255929 Zanovitch et al. Nov 2006 A1
20060257005 Bergeron et al. Nov 2006 A1
20060262902 Wattenburg Nov 2006 A1
20060269135 Ramsay et al. Nov 2006 A1
20060273257 Roos et al. Dec 2006 A1
20060274916 Chan et al. Dec 2006 A1
20060282886 Gaug Dec 2006 A1
20070003122 Sirohey et al. Jan 2007 A1
20070041612 Perron et al. Feb 2007 A1
20070041613 Perron et al. Feb 2007 A1
20070058037 Bergeron et al. Mar 2007 A1
20070147585 Eilbert et al. Jun 2007 A1
20070168467 Hu et al. Jul 2007 A1
20070195994 McClelland et al. Aug 2007 A1
20070200566 Clark et al. Aug 2007 A1
20070206719 Suryanarayanan et al. Sep 2007 A1
20070210921 Volpi et al. Sep 2007 A1
20070269005 Chalmers et al. Nov 2007 A1
20080236275 Breed et al. Oct 2008 A1
20080260097 Anwar et al. Oct 2008 A1
Foreign Referenced Citations (27)
Number Date Country
2307439 May 2000 CA
2319958 Sep 2000 CA
2574402 Jan 2006 CA
0 577 380 Jan 1994 EP
WO 02082290 Oct 2002 WO
WO 03069498 Aug 2003 WO
WO 03107113 Dec 2003 WO
WO 2005086616 Sep 2005 WO
PCTCA2005000716 Feb 2006 WO
PCTCA2005001930 Apr 2006 WO
PCTCA2006000655 Aug 2006 WO
PCTCA2006000751 Aug 2006 WO
WO 2006119603 Nov 2006 WO
PCTCA2007000779 Aug 2007 WO
PCTCA2007000840 Aug 2007 WO
PCTCA2005000716 Nov 2007 WO
PCTCA2005001930 Nov 2007 WO
PCTCA2006000655 Nov 2007 WO
PCTCA2006000751 Nov 2007 WO
PCTCA2007001297 Nov 2007 WO
PCTCA2007001298 Nov 2007 WO
PCTCA2007001658 Jan 2008 WO
PCTCA2007001749 Jan 2008 WO
WO2008009134 Jan 2008 WO
WO2008034232 Mar 2008 WO
WO2008040119 Apr 2008 WO
PCTCA2008000275 Oct 2009 WO
Related Publications (1)
Number Date Country
20080170660 A1 Jul 2008 US
Continuation in Parts (1)
Number Date Country
Parent 11431627 May 2006 US
Child 11747639 US