Certain embodiments generally relate to methods and systems for providing remote access to baggage scanned images and passenger security information.
In recent years there has been increasing interest in the use of imaging devices at airports to improve security. The President of the United States signed the Aviation and Transportation Security Act on Nov. 19, 2001, which, among other things, mandated that all checked luggage should be inspected by an explosives detection system (EDS). The Federal Aviation Administration (FAA), now the Transportation Safety Administration (TSA), a division of Homeland Security Administration (HAS), has set standards for qualifying explosives detection systems. To date all certified systems have represented computed tomography (CT) scanners and in one instance, a diffraction imaging (DI) scanners. Today thousands of CT scanners are installed at airports to scan checked baggage. The CT and DI scanners generate data sets that are used to form images representative of each scanned bag. The data sets are processes by an automated image recognition system, such as for certain patterns, characteristics and the like. When the image recognition system identifies a potential threat, the images are brought to the attention of an operator. The scanners are operated by TSA personnel who view cross sectional images of the baggage that is identified by the automated detection software to be a possible threat.
The scanners are capable of producing fully 3-dimensional (3-D) images. However, the software required to view such 3-D images is complex and generally requires sophisticated operators with expertise in 3-D rendering software tools. CT scanners are able to generate a 3-D voxel data set that represents the volume of the scanned bag. Conventionally, scanners provide 3-D images by stacking a series of closely spaced cross section images into a 3-D matrix. The 3-D image may then be viewed by an operator/screener. The operator usually steps through two-dimensional (2-D) slices (e.g., planes) of the 3-D matrix to detect and identify potential threats within the packed bag. Thus, the current procedure requires a highly trained operator with substantial expertise for the reliable detection of threats.
Currently, existing CT based explosive detection systems (EDS) are deployed at airports to detect various threats within packed bags. The suspicious bags are passed onto a human screener who examines individual CT slice images of the scanned bag. The CT slice images of alarmed bags are carefully examined by the human screener who then either accepts or redirects the bag for explosive trace detection (ETD) and/or manual unpacking for a visual inspection. This two step process allows approximately 250 bags per hour to be examined with a false-alarm rate of about 20-30%.
Unfortunately, the individual CT scanners will be operating on a stand-alone basis even though current and potentially future CT scanners may be on a network such as the Internet that allows access from the outside world. Without the appropriate tools, the screener, who sits next to one of the scanners, views the scanned CT images either to accept or reject the bag for further inspection. The overall false alarm rate will depend upon the ability and experience of the screener sitting next to the scanner, for a more experienced screener will have a lower false alarm rate than a less experienced screener. A screener operating in this stand-alone basis does not have the capability to receive advice or consult with a more experienced screener. Currently, one in five bags must be further inspected by carefully reviewing CT slice images.
Nondestructive testing techniques aim to detect certain features inside or outside of an object of interest to evaluate physical and mechanical characteristics of the object without harming the object. For instance, an ultrasonic pulse-echo technique is conventionally used to detect metal objects that are hidden inside of a package. However, the shape and position of potential explosive devices may vary and sometimes threat resolution may require a detailed knowledge of the chemical properties of the explosives and the physics of the packaging.
Further, the demands placed on scanner operators are further exaggerated by the time pressures of the application. The time pressures result from the need to examine baggage between the time that the baggage is checked and loaded on a flight. Often travelers check-in only shortly before their scheduled departure time, thereby permitting little time for the scanner operator to view the baggage.
After the baggage is check-in, the baggage is scanned by a CT scanner and axial slices or images are created of the baggage. The operator/screener views the axial slices or images by scrolling through each image slice one by one to determine if any potential threats are present in an image. Scrolling through over dozens of images (or even more for future generation scanners) for each bag is a laborious task, and the operator/screener must be alert to detect features of any potential threats within an image in order to flag the possible threats. Examination of each axial slice image gives rise to operator/screener fatigue that eventually will lead to sub-optimal performance by the operator causing him/her to miss some threats. For example, a bag that simulates a piece of luggage containing explosive stimulants (e.g., two bars of soap) that are hidden inside a radio underneath one the speakers is scanned. After a bag is checked, a CT 3-D data set of a packed bag is obtained and may, for example, include hundreds of axial slice images. Of these images only a few images may show the potential threat. If the operator misses anyone of these few images, the undetected threats could result in disaster.
There is a need for an improved baggage scanning system and method, to electronically unpack a scanned bag that provides views of inside of a packed bag such as without having to physically unpack the bag, or without having to harm the bag.
In accordance with certain embodiments a method and system to analyze the content of a packed bag utilizing a scanner is provided. The bag is scanned for a scannable characteristic to acquire scan data representative of a content of the piece of baggage. A volumetric data set is generated from the scan data, wherein the volumetric data set includes voxel values of the scannable characteristic throughout a volume of interest in the baggage. A rendered view is produced of the content of the piece of baggage based on the voxel values within a selected range from the volumetric data set. The method and system also provide identifying a threat by determining a Hounsfield value of the material of interest having a value close to that of explosives.
An electronic unpacking process and system are provided that simulates physical unpacking of packed bags to visualize various objects within the bag, where the objects may represent threats. The electronic unpacking begins by generating a three-dimensional (3-D) data set of the packed bag utilizing, for example, the computed tomography (CT) scanners currently available. A screener or operator upon scrolling through slices of the 3-D data set would examine separate CT slices and decide whether an object that may be a threat is present or not. Typically, one piece of baggage is divided into hundreds of slices or images that are viewed for a complete screen.
Electronic unpacking utilizes the same CT data set to electronically unpack the same identical bag by performing a slice-by-slice processing. First, the 3-D computed tomography data may be interpolated to generate isotropic volume data. Just as there are many ways to unpack real physical bags, electronic unpacking offers a number of different visualization techniques/algorithms to visualize the objects within the packed bags. For instance, the bag may be electronically unpacked in 3-D using surface rendering (SR), volume rendering (VR), maximum intensity projection (MIP), or a combination thereof. The unpacking process may be used to visualize organic material (e.g., typical bombs) or metals (e.g., guns, knifes, and the like) for detection of certain threats while unpacking the bag.
Optionally, a 3-D threat detection algorithm may be provided to detect possible threats. The electronic unpacking and threat detection enhance both the threat detection accuracy and the image quality of the electronically unpacked bag. The bags that are automatically flagged by the threat detection algorithm are sent, along with images of the electronically unpacked bag clearly marked with threats, to a local screener who carefully inspects the images.
The electronic unpacking is superior to manual unpacking, in terms of bag throughput, as the electronic unpacking may be performed within the time that it takes to take the bag from the conveyor and to place the bag on a table for manual unpacking by a local screener. In addition, unlike manual unpacking, the electronic unpacking can identify whether various objects within the bag are innocuous or not, based on the Hounsfield Unit (HU). Moreover, electronic unpacking can determine the exact volume of all objects, both threats and innocuous, within the packed bag.
In accordance with certain embodiments, an electronic unpacking system seamlessly integrates with both the currently deployed and future generation computed tomography (CT) scanners as well as current and future generation explosives detection systems (EDS), while allowing a desired allocation of expert screening capabilities and remote monitoring of the inspection process itself. The electronic unpacking system integrates the EDS at all installations (e.g., airports, seaports, buses, trains, and the like) and all expert (and otherwise) screeners via a secure network. The system also integrates other orthogonal sensors (e.g., ETD) and the passenger information database. The system provides an expert-on-demand (EoD) service, through which, a local inexperienced screener at a particular EDS site, has instant access to remote screening experts located anywhere in the world. The system also allows real-time remote monitoring of the inspection process for each and every EDS site, enabling supervisors and government personnel to continually improve the inspection process. The monitoring capability may be used to log and store an inspection process that can be analyzed at a later date for various performance studies.
Each scanner device 8 includes a scanner source and detector that are capable of obtaining a volumetric (or a cross-sectional) scan of each item of interest, a controller module to control operation of the scanner device, a user interface to afford operator control, and a monitor to display images obtained by the scanner. For example, the scanner and detector may rotate about the baggage as the baggage is conveyed along a belt (e.g., to perform a helical scan). The scanner device 8 communicates bi-directionally with a local terminal/server 12 that is configured to, among other things, operate as a local server. The scanning device 8 scans objects of interest, such as baggage (e.g. luggage, backpacks, briefcases, purses, and the like) to obtain volumetric data set representative of every voxel within the object of interest. The scanning device 8 conveys the volumetric data set for each piece of baggage to the local terminal 12. The local terminal 12 (or a local workstation) is configured to perform electronic unpacking and/or threat detection in real-time as a bag is conveyed along the belt. The local terminal 12 captures scan data in real-time and stores the scan data in local memory as the 3D volumetric data set, such as on the hard drive of the local terminal 12. The local terminal 12 includes a monitor 14 to display the volumetric and 2D images in real-time as an object is passing through the scanner device 8. The local terminal 12 also includes a user interface 16 to provide an operator control over the local terminal 12 and scanner device 8. Optionally, a single local terminal 12 may be connected to one or more nearby scanner devices 8 that are located in close proximity to one another, so that each operator can have access to the console of the local terminal 12. Optionally, a local terminal may be a local workstation that may produce a rendered view or may be connected to multiple display terminals to show the rendered view. The rendered views are pre-sorted and stored as a sequence of images. In an embodiment, the workstation communicates with a plurality of other processors and local workstations over a high-speed connection to display the rendered view.
The volumetric data set is sent from the local terminal 12 over a private communications link, such as a local area network (LAN) 18, to an enterprise server 20. The transfer of the volumetric data set may be initiated independently by the local terminal 12 or under the command of the enterprise server 20. The scan data is conveyed to the enterprise server 20 substantially in real-time. The term “real-time” as used through out this document shall include the time period while the object being scanned is still within the scanner device 8, and shall also include a period of time immediately after the object exits the scanning device 8 while the object is still within the restricted access area. For example, “real-time” would include the time from when a bag is checked up, the time in which the bag is transported to the flight, the time in flight, and the time in which the bag is transported from the flight to the bag retrieval area at the destination airport. In the example of a government building, “real-time” would include the time from when the object first enters the building until the object is carried out of the building. In the example of a live event, “real-time” would include the time from when the object enters the event area (e.g., fair ground, stadium, etc.) up until the object leaves the event area.
The scan data is conveyed from the local terminals 12 to the ES 20 in one of several image formats (e.g., DICONDE, TIFF, JPEG, PDF, etc.). Each image file is assigned a header that identifies which scanner device 8 produced the image, the time of the scan, the passenger ID, and other data obtained at the point of scan. The image files are stored for 48 hours or more depending on the needs of the restricted access area.
The enterprise server 20 is connected, through a high-speed connection 24 (e.g., the internet, a private network, etc.), to multiple remote terminals 26 that may be used by experts in a manner explained hereafter. The enterprise server 20 and/or remote terminals 26 (as well as local terminals 12) are configured to perform electronic unpacking and threat detection in accordance with various techniques described thereafter. The enterprise server 20 performs numerous operations, such as responding to inquiries for specific scan data. The inquires may come from a remote terminal 26 over the high-speed connection 24 or from another local terminal 12 over the LAN 18. The enterprise server 20 obtains the requested data sets from memory, compresses and encrypts the data sets and sends the compressed data sets in a compressed, encrypted manner to the requesting remote terminal 26 or local terminal 12. The compressed data sets are conveyed with a standard internet transport protocols. By way of example, an enterprise server may service all of the local terminals 12 in a single large airport terminal building, or the entire airport for medium-size, and smaller airports. A larger airport such as Los Angeles International airport (LAX) or John F. Kennedy International airport (JFK) may have several enterprise servers 20, corresponding to the different terminal buildings at their respective locations. Various functions of ES can also be distributed to local terminals 12.
Upon review of the 3-D electronically unpacked bag images, the local screener 72 has the option to either accept the bag or request consultation with one or more remote screening experts 74 using the expert-on-demand (EoD) service (as shown in
The sequence begins by the local screener 72, using a local workstation to request Expert-on-Demand (EoD) service through a TeleInspection Server by using a local instance of TeleInspection Client (TIC). The TIC may have a list of preferred experts. The user clicks an EoD button (not shown) on the local screener's 72 TIC window to request 80 the expert-on-demand (EoD) service. The request 80 is received by the central TIS 78, which routes 82 the EoD request to the next available remote screening expert 74. The EoD request 80 is acknowledged 84 by the currently available remote expert 74, and the EoD request 80 is serviced by the first available remote screening expert 74. Upon completion of this sequence, a direct TLS link 70 from the screening expert's 74 TIC to the local screener's 72 TIC is established
The local screener 72, through the local screener TIC, and the expert screener 74, through the expert screener TIC, is able to view and manipulate the 3-D images of unpacked bags through a TeleInspection View Sharing (TVS) protocol built into the TIC. The communication between the local screener 72 and the remote expert 74 is similar to public domain messenger services such as MSN®, AOL® and Yahoo® messengers. However, the TeleInspection system is specially developed for security applications with careful considerations of airport inspection process and EDS requirements. Therefore, users or public messenger services are unable to establish a link with expert screeners 74.
The system architecture allows remote experts 74 to be off-site by using instances of the TIC anywhere in the world. A remote expert 74, thus, may be off-site with a laptop computer 86, at home with a PC 88 or on the road with a PDA 90. The TIC supports transmitting text, voice, video, white board, and the like. Various passenger information data, such as passenger itinerary, travel history, credit information, passenger profile, passport information, passenger photograph, family history, age, physical characteristics, job information and the like are also available for review to assist in the decision whether to accept or reject the suspect bag. Other screeners and experts (not shown) running instances of TICs, may be invited to join an on-going conference.
The existence of the TIS is transparent to all users of TICs. From the viewpoint of the local screener 72, the conference begins at the click of the EoD button (not shown) and the conference begins from the viewpoint of the remote expert 74 at the click of the Acknowledge button (not shown) on the expert's 74 TIC windows. Thus, one remote expert 74 is able to provide service to a number of different EDS sites. The number of false-positives at each EDS site is reduced by the availability of remote screening experts 74; for, the remote experts 74 appear to an outside observer as being locally stationed with each and every EDS site that utilizes a remote expert 74. The TeleInspection system improves the performance of each EDS by effectively sharing expert screeners 74 and allowing the expert screeners 74 to be located anywhere in the world.
As the scan data becomes available from the CT scanner 102, a 3-D electronic unpacking module 104 unpacks the 3-D bag electronically and generates inside views of the unpacked bag with clearly marked threats in 3-D. If there are no threats automatically detected, the bag is accepted 114. If the bag is not accepted, the flow continues along 113, because potential threats are displayed, the 3-D images with clearly marked threats are passed onto the local screener 72 who visually inspect the 3-D images. The local screener 72 can either accept the bag 114 or request assistance via the expert-on-demand (EoD) service of the TeleInspection system 106. The TeleInspection system 106 provides a secure communication link (shown in
Running parallel to the electronic unpacking process 104 (as shown in
Upon completion of the segmentation, the volume rendered 304 images are compared against the segmented regions for consistency, and the initial segmentation is modified in accordance. The rendered views are generated using a portion of the 3-D data to render a particular object (e.g., the threat) or objects within the packed bad, and to discard obstructing structures to clearly render the object of interest (e.g., the threat). Once the final segmentation is completed, the detected threat objects are automatically indicated with ellipsoids on the 3-D rendered images for the local screener. By combining volume rendering 304 with segmentation, the threats (e.g., explosive stimulants) are clearly visible and identified, and the via the consistency check the detection accuracy is improved and the false-alarm rate is lowered. The rendered views of the threats may be shared across a network.
Both the 3-D electronic unpacking 104 and threat detection are executed in real-time and will be completed by the time the data for the next bag becomes available. Alternatively, the workstation 101 can be augmented with staggered computer processing units (CPUs) for meeting throughput requirements for future improved threat detection algorithms that may require more processing requirements.
The user interface 402 may include scroll bar or buttons (e.g., up-button down-button, left-button, right-button, and the like) to facilitate selection of a region of interest in the object. Alternatively, the user can utilize a drawing function to trace, to sketch, or to outline around the area of interest. The user interface 402 may allow a user to toggle-on and toggle-off various portions of the display portions (e.g., 3-D rendering window or 2-D rendering window). If a display portion is not shown, the remaining portion may be re-sized and/or rotated. For example, the object displayed in any of the windows can be rotated about at least two axes, typically a vertical axis and one or both horizontal axes. The user interface 402 also allows the user to measure a plurality of distances and save each distance in memory. The distances may include a length, a diameter, a radius, and the like. The distances can be utilized to determine a volume of an object. Further, user interface 402 provides a variety of markers for the user to identify potential areas of interest, cut-out specific sections or cut-out specific portions of an object, and to identify threats.
The 3-D rendering window 404 provides a volumetric view of the object. By utilizing a rotation function provided by user interface 402, the volumetric view is rotatable in 360 degrees. The 2-D rendering window 406 provides a cross-sectional view of the object or of a region of interest as selected by the user via the user interface 402. The 3-D rendered window 404 and the 2-D rendered window 406 are related to one another and based on a common region of interest. For instance, both the 2-D rendering window 406 and the 3-D rendering window 404 provide interactive 2-D. Furthermore, both displays allow a selected portion of the object to be magnified and rotatable.
The cut-plane window 408 allows a variety of planes through the object to be displayed, such as horizontal, vertical, transverse, and other planes at various angles (e.g., saggital, coronal and axial planes). For instance, the user may select a singular planar section, two sections, four sections, six sections, or eight sections to be displayed in cut-plane window 408. Each plane is selectably rotatable in 360 degrees. Further, as the region of interest shown in the 3-D rendered window 404 and 2-D rendered window 406 is updated, changed, or reconfigured, the plane(s) shown in cut-plane window 408 are interactively updated.
The magnification window 410 displays a region of interest and allows the user to magnify a selected region of interest by depressing a button located on the user interface 402 (e.g. a mouse button, a touch screen button, a trackball button, and the like). Further, the user has the ability to select the amount of magnification ranging from approximately 0.1× to 100×.
{xp}p=0N
where Nx×Ny×Nz in equation (1) denotes the number of voxels available in the 3-D data. The CT scanner scans a piece of baggage for a scannable characteristic to acquire scan data representative of a content of the piece of baggage, wherein the scannable characteristic is an attenuation measurement. The CT scanner provides axial slices (or z-slices) with isotropic pixels, where xp=p Δx, and yq=q Δy, and Δx=Δy. However, as the slice spacing of axial cuts may be different from the pixel size and may actually vary even within a single 3-D data set, the arbitrary slice spacing, shown below in equation (2), may be assumed:
{zk}k=0N
Thus, assuming the data to be corrupted by additive noise, the equation (3) is a model of the measured 3-D data set:
ƒ(xp, yq, zk)=μ(xp, yq, zk)+n(xp, yq, zk) (3)
The 3-D non-isotropic data set must be interpolated across the z-axis 312 to generate an isotropic 3-D data set. The exact interpolation may be performed on the 3-D data using Shannon's interpolation formula (i.e., also known as Whittaker-Shannon interpolation formula). The interpolation may also be implemented using fast Fourier transforms (FFT) with zero padding if the slice spacing does not change within the 3-D data set; however, the size of the required FFT may be too large to be implemented on a general purpose computer. A general solution for the arbitrary slice spacing given in equation (2) must utilize Shannon's original interpolation formula or at least the sub-optimal approximation to Shannon's interpolation formula. However, conventionally the following linear interpolation 321 shown in equation (4) is used:
Although the above technique is a linear interpolation across z-slices 312, any other interpolation method may also be used. The isotropic volume data of the specimen that can be visualized in 3-D results upon interpolation of the 3-D CT data. Thus, a volumetric data set is generated from the scan data, where the volumetric data set includes voxel values that are in Hounsfield units, for the scannable characteristic throughout a volume of interest in the piece of baggage. A portion of the volumetric data set is segmented based on the voxel values to identify an object and provide a visual marker outlining the object. The voxel values are segmented into categories. The categories are selected by the Hounsfield unit value of the voxel, and the categories being an innocuous material, an organic material and a metallic material.
At 202, depending on the type of bag and/or the screener's 72 preference, a visualization technique is selected. A variety of visualization techniques may be utilized to render the isotropic volume data in 3-D. Furthermore, the visualization technique may be selected automatically depending on the size of the bag and the density distribution of objects inside the bag. Depending on which visualization technique selected by the screener 72, the electronic unpacking renders the scanned bag data using surface rendering (SR) 303, volume rendering (VR) 304, or maximum intensity projection (MIP) 305. Alternatively, minimum intensity projection (MinIP), multi-planar reformatting (MPR), or radiographic projection (RP) may also be utilized in place of MIP to visualize the 3-D results. The selected visualization technique produces a rendered view of the content of the piece of baggage based on voxel values within a selected range from the volumetric data set. The rendered view is produced from voxel values that lie within a user selected range of thresholds of a selectable range of voxel values wherein the user has the ability to interactively adjust the selectable range.
Surface Rendering (SR) 203 is a visualization that is utilized to unpack a bag to show the exterior surface of objects within the packed bag. But, before the actual visualization step occurs, SR 203 requires the volume data to be preprocessed, i.e., a surface extraction is performed to determine the exterior surface of an object. In general, surface boundary voxels are determined using a threshold value from the isotropic volume data. Then a marching cube algorithm or a marching voxel algorithm is applied to the surface boundary voxels, which provides the surface data of the object.
After surface extraction, the object is rendered in 3-D using a light source 332 and a reflection 334 resulting from the light source 332. There are three types of reflection, i.e., diffuse, specular, and ambient reflections.
An object is represented by the sum of the diffuse, specular, and ambient reflections. If the unit vectors on the surface points 338 of the object are denoted to correspond to the viewing plane points (X,Y) 328 as {right arrow over (L)}(X,Y), {right arrow over (N)}(X,Y), {right arrow over (R)}(X,Y) and {right arrow over (V)}(X,Y). The resulting surface-rendered image PSR(X,Y) 340 can be expressed by equation (5) as:
PSR(X,Y)=I{Kdiff({right arrow over (L)}(X,Y)·{right arrow over (N)}(X,Y)+Kspec({right arrow over (V)}(X,Y)·{right arrow over (R)}(X,Y))n}+A (5)
As shown in equation (5), Kdiff and Kspec are the coefficients that control the ratio of diffuse and specular reflections, and I is the intensity of the light source from a specific direction {right arrow over (L)}. The constant A denotes the ambient reflection by the ambient light source in a specific space. The inner product values are in the range between zero and unity.
SR 303 handles only surface data of the object after surface extraction, so the surface rendering speed is higher than the other visualization techniques. SR 303 is very good for various texture effects, but SR 303 needs preprocessing of volume data such as surface extraction. Therefore, SR 303 is not suitable for thin and detailed objects as well as that object that have a transparent effect.
On the other hand, Volume Rendering (VR) 304 is a visualization technique that uses volume data directly without preprocessing of the volume data, such as required by surface extraction in SR 303. A characteristic aspect of VR 304 is opacity 342 and color 344 (shown in
While the rays traverse the voxels, the pixel values are computed from the opacities 342 and colors 344. The composite ray-tracing algorithm used in VR 304 is based on the theory of physics of light transportation in the case of neglecting scattering and frequency effects. Given an emission q and an absorption κ along a ray s, an intensity I can be computed from the following equation (6):
that can be discretized as show in equation (7) below:
VR 304 is known as a superb method for thin and detailed objects and suitable for good transparent effects. But VR 304 uses whole volume data, so the volume rendering speed is relatively low due to the expensive computation.
Similar to VR 304, the resampling of voxel intensities at the new voxel locations according to the viewing direction is also needed before ray-tracing can be performed. The MIP 305 process can be expressed shown in equation (9):
The viewing direction 326 is in the direction of the Z-axis 324 and {circumflex over (ƒ)}INT(X,Y,Z) 321 is the resampled voxel intensity at the new voxel location in the new coordinate (X,Y,Z) 328.
MIP 305 is a good visualization method, especially for vessel structures. But MIP 305 discards the information of depth while transforming 3-D data to 2-D data that results in an ambiguity of geometry of an object in the MIP 305 images. But, this ambiguity can be solved, for example, by showing MIP 305 images at several different angles in a short sequential movie.
At 306, if the rendered image is not sufficiently detailed enough to make the decision whether to accept or reject the bag, the unpacking process is repeated with a new set of rendering parameters. The new rendering parameters may include, for example, the opacity 342 and coloring 344 scheme for volume rendering 304 and lighting conditions for surface rendering 303, a new rendering region, orientation of the viewpoint, and the like. For instance, the rendered image as displayed to a local screener may simultaneously co-display at least one of a surface and volume rendered view of the content of the piece of baggage, and an enlarged image of a region of interest from the rendered view. The user may zoom on a region of interest within the rendered view or may rotate the rendered view to display the content of the piece of baggage from a new viewpoint.
At 307, the screener 72 must decide whether one or more threat objects exist within the packed bag. In an embodiment, an automatic threat detection analysis of at least a portion of the volumetric data set based on the scannable characteristic is performed. Based on screener's 72 experience, perhaps aided by a remote expert 74 through the Expert-on-Demand service of TeleInspection 78 (shown in
In the above examples, the scanners are described in connection with CT and DI scanners and the data sets are described in connection with attenuation measurement data. For instance the scanners may include a cine computed tomography scanner, a helical CT scanner, a dual-energy x-ray scanner, dual-energy CT scanner, and a four-dimensional (4-D) cine computed tomography scan. However, alternatively other types of scanners and other types of data may be obtained, processed and displayed without departing from the meets and bounds of the present invention. For example, the scanner may represent an electron beam scanner. Alternatively, the scanner may transmit and receive non-x-ray forms of energy, such as electromagnetic waves, microwaves ultraviolet waves, ultrasound waves, radio frequency waves and the like. Similarly, in the above described embodiments, the data set is representative of attenuation measurements taken at various detector positions and projection angles, while the object is stationary within the scanner or while the object is continuously moving through the scanner (e.g., helical or spiral scanning). Alternatively, when non-x-ray forms of energy are used, the data set may represent non-attenuation characteristics of the object. For example, the data may represent an energy response or signature associated with the object and/or the content of the object, wherein different types of objects may exhibit unique energy responses or signatures. For example, explosives, biological agents, and other potentially threatening medium, may exhibit unique electromagnetic responses when exposed to certain fields, waves, pulse sequences and the like. The electromagnetic response of the object and the content of the object are recorded by the scanner as scan data. As a further example, the scanner may be used to obtain finger prints from the object. The finger prints would be recorded as scan data.
The modules discussed above in connection with various embodiments are illustrated conceptually as a collection of modules, but may be implemented utilizing any combination of dedicated hardware boards, digital signal processors (DSPs) and processors. Alternatively, the modules may be implemented utilizing an off-the-shelf PC with a single processor or multiple processors, with the functional operations distributed between the processors. As a further option, the modules may be implemented utilizing a hybrid configuration in which certain modular functions are performed utilizing dedicated hardware, while the remaining modular functions are performed utilizing an off-the shelf PC and the like.
While the invention has been described in terms of various specific embodiments, those skilled in the art will recognize that the invention can be practiced with modification within the spirit and scope of the claims.
This application claims priority to U.S. Provisional Patent Application No. 60/779,133, filed Mar. 3, 2006, the complete subject matter from all of which is expressly incorporated herein by reference in its entirety.
Number | Date | Country | |
---|---|---|---|
60779133 | Mar 2006 | US |