Non-Destructive Evaluation (NDE) has been one of the engineering disciplines which has mostly revolutionized diagnostic techniques in industry and in medicine during the last decades. MR (Magnetic Resonance), CT (Computerized Tomography), US (Ultra-Sound), and other NDE devices are being standard tools for a wide range of diagnostic fields. Furthermore they are currently changing significantly medical surgery, as their capability of visualizing intra-body structures enables surgeons to minimize the invasiveness of their interventions. Even though NDE techniques are by themselves expensive, they are potentially even interesting from a financial point of view, as they can significantly decrease the expensive hospitalization time of patients. Similar arguments apply to industrial NDE which was able to bring quality assurance and failure prediction to an impressive performance.
Information coming from one NDE device is always restricted to a particular characteristic of the intra-body physics. For example, the information that can be drawn from MR devices or CT scanners is related but still very different. While medical MR devices image the soft tissues, CT scanners are unable to discriminate different types of soft tissues while giving information about the skeleton structure of the patient. In practice, both imaging systems are therefore used in conjunction when necessary, in order to exploit the complementary information provided by both devices. The result is multimodal information, i.e. information which can practically not be provided by one single system.
As used herein, “multimodal” means use of at least two imaging modes which differ by the physical characteristics of the scene they image during the data acquisition process.
One of nowadays main research directions consists of determining new imaging modalities which are able to measure more intra-body characteristics. One of these characteristics is the intra-body temperature of industrial specimen or patients. Temperature is a highly discriminative characteristic for e.g. cancer tissues or design errors in electrical circuits. Several approaches to temperature imaging have been proposed, but just few have been able to reach any real market value.
Devices enabling data acquisition of thermal maps comprise MR scanners (intra-body), infrared devices (surface), and passive microwave imaging systems (intra-body). Thermal maps are considered to give pertinent information. In the case of medical imaging, the temperature maps acquired by the imaging device can give information about the intra-body temperature distribution during thermal ablation of e.g. a tumour, or about the presence, respectively absence, of cancerous tissues in human bodies. In the context of industrial applications, thermal maps are particularly valuable for instance for the prediction and prevention of electrical equipment circuit failures.
But temperature by itself has just a restricted practical value if not combined with other imaging modalities such as US, CT, or MR data, in order not only to give intra-body temperature maps, but also to relate them to the underlying physical structure of the patient or specimen under study. For example, increased intra-body temperature is not sufficient to conclude on the presence of cancerous tissue due to the inhomogeneity of biological tissue. It could also come from a local inflammation of the biological tissue. Underlying anatomical data reflecting this inhomogeneity can therefore increase significantly the diagnostic specificity of the acquired temperature maps.
Thermal maps are thus medically and industrially most useful in conjunction with anatomical or structural images in order to provide the medical doctors and investigators with information complementary to the thermal data. In the case of cancer detection, the anatomical images acquired by e.g. US or CT devices ensure that not only cancerous tissues are detected, but also they define their exact anatomical location, enabling the medical doctors to optimize their treatment plans. In the context of electrical circuits, the temperature maps get expressive when being related to the underlying circuit design.
Presently, different imaging modalities, such as thermal maps and anatomical information, could only be provided sequentially, but not simultaneously. As a result, the following disadvantages of existing approaches to thermal imaging can be listed:
First, anatomical and structural information is a fundamental prerequisite for medical doctors and investigators during medical diagnosis or industrial quality testing in order to locate sensitive anatomical structures, e.g. during thermal ablation, or to relate failure sensitive regions of an integrated circuit to the underlying circuit design which can be improved thereafter. Therefore, having access to mere temperature maps without having simultaneously the related structural information is a highly limiting factor, as temperature maps by themselves do not provide any information about the underlying structure.
Second, in the case of medical applications, the medical doctors are used to inspect anatomical data, but do not have much practical experience in interpreting thermal maps due to the novelty of thermal imaging devices. Therefore an appropriate combination of thermal maps with conventional anatomical imaging modalities would heavily facilitate the medical doctors' tasks of interpreting the acquired thermal maps.
Third, if thermal and structural information are acquired in a sequential manner, it is required to physically fix the patient or the investigated industrial piece during and between the successive imaging processes and to exactly determine the position of target with respect to the imaging devices in order to be able to fuse and visualize the multimodal data sets without mutual displacements.
Fourth, even if both data from the same body portion could be acquired simultaneously by two different imaging devices, a pure superimposition of the resulting images would not be optimal, as redundant or irrelevant information would get displayed as well.
Fifth, the only device currently being in practical use for the non-invasive determination of intra-body temperatures is MR. This is a very large and expensive device not being easily accessible by the majority of medical doctors.
The present invention solves the problems described above by providing means to optimally fuse and visualize the data coming from individual imaging devices.
An object of the invention is a virtual multimodal non-invasive imaging device comprising:
As used herein, “spatial registration” refers to bringing two datasets into spatial correspondence, i.e. for every position in one dataset the corresponding position in the second dataset is known.
Thus, the processing system which receives the data streams from the respective imaging devices performs signal conditioning, image registration, and data fusion, in order to provide the information from the plurality of input data streams in a virtually fused single data stream.
Preferably, said processor extracts corresponding image features of said thermal map and of said structural map for the registration process. The resulting digital data stream guarantees a spatial correspondence of the input data streams from the different imaging devices.
Preferably, said processor extracts complementary image features of said thermal map and of said structural map for the data fusion step. Thereby, the system removes redundant information from the different data streams in order to guarantee the visualization of both a complete and very compact representation of all the information coming from the different input data streams.
Preferably, the thermal data of said multimodal data set are color coded and overlaid to the structural data. The system visualizes the resulting single data stream on a computer screen.
Preferably, configuration information is streamed back from said processing system to said imaging subsystems.
The system has several inputs for the different imaging devices, one of which provides the thermal data, the other(s) the complementary structural information. Data streams from the thermal imaging device and one, or more than one, structural imaging device are streamed over standard networking and data streaming connections, such as USB2, IEEE1394, Ethernet, S-Video, or others, towards the processing system. In the case of analogue streaming connections, such as S-Video, a state-of-the-art analogue-to-digital converter or frame-grabber digitizes the analogue data before passing the resulting digital representation to the processor.
In order to bring the data into spatial correspondence, the imaging devices can be linked mechanically and calibrated to pre-determine the transformation which brings the data into correspondence. The resulting transformation can thereafter be applied in real-time. As an alternative, a software tool performs this spatial registration task by maximizing an information theoretic measure between image features with respect to possible image rotations and displacements. Further, the software will extract the most complementary and pertinent image features and characteristics of the different image modalities before fusing the resulting data into one multi-modal data set. As a final step, the resulting image representing the multimodal information is being visualized on a computer screen.
As used herein “real-time” capability of the final system refers to a real-time capability which does not significantly differ from the real-time capabilities of the individual imaging subsystems which are connected to the core device.
As used herein, “information theoretic measures” refer to all functionals constructed from information theoretic statistics, such as entropy, joint entropy, mutual information etc.
The system streams two data streams to the signal processor, e.g. the 2D thermal data from a MW thermal imaging device, noted T(x,y), and the 2D structural data from a structural imaging device, noted S(x′,y′), where x and y, resp. x′ and y′, parameterize the discrete imaging space of the thermal MW data T, resp. the structural data S. A mapping mα: (x′,y′)−>(x,y) provides a spatial correspondence between both, the imaging space of T and the imaging space of S, with α being the registration parameters for rigid or non-rigid registration, e.g. α parameterizes translations and rotations of the 2D structural data frames in 3D to register them with the thermal map. Analoguously, the mapping might be from the temperature imaging space to the structural data imaging space: mα′: (x,y)−>(x′,y′). Then, α′ parameterizes translations and rotations of the 2D MW data frames. The mapping parameters α and image features which allow the determination of α have to be determined simultaneously. Mathematically, the feature extraction can be formalized as an image mapping, kβ: T(x,y)−>kβ(T(x,y)), resp. lγ: S(x′,y′)−>lγ(S(x′,y′)), with β, resp. γ, being the feature extraction parameters. The aim of using image features instead of the raw image data for the registration process reflects the fact that not all information contained in multimodal data is pertinent for the registration process. Some image characteristics solely present in one of the modalities, such as imaging noise, cannot give any reliable input to the registration process, but rather decrease reliability of the algorithm. Therefore the feature extraction block within the multimodal registration process detects the image features most pertinent for the spatial correspondence of the input images and removes those not being of any use for this aim.
The optimization objective which allows simultaneous extraction of the most related image features and the determination of the optimal registration parameters α is written
e(FT,FS)=I(FT,FS)/H(FS,FT).
This functional is called feature efficiency. I(.,.) stands for mutual information, H(.,.) for joint entropy, FT is a random variable with a probability density pT estimated from the features of the thermal data T, and FS is a random variable with a probability density pS estimated from the features of the structural data S. Using histogramming, the following formulas are being employed to estimate the probability densities, pS, pT, and joint probability pS,T, for the random variables, FS, FT, and joint random variable FS,T, respectively. fS and fT are the features extracted from the data S and T, respectively:
where N is the number of features extracted from the datasets, and δa,b is the Kroenecker delta function, which is 1 if a=b and 0 otherwise. Other probability estimators might be used as well, such as Parzen-window probability estimation. This probability estimation step is important to the registration process of
Using these definitions, the exact expressions for mutual information and entropy can be given:
In
As an alternative to feature efficiency, normalized entropy might be used to drive the optimization process. It is defined by:
NE(FT,FS)=((H(FT)+H(FS))/H(FS,FT)).
In the case of feature efficiency, the resulting data registration process can be formalized as
(αopt,βopt,γopt)=maxα,β,γe(Fkβ(T(max(x,y))),Flγ(S(x′,y′))).
This process is outlined in
In contrast to the first data feature extraction related to image registration as described in the previous paragraph, the second feature extraction for data fusion aims to extract the features of the initial datasets which are most complementary to each other while removing redundant information from the datasets. Mathematically, the feature extraction process is written the same way as in the previous paragraph, even though its implementation might differ. Therefore, the feature extraction from the input thermal and structural maps is represented by a mapping oδ: T(mα(x,y))−>o67 (T(mα(x,y))), resp. uε: S(x′,y′)−>uε(S(x′,y′)). δ, resp. ε, represent again feature extraction parameters of the initial MW thermal map, T(x,y), resp. of the structural data device, S(x′,y′). The extraction process is driven by the minimization of the same optimization function as in the previous paragraph. Mathematically this can be written
(δopt,εopt)=minδ,εe(Foδ(T(m(x,y))),Fuε(S(x′,y′))).
For the optimization process, again any adapted algorithm can be employed.
The fact that while for registration the optimization objective has to be maximized, the optimization objective has to be minimized for data fusion, reflects the fact that data fusion aims to keep all available information of the input data and to remove the redundant information. In the case of registration, it is just this redundant information, i.e. the information that is present in both input datasets, that is able to drive the registration process towards the optimal spatial correspondence.
The resulting datasets, oδ(T(mα(x,y))) and uε(S(x′,y′)), are fused thereafter. The final data fusion is a fundamental step with respect to the general design of the system, as it is thanks to the fusion result that the medical doctor or industrial investigator has the impression of being working with only one single physical system. For this aim, the thermal map data, oδ(T(mα(x,y))), resulting from the previously described signal processing steps, are getting color coded.
The resulting color mapped thermal data can be overlaid on the structural data, uε(S(x′,y′)), resulting in a virtually augmented image. This process is the implementation of the so called “fusion rule” of
Further features and advantages of the invention will appear to those skilled in the art by means of the following description of a particular embodiment in connection with the drawings.
If the imaging subsystems have been designed in such a way, configuration information is getting streamed back from the processing system to the imaging subsystems in order to provide the investigator with the impression of being interacting directly with the subsystems. The data from the input devices are getting processed and fused inside the processing system before getting visualized on a computer screen.
Performing the different image processing steps shown in
In
There are two imaging subsystems connected to the processing system implemented on a state-of-the-art personal computer (PC): On the one hand, a microwave (MW) imaging device which provides 2D intra-body temperature maps of the patient, and on the other hand an ultrasound (US) imaging system which images 2D slices of the patient's anatomy and the needle positions. As schematically shown in
The Personal Computer (PC) 4 is provided with following components:
Using commercially available drivers and C/C++ application interfaces (APIs), the specifications of which may be found in ref. [1] and ref. [6], the processing system receives in real-time the respective datasets from the connected subsystems. The APIs are called from individual C-threads on the PC 4 which are implemented in order to receive the data from the individual imaging subsystems individually, but in a synchronized fashion.
The employed US subsystem is the commercially available Echoblaster 128 produced by Telemed, Lithonia (see ref. [1]). The two dimensional US frames are streamed over the USB2 connection 3 from the US beamformer 2 to the PC 4, and changes in configuration are streamed back from the PC 4 to the beamformer 2. Furthermore, Telemed provides the hardware drivers and C++ application interface (API) for Windows, based on the DirectX technology from Microsoft. A variety of US transducers 1 is also proposed by Telemed, enabling any programmer to easily implement a fully functional PC based US device.
The brightness temperatures are directly related to the real intra-body temperatures at the different locations. In order to reconstruct a two dimensional grid of real intra-body temperatures from the brightness temperatures, the algorithm disclosed by Jacobsen and Stauffer in ref. [4] is getting applied to the output grids from the Orsys analogue-to-digital converter 9. In fact, as the algorithm of ref. [4] reconstructs 1D temperature profiles, it is applied consecutively to the brightness temperatures of the individual antennas in the antenna array 7. The combination of the reconstructed 1D temperature profiles results in a 2D temperature map. The reconstruction algorithm from ref. [4] is implemented on the embedded Compact C6713 system 10, sold by Orsys (see ref. [5]). The analogue-to-digital converter 9 from Orsys can actually be plugged directly on the microbus of the Compact C6713 embedded system 10. Consecutive frames of 2D temperature maps are streamed over the firewire connection 5 of the Compact C6713 system to the firewire connector of the PC 4, while system configuration parameters are streamed back from the PC 4 to the Compact C6713 embedded device. On the PC 4, the firewire drivers and C-APIs from Unibrain (see ref. [6]) provide the programmer with an easy to use tool to implement a completely functional temperature monitoring imaging device.
In order to overlay the temperature maps from the MW device on the US frames with a guaranteed spatial correspondence, the US transducer 1 is physically linked to the MW antenna array 7. This guarantees that both datasets are acquired within the same imaging plane and that the imaged regions of both devices overlap significantly. The link between the two transducers is outlined in
The registration process which enables to compensate for the rotational offset of a between the US and MW scans is outlined in
In fact both, the rotational angle α and image features which allow the determination of α have to be determined simultaneously. This is because by nature of the two imaging modalities, the raw data does not contain corresponding information which allows direct mutual registration. Rather, the data features which represent pertinent information for the determination of the rotational angle α have to be determined. As disclosed in this invention, the determination of the registration parameter α and of the features pertinent to the determination of α are done simultaneously.
The feature extraction step can have a variety of specific implementations, e.g. considering prior information about the features to be extracted, discretely or continuously parameterized features, etc. With respect to the specific implementation described in this example, the feature extraction parameters β and γ represent a particular scale of the scale space image decomposition described by Lindenberg in ref. [7]. The feature extraction block within the multimodal registration process detects the image features most pertinent for the determination of the spatial correspondence between the input images and removes those not being of any use. The fact that in this particular implementation the imaging features are restricted to a specific scale of the scale space decomposition of the initial datasets reflects a known prior information about which image features will result in good registration. Still, the exact scale is not being fixed from the beginning, as the best scale in the scale space decomposition is not guaranteed to remain constant. Rather, it might change with changing parameters in the image acquisition process, such as e.g. a changing frequency used for the US image acquisition, since Telemed provides multi-frequency US transducers.
The optimization objective which allows simultaneous extraction of the best scale of the scale space image decomposition, i.e. the best features for the spatial registration, and the determination of the optimal registration angle α is called feature efficiency and is written
e(FT,FS)=(I(FT,FS)/H(FS,FT))
as indicated above. FT is estimated from the thermal data T, and FS is estimated from the structural data S, according to the teaching of Thomas A. Cover in ref. [8]. Histogramming is being employed as the probability estimator for FT, FS, and the joint random variable FT,S as described by T. Butz in ref. [9]. This means that the following formulas are being employed to estimate the probability densities, pS, pT, and pS,T, for the random variables FS, FT, and FS,T respectively:
where N is the number of features in the datasets. δa,b is the Kroenecker delta function, which is 1 if a=b and 0 otherwise. Other probability estimators might be used as well, such as Parzen-window probability estimation. The probability estimation step is important to the registration process of
The action of adapting registration parameters and image features refers to adapting the parameters α, β, and γ. For maximization, an optimization algorithm such as Powell (see ref. [10]) or genetic optimization (see ref. [11]) can be used.
In contrast to the data feature extraction related to image registration as described in the previous paragraph, the feature extraction for data fusion aims to extract the features of the initial datasets which are most complementary to each other while removing redundant information from the datasets. The feature extraction from the input thermal maps, resp. the US data, is represented by a mapping oδ: T(mα(x,y))−>oδ(T(mα(x,y))), resp. uε: S(x′,y′)−>uε(S(x′,y′)). δ, resp. ε, represent again a particular scale of the scale space decomposition according to ref. [7] of the initial MW thermal maps, T(x,y), resp. the anatomical data of the US device, S(x′,y′). The extraction process is driven by the minimization of the same optimization functional e as in the previous paragraph.
For the optimization process again, an algorithm, such as Powell or genetic optimization, can be employed. The resulting datasets, oδ(T(mα(x,y))) and uε(S(x′,y′)), will be fused thereafter.
For the data fusion, the thermal maps, o67 (T(mα(x,y))), resulting from the previously described signal processing steps, are color coded so as to reflect a natural interpretation, e.g. hot spots being red and cold spots being blue. When a thermal data, t(x,y)=o67 (T(mα(x,y)), is initially represented by a scalar within [tmin,tmax], such a thermal mapping is being performed by the following color coding equation:
The resulting color mapped thermal data can be semi-transparently overlaid on the structural data, uε(S(x′,y′)), resulting in a virtually augmented image. In this example, the software VTK (Visualization Tool Kit) (see ref. [12]) is employed as the data fusion and visualization package.
The software Qt available from Trolltech (see ref. [13]) is being employed for the implementation of the graphical user interface, which is designed in a way that the doctor can interact directly over the USB2, resp. Firewire, connection to the US device, resp. MW device, with the imaging subsystems. Thus, performing the described different image processing steps continuously as the data arrive over the input networking connections, results in a virtual multi-modal imaging device. The doctor can therefore interpret the acquired multimodal datasets simultaneously and interact with the MW and US subsystems, as if they were just one single multi-modal device which generates just one single data stream.
The invention as described herein above is capable to provide the industrial investigators and medical doctors simultaneously and in real-time, or off-line, with both structural data on the one hand and thermal maps on the other hand. Furthermore, as the information from different imaging modalities is combined by the multimodal signal processor, also the pertinent and complementary information from the different modalities is combined, resulting in a virtually augmented single system with increased value versus the individual subsystems.
The open standard technology of the data stream connections between the individual data acquisition systems and the processor enables removal and addition of imaging subsystems according to need and comfort of the industrial investigators or medical doctors. This applies to 1D, 2D, 3D or mixed data and data sequences.
Number | Date | Country | Kind |
---|---|---|---|
05405026.5 | Jan 2005 | EP | regional |
Filing Document | Filing Date | Country | Kind | 371c Date |
---|---|---|---|---|
PCT/CH2006/000033 | 1/13/2006 | WO | 00 | 7/17/2007 |