This invention relates to methods and apparatus for analysis of images, and is especially related to analysis and classification of image textures.
Certain visual characteristics of regions in images, relating to the regularity, coarseness or smoothness of the intensity/colour patterns are commonly referred to as texture properties. Texture properties are important to human perception and recognition of objects. They are also applicable for various tasks in machine vision, for example for automated visual inspection or remote sensing, such as analysing satellite images.
Texture analysis usually involves extraction of characteristic texture features from images or regions, which can be later used for image matching, region classification, etc.
Many existing approaches to texture analysis can be classified into one of three broad classes: i) structural approaches, ii) statistical approaches and iii) spectral approaches.
In a structural approach, the texture is characterised by features and spatial arrangements of certain visual primitives, such as blobs, line segments, corners, etc.
In a statistical approach, the texture is characterised by statistical distribution of intensity values within a region of interest.
In a spectral approach, a set of filters with varying properties is used, and their response to the underlying image is used as feature vector. For example Gabor filters with varying directional and frequency responses can be used. (See D. Dunn, W. Higgins, and J. Wakeley, “Texture segmentation using 2-D Gabor elementary functions”, IEEE Trans. Pattern Anal. And Machine Intell., vol. 16, no. 2, February 1994)
These known methods generally operate in the image domain, usually defined as a two-dimensional (2-D) lattice.
It is also known than an image can be mapped into a one-dimensional (1-D) representation using a mapping function, for example a plane-filling curve such as a Peano curve or a Hilbert curve. (See Peano G. “Sur une courbe que remplit toute une aire plane”, Math Annln., 36, pp. 157-160 (1590), and D. Hilbert, “Uber die stetige Abbildung einer Linie auf ein Flachenstuck”, Math. Annln, 38, pp. 459-460 (1891).) Subsequently the properties of the 1-D signal could be analysed, for example by Fourier analysis, to determine the texture features of the image.
The majority of the existing approaches are computationally intensive.
It would be desirable to provide a method and apparatus for texture description, classification and/or matching which is invariant to changes in image intensity, region translation and rotation, the method being computationally simple.
Aspects of the present invention are set out in the accompanying claims.
In accordance with a further aspect of the invention, a one-dimensional representation of an image is statistically analysed to derive a feature vector representative of the image.
The analysis preferably involves comparison of the one-dimensional representation with at least one threshold, and may be arranged to determine any one or more of the following:
(a) the rate at which the representation crosses a threshold;
(b) the average slope of the representation at points where a threshold is crossed. The points could be selected to be those where the representation values are increasing (to obtain the average “upslope”), or those where the representation values are decreasing (to obtain the average “downslope”). Alternatively, both the average upslope and the average downslope could be determined, or simply the average slope at all the points; and
(c) the average interval for which the representation remains above (or below) a threshold.
It has been found that the above parameters, which can be obtained in a computationally simple manner, provide a good discriminant for the many image classes, and particularly image textures. A combination of parameters (a) and (b) has been found to be especially effective.
A preferred embodiment of a method according to the present invention comprises the following steps:
This feature vector may relate to only a part of the image represented by a part of the one-dimensional target function. Further feature vectors can also be derived for other parts of the image. In the preferred embodiment, successive overlapping segments of the one-dimensional function are analysed to derive respective feature vectors. It is, however, not essential that the segments be overlapping.
In the preferred embodiment, each of the statistical characteristics is determined by comparing the one-dimensional representation with a threshold level. The threshold may be different for different characteristics, or may be the same for at least some of those characteristics. It is also possible to replace a simple fixed-value threshold by a varying threshold (the term “threshold function” being used herein to refer both to a varying threshold and to a fixed-value threshold, wherein the function is a predetermined constant).
In an enhancement of the invention, better discrimination is achieved by separately determining the rate at which the target function crosses respective different threshold functions. Preferably, for at least one threshold function, two separate values for the average slope of the target function at the crossing points are derived, one value representing the slope of the function when the function is increasing (the “upslope”), and the other value representing the slope of the function when the function is decreasing (the “downslope”).
Statistical characteristics other than averages may be used for deriving any of the values used to construct the feature vector, such as means, medians or variances.
Although the invention is primarily described in the context of analysing texture represented by the grey levels of an image, the texture could additionally or alternatively be represented by other characteristics, such as colour.
Arrangements embodying the present invention will now be described by way of example with reference to the accompanying drawings, in which:
An input image mapper (IIM) 100 employs the so-called Peano scanning to represent grey-level values of the two-dimensional (2-D) input image received at input 210 by a one-dimensional (1-D) function produced at output 212, referred to as the target function.
A scale-invariant transformer (SIT) 101 uses a suitable logarithmic transformation to convert the target function at output 212 of the IIM 100 into a target-function representation at 214, with values independent of the dynamic range of the 2-D input image. The dynamic range of the input image may be affected by varying illumination conditions, changes in local sensitivity of an image sensor, etc.
A moving window selector (MWS) 102, driven by the signal at 214 from the scale-invariant transformer (SIT) 101, selects segments of the target function representation suitable for further processing. This is illustrated in
The output of the MWS 102 is applied in parallel to the signal inputs of a plurality of feature estimators, including a crossing rate estimator (CRE) 104, a crossing slope estimator (CSE) 105, and a sojourn interval estimator (STE) 106.
The control input of the crossing rate estimator (CRE) 104 is connected to a reference level generator (RLG) 103 to receive on line 204 a signal defining a suitable rate threshold function (in this embodiment a simple constant value) for setting a discriminating level used for feature extraction from the representation of the target function. Similarly, the crossing slope estimator (CSE) 105 and sojourn interval estimator (STE) 106 receive from reference level generator (RLG) 103 on lines 205 and 206 respectively signals defining a suitable slope threshold function and a suitable duration threshold function for setting the discriminating levels which those estimators use for feature extraction.
In the present embodiment, all three estimators receive signals which define a common, fixed-value discriminating level, shown at 401 in FIGS. 4 to 6. This may be chosen in different ways; the level 401 could represent the median of the values in the one-dimensional output of the transformer 214, or the median of the values in the current window. However, the discriminating levels of the estimators may alternatively differ from each other, and could be variable.
Referring to
Referring to
Referring to
The image texture classifier (ITC) 107 processes jointly feature data available at its inputs to perform texture classification of the 2-D input image. The procedure used for texture classification may be based on partitioning of the entire feature space into a specified number of regions that represent texture classes of interest.
The regions are non-overlapping, i.e.,
Si∩Sj=Ø, i,j=1,2, . . . ,M i≠j
and the partition of the entire feature space S is exhaustive, i.e.,
S1∪S2∪. . . ∪SM=S
An image analysis procedure, according to the present invention, produces numerical values from the three estimators, CRE, CSE and STE, available at the outputs 220, 221 and 222, respectively. In the 3D feature space, such a triplet can be viewed as a point which must fall into one of the regions S1, S2, . . . , SM. If the point falls into Sk, 1≦k≦M, then a decision is made that an image under test exhibits the texture belonging to class k of M texture classes.
Partitioning of the feature space S into M regions may be performed according to some optimisation criterion based on minimum cost, minimum probability of misclassification, etc. The required partitioning procedure is a standard operation carried out for various applications of statistical decision theory.
Referring again to
Also, referring to
In this embodiment, the six values provided by the crossing rate estimator 104, the crossing slope estimator 105 and the sojourn interval estimator 106 are used by the image texture classifier 107 to classify the image within a six-dimensional feature space.
In an alternative arrangement, the three values from the crossing rate estimator 104 and/or the two values from the crossing slope estimator 105, can be combined, for example by using various weighting coefficients, to form a single respective value.
It is anticipated that in many applications the one-dimensional function will occupy the time domain, for example when the function is derived from a repetitively scanned image as might occur in some video systems. A time interval would thus represent a segment of the image that would be scanned during this notional time period. Accordingly, the argument of the target function in this situation may be the distance from a selected point on the scanning curve, or may be the time elapsed from a selected reference time instant.
The example implementation is rather simple for the sake of description clarity. A large number of alternative implementations exist. Alternative implementation may be obtained by:
Although the invention has been described in the context of analysis of two-dimensional images, the techniques can be extended to analysis of multidimensional data, and in particular multidimensional images, by employing suitable space-filling curves. The image may be a conventional visual image, or may be an image in a non-visual part of the electromagnetic spectrum, or indeed may be in a different domain, such an ultrasound image.
Number | Date | Country | Kind |
---|---|---|---|
01309065.9 | Oct 2001 | EP | regional |
Filing Document | Filing Date | Country | Kind |
---|---|---|---|
PCT/GB02/04338 | 9/25/2002 | WO |