Method and device for measuring similarity between images

Information

  • Patent Grant
  • 7308140
  • Patent Number
    7,308,140
  • Date Filed
    Monday, April 2, 2001
    23 years ago
  • Date Issued
    Tuesday, December 11, 2007
    16 years ago
Abstract
A method of measuring a similarity between images is provided. The method includes the steps of (a) computing a statistical dissimilarity between the images; (b) computing a perceptual dissimilarity between the images; and (c) computing a dissimilarity between the texture features of the images based on the statistical dissimilarity and the perceptual dissimilarity. When images perceptually similar to a query image are searched and retrieved according to the above measuring method, the possibility that the retrieved images are not perceptually similar to the query image is small. Accordingly, the image searching performance can be improved when searching images having texture features similar to that of a query image.
Description
BACKGROUND OF THE INVENTION

1. Field of the Invention


The present invention relates to a device for measuring the similarity between images and a method used by the same, and more particularly, to a method and device for accurately measuring the similarity between images.


2. Description of the Related Art


For applications of image search and retrieval, the characterization of the statistical attribute of textures is widely used. Texture descriptors are obtained according to such characterization. A texture descriptor is composed of two components: a similarity retrieval component (SRC) based on statistics and a perceptual browsing component (PBC) which is developed for describing the perceptual attributes of textures, such as the directionality and the texture's regularity. Such a texture descriptor is used for browsing and classifying images. The browsing and classifying of images are based on a measure of a distance. The measure of a distance is defined based on a feature vector and provides the similarity/dissimilarity between images.


A conventional similarity measuring method uses a statistical-based texture descriptor. According to this method, when the radius of a sphere centered at the feature vector of a query image is expanded in a feature vector space, if other feature vectors fall within the sphere, it is determined that the possibility that the query image is perceptually similar to the texture features of images corresponding to the feature vectors is very high.


However, for images corresponding to feature vectors newly falling within the sphere when the radius of the sphere is continuously increased in such a conventional similarity measuring method, the possibility that they are perceptually similar to the query image drops rapidly. Accordingly, the conventional similarity measuring method using a statistical-based texture descriptor has a problem in that an image which has been retrieved as being perceptually similar to a query image may not have a perceptual similarity to the query image.


SUMMARY OF THE INVENTION

To solve the above problem, it is a first object of the present invention to provide a method of accurately measuring the similarity between two images.


It is a second object of the present invention to provide a computer-readable recording medium for storing a computer program for executing the above method of measuring the similarity between images.


It is a third object of the present invention to provide a device for performing the above method of measuring the similarity between images.


Accordingly, to achieve the first object of the invention, in one aspect, there is provided a method of measuring the similarity between images. The method includes (a) computing a statistical dissimilarity between the images; (b) computing a perceptual dissimilarity between the images; and (c) computing a dissimilarity between the texture features of the images based on the statistical dissimilarity and the perceptual dissimilarity.


Preferably, the step (b) includes the steps of (b-1) computing regularity of the texture of the images; and (b-2) computing a dissimilarity between the computed regularities of texture of the images.


Preferably, the step (c) includes the steps of (c-1) determining the dissimilarity between the texture features as a value proportional to the statistical dissimilarity when the perceptual dissimilarity is smaller than a predetermined threshold.


In addition, it is preferable that the step (c) includes the steps of (c-1) determining the dissimilarity between the texture features based on the exponent of the power of the statistical dissimilarity, the exponent being the perceptual dissimilarity, when the perceptual dissimilarity is smaller than a predetermined threshold.


In another aspect, there is provided a method of measuring the similarity between images. The method includes the steps of (a) computing a statistical dissimilarity d(i, j) between two images i and j using a statistical-based texture descriptor; (b) obtaining quantitative measurements P(i) and P(j) of texture patterns of the two images i and j in terms of regularity; and (c) obtaining a dissimilarity between the texture features by computing a dissimilarity {circumflex over (d)}(|P(i)−P(j)|) between the texture patterns of the two images i and j in terms of regularity, where {circumflex over (d)} is assumed to be a predetermined function whose value is determined according to the range of a magnitude of a perceptual dissimilarity.


To achieve the second object of the invention, there is provided a computer-readable recording medium storing a computer program for executing a method of measuring a similarity between texture features of images. The medium includes a computer readable means for (a) computing a statistical dissimilarity d(i, j) between two images i and j using a statistical-based texture descriptor; (b) obtaining quantitative measurements P(i) and P(j) of texture patterns of the two images i and j in terms of regularity; and (c) obtaining a dissimilarity between the texture features by computing a dissimilarity {circumflex over (d)}(|P(i)−P(j)|) between the texture patterns of the two images i and j in terms of regularity, where {circumflex over (d)} is assumed to be a predetermined function whose value is determined according to the range of a magnitude of a perceptual dissimilarity.


To achieve the third object of the invention, there is provided a device for measuring a similarity between texture features of images. The device includes a statistical dissimilarity computing part for computing a statistical dissimilarity between the images; a perceptual dissimilarity computing part for computing a perceptual dissimilarity between the images; and a texture feature dissimilarity computing part for computing a dissimilarity between the texture features of the images based on the statistical dissimilarity and the perceptual dissimilarity.





BRIEF DESCRIPTION OF THE DRAWINGS

The above objectives and advantages of the present invention will become more apparent by describing in detail preferred embodiments thereof with reference to the attached drawings in which:



FIG. 1 is a block diagram illustrating the configuration of a device for measuring the similarity between images according to an embodiment of the present invention;



FIG. 2 is a flowchart illustrating the main steps of a method of measuring the similarity between images, which is performed in the device of FIG. 1 according to an embodiment of the present invention; and



FIG. 3 is a diagram illustrating a framework of measuring the similarity between two images.





DETAILED DESCRIPTION OF THE PRESENT INVENTION

Hereinafter, preferred embodiments of the present invention will be described in detail with reference to the attached drawings.


Referring to FIG. 1, a device for measuring the similarity between images according to an embodiment of the present invention includes a statistical dissimilarity computing part 102, a perceptual dissimilarity computing part 104 and a texture feature dissimilarity computing part 106. The perceptual dissimilarity computing part 104 includes a texture regularity measurer 104a1 and a regularity-dissimilarity computing part 104a2. FIG. 2 is a flowchart illustrating the main steps of a method of measuring the similarity between images, which is performed in the device of FIG. 1 according to an embodiment of the present invention.


In the operation of the device of FIG. 1, the statistical dissimilarity computing part 102 computes the statistical dissimilarity between two images i and j using a statistical-based texture descriptor in step 202. The computed dissimilarity is represented by d(i, j). The statistical dissimilarity may be computed according to the expanding radius of a sphere centered at a feature victor as described above. Other statistical descriptors may also be used.


Next, the texture regularity measurer 104a1 in the perceptual dissimilarity computing part 104 quantitatively measures the perceptual attribute of texture. In this embodiment, the texture regularity measurer 104a1 quantitatively measures the regularity of texture as the perceptual attribute of the texture. In other words, in step 204, the texture regularity measurer 104a1 obtains quantitative measurements P(i) and P(j) of regularity with respect to the texture pattern of the two images i and j. The quantitative measurements P(i) and P(j) are integers within a certain range. Larger values of the quantitative measurements P(i) and P(j) indicate stronger regularities of the texture. The regularities of the two images i and j can be computed using the quantitative measurements of the perceptual attribute of the texture. A method of obtaining a texture descriptor containing a perceptual browsing component (PBC) by extracting the PBC based on a quantitative measurement of a perceptual attribute of texture is disclosed in U.S. patent application Ser. No. 09/272,321, which was filed by this applicant on Mar. 19, 1999 and which is incorporated into the present specification by reference. Accordingly, the texture descriptor containing the PBC will not be further described in this specification.


Next, in step 206, the regularity-dissimilarity computing part 104a2 computes the dissimilarity between the texture patterns of the two images i and j in terms of regularity and outputs regularity-dissimilarity data |P(i)−P(j)| as the result. FIG. 3 is a diagram illustrating a framework for measuring the similarity between two images. Referring to FIG. 3, the statistical components and the perceptual components of the images i and j are obtained. The statistical dissimilarity d(i, j) and the perceptual dissimilarity {circumflex over (d)}(|P(i),P(j)|) of the two images i and j are obtained. In this embodiment, the perceptual dissimilarity {circumflex over (d)}(|P(i),P(j)|) is obtained from the regularity of texture features. Other perceptual measures of texture, such as directionality, may also be used.


The texture feature dissimilarity computing part 106 computes the dissimilarity between texture features of the two images using the statistical dissimilarity data output from the statistical dissimilarity computing part 102 and the regularity-dissimilarity data output from the regularity-dissimilarity computing part 104a2. In this embodiment, assuming α is a predetermined scaling factor, the dissimilarity D(i, j) of texture features of the two images i and j is computed by Equation (1) in step 208.

D(i,j)=d(i,j)+d(i,j)α{circumflex over (d)}(P(i),P(j))  (1)

The value of the scaling factor α depends on the effectiveness of the statistical description d(i, j). For a more reliable descriptor, the value of α is relatively small. Here, d(i, j) indicates a dissimilarity computed using a statistical-based texture descriptor, {circumflex over (d)} indicates a predetermined function whose value is determined depending on the range of magnitude of the perceptual dissimilarity, and {circumflex over (d)}(|P(i),P(j)|) indicates the measurement of dissimilarity between the texture patterns of the two images i and j in terms of regularity. For example, the computation of {circumflex over (d)} is defined as











d
^



(


P

(
i
)


,

P

(
j
)



)


=

{



0







P

(
i
)


-

P

(
j
)






1









P

(
i
)


-

P

(
j
)












P

(
i
)


-

P

(
j
)





>
1









(
2
)







The results of measuring the similarity between images using Equation 1 according to the similarity measuring method in different circumstances will be described.


First, according to the present invention, when both the statistical dissimilarity and the perceptual dissimilarity exceed a predetermined threshold, the dissimilarity between texture features is determined by a combination of the statistical dissimilarity and the perceptual dissimilarity. Assuming that the image j is an object to be measured and the image i is a query image in this embodiment, if the statistical dissimilarity d(i, j) and the perceptual dissimilarity between the two images i and j are both large, that is, if |P(i)−P(j)|>1, d(i,j)α{circumflex over (d)}(P(i),P(j)) increases exponentially. Accordingly, the dissimilarity D(i, j) of texture patterns of the two images i and j rapidly increases, which means that the image j is very different from the image i in terms of the regularity of a texture feature. With regard to search results, the exponential increase of d(i, j)α{circumflex over (d)}(P(i), P(j)) will push one pattern away and provide a chance for another pattern which is a little bit beyond the closest N matches, but which has similar regularity as the query pattern to be retrieved.


Second, according to the present invention, when a perceptual dissimilarity is smaller than the predetermined threshold, the dissimilarity between texture features is determined by a value proportional to a statistical dissimilarity. When the perceptual dissimilarity between the two images i and j is small in this embodiment, that is, when |P(i)−P(j)|≦1, the magnitude of an exponent term, d(i,j)α{circumflex over (d)}(P(i),P(j)), drops very fast and approaches 0. Accordingly, the dissimilarity D(i, j) of texture patterns approaches the statistical dissimilarity d(i, j), where d(i, j)>>1.


Third, when a statistical dissimilarity is smaller than the predetermined threshold, but a perceptual dissimilarity exceeds the predetermined threshold, the dissimilarity between texture features is partially determined depending on the perceptual dissimilarity. When the statistical dissimilarity d(i, j) is small but the perceptual dissimilarity is large, that is, when |P(i)−P(j)|>1, the dissimilarity D(i, j) of texture patterns is determined by the exponent of the second statistical dissimilarity component, wherein the exponent is the perceptual dissimilarity. However, since the statistical dissimilarity is relatively small, the dissimilarity between the texture features is only slightly influenced by the perceptual dissimilarity.


According to a method of measuring the similarity between images described above, the dissimilarity between texture features of the images is determined using the statistical dissimilarity and the perceptual dissimilarity between the texture features. When images perceptually similar to a query image are searched and retrieved according to the above measuring method, the possibility that the retrieved images are not perceptually similar to the query image is small. Accordingly, the present invention can improve the image searching performance when searching images having texture features similar to that of a query image. It is preferable that such a method of measuring the similarity between images according to the present invention is applied to content-based image searching.


Since when a dissimilarity is large, a similarity is small, but when a dissimilarity is small, a similarity is large, measuring of a dissimilarity is equivalent to measuring a similarity. Accordingly, the present invention is referred to as a similarity measuring method although it measures a dissimilarity.


A method of measuring the similarity between images according to the present invention can be written as a program which can be executed in a personal computer or a server computer. Program codes and code segments constructing the programs can be easily inferred by computer programmers skilled in this art. The programs can be stored in a computer-readable recording medium. The computer-readable medium can be a magnetic recording medium, an optical recording medium or a radio medium.


In the above embodiment, the dissimilarity between texture features has been obtained using the exponent of the power of a statistical dissimilarity, wherein the exponent is a perceptual dissimilarity. However, it will be apparent to one of ordinary skill in the art that modifications to the described embodiment may be made without departing from the spirit and scope of the invention. Accordingly, the scope of the present invention defined by the attached claims is not restricted to the above embodiment.

Claims
  • 1. A method of measuring a textural similarity of images, the method comprising: automatically determining a statistical dissimilarity (d(i, j)) between the images (i, j); andautomatically determining a perceptual dissimilarity ({circumflex over (d)}|P(i)−P(j)|) between the images; andautomatically determining a textural dissimilarity (D(i, j)) based on the statistical dissimilarity (d(i, j)) and a function of the perceptual dissimilarity ({circumflex over (d)}(|P(i)−P(j)|))where {circumflex over (d)} is a function whose value is dependant on a range and magnitude of the perceptual dissimilarity,wherein the textural dissimilarity D(i, j) is determined according to an equation D (i, j)=d(i, j)+d)i, j)α{circumflex over (d)}(P(i), P(j)) where α is a predetermined scaling factor and {circumflex over (d)} is a function defined as:
  • 2. A method according to claim 1, wherein determining the perceptual dissimilarity (|P(i)−P(j)|) comprises: determining quantitative measurements (P(i), P(j)) of a textural regularity of the respective images (i, j); anddetermining a difference between said quantitative measurements (P(i), P(j)).
  • 3. A method according to claim 1, wherein the textural dissimilarity (D(i, j)) is a value proportional to the statistical dissimilarity (d(i, j)) when the perceptual dissimilarity (|P(i)−P(j)|) is not larger than a predetermined threshold.
  • 4. A method according to claim 1, wherein a degree of influence of the statistical dissimilarity on the textural dissimilarity (D(i, j)) is determined based on a magnitude of the perceptual dissimilarity (|P(i)−P(j)|) when the perceptual dissimilarity (|P(i)−P(j)|) is greater than a predetermined threshold.
  • 5. A computer-readable recording medium storing a computer program for executing a method of measuring a textural similarity of images, wherein the method comprises: automatically determining a statistical dissimilarity (d(i, j)), between the images (i, j); andautomatically determining a perceptual dissimilarity (|P(i)−P(j)|)between the images andautomatically determining a textural dissimilarity (D(i, j)) based on the statistical dissimilarity (d(i, j)) and a function of the perceptual dissimilarity ({circumflex over (d)}(|P(i)−P(j)|)) where {circumflex over (d)} is a function whose value is dependant on a range and magnitude of the perceptual dissimilaritywherein the method comprises determining said textural dissimilarity D(i, j) in accordance with the equation D(i, j)=d(i, j)+d(i, j)α{circumflex over (d)}(P(i),P(j)), where α is a predetermined scaling factor and the function {circumflex over (d)} is defined as:
  • 6. An apparatus for measuring a textural similarity of images, the apparatus comprising: means for automatically determining a statistical dissimilarity (d(i, j)) between the images (i,j); andmeans for automatically determining a textural dissimilarity (D (i, j)) based on the statistical dissimilarity (d(i, j)) and a function of a perceptual dissimilarity ({circumflex over (d)}(|P(i)−P(j)|)) where {circumflex over (d)} is a function whose value is dependant on a range and magnitude of the perceptual dissimilaritywherein the means for determining the textural dissimilarity D(i, j) is configured to determine the textural dissimilarity D(i, j) according to the equation D (i, j)=d(i, j)+d)i,j)α{circumflex over (d)}(P(i), P(j)), where α is a predetermined scaling factor and {circumflex over (d)} is a function defined as:
  • 7. An apparatus according to claim 6, wherein the means for determining the perceptual dissimilarity (d(|P(i)−P(j)|)) comprises: means for determining quantitative measurements (P(i), P(j)) of a textural regularity of the respective images (i,j); andmeans for determining a difference between said quantitative measurements (P(i), P(j)).
  • 8. An apparatus according to claim 6, wherein the means for determining the textural dissimilarity (D(i, j) is configured to determine the textural dissimilarity (D(i, j)) as a value proportional to the statistical dissimilarity (d(i, j)) when the perceptual dissimilarity (|P(i)−P(j)|) is not larger than a predetermined threshold.
  • 9. An apparatus according to claim 6, wherein the means for determining the textural dissimilarity (D(i, j)) is configured to control a degree of influence of the statistical dissimilarity on the textural dissimilarity (D(i, j)) in dependence on the magnitude of the perceptual dissimilarity (|P(i)−P(j)|) when the perceptual dissimilarity (|P(i)−P(j)|) is greater than a predetermined threshold.
Priority Claims (3)
Number Date Country Kind
00-48324 Aug 2000 KR national
01303087 Mar 2001 EP regional
2001-165519 Mar 2001 JP national
Parent Case Info

This is a non-provisional application of pending provisional application 60/207,958 (filed May 31, 2000), and further claims the benefit of Korean Patent Application No. 00-48324. The contents of both documents are incorporated by reference.

US Referenced Citations (9)
Number Name Date Kind
5394483 Daly Feb 1995 A
5644655 Windsor Jul 1997 A
5852823 De Bonet Dec 1998 A
5949055 Fleet et al. Sep 1999 A
6229931 Essafi et al. May 2001 B1
6256409 Wang Jul 2001 B1
6360017 Chiu et al. Mar 2002 B1
6529631 Peterson et al. Mar 2003 B1
6591006 Niemann Jul 2003 B1
Foreign Referenced Citations (6)
Number Date Country
9-101970 Apr 1997 JP
9-101970 Apr 1997 JP
10-154149 Jun 1998 JP
10-154149 Jun 1998 JP
11-110556 Apr 1999 JP
10-2000-0014123 Mar 2000 KR
Related Publications (1)
Number Date Country
20020006221 A1 Jan 2002 US
Provisional Applications (1)
Number Date Country
60207958 May 2000 US