This patent application is based on and claims priority to Japanese patent application No. JPAP2004-059694 filed on Mar. 3, 2004, in the Japanese Patent Office, the entire contents of which are hereby incorporated by reference.
The present invention relates to an apparatus, method, system, and computer program and product, each capable of generating a normalized image from an original image.
In today's office environment, various document management functions, including a document reading function, a document editing function, and a document outputting function, are usually performed separately in different apparati. For example, an original image may be read by a scanner, and stored as image data in a personal computer. The stored image data may be displayed using a display or may be output by a printer as a printed image. In such situations, the original image may be reproduced in lower image quality or different colors as the scanner, the display, and the printer have different characteristics from one another.
To solve this problem of characteristics differences, image processing is applied to the original image, using characteristic information obtained from each of the scanner, display, and the printer. However, such characteristic information may not be always available. Further, image processing optimal for one apparatus may not be always optimal for another apparatus.
In addition to the above-described problem, there is a problem of having the original image of lower image quality. For example, the original image may include undesired information caused during a scanning process, such as noise, abnormal illumination, geometric distortion, etc. In such a case, image processing capable of eliminating the undesired information is desired.
Preferred embodiments of the present invention include an apparatus, method, system, and computer program and product, each capable of generating normalized image data from original image data.
In one exemplary embodiment, the normalized image data is generated using feature information. The feature information includes at least one of pixel color information, pixel frequency information, and pixel position information, of the original image data. Using the feature information, predetermined image processing is applied to the original image data to generate the normalized image data.
In another exemplary embodiment, the normalized image data is generated using feature information, which corresponds to image preference information. The feature information includes at least one of pixel color information, pixel frequency information, and pixel position information, of the original image data. Using the feature information, image processing selected based on the image preference information is applied to the original image data to generate the normalized image data.
In another exemplary embodiment, the normalized image data is generated using at least one of feature information and characteristic information. The characteristic information relates to inputting of the original image data. At least one of the feature information and the characteristic information is selected, and image processing corresponding to the selected information is applied to the original image data to generate the normalized image data.
A more complete appreciation of the disclosure and many of the attendant advantages thereof will be readily obtained as the same becomes better understood by reference to the following detailed description when considered in connection with the accompanying drawings, wherein:
In describing preferred embodiments illustrated in the drawings, specific terminology is employed for the sake of clarity. However, the disclosure of this patent specification is not intended to be limited to the specific terminology so selected and it is to be understood that each specific element includes all technical equivalents that operate in a similar manner. Referring now to the drawings, wherein like reference numerals designate identical or corresponding parts throughout the several views, particularly to
The image processing system 100 includes an image input 101, an image processor 102, an image output 103, and a controller 104. The image input 101 includes any kind of device capable of inputting an original image as original image data. The image processor 102 includes any kind of device capable of generating normalized image data from the original image data. The image output 103 includes any kind of device capable of outputting the normalized image data. In this specification, image normalization refers to eliminating image variations, which are related to characteristics of the image input 101 but are irrelevant to object identity. The normalized image data may be generated using at least one of feature information of the original image data and characteristic information of the image input 101.
The controller 104 controls an entire operation of the image processing system 100, according to an instruction given by a user.
For example, the controller 104 may cause the image processor 102 to generate the normalized image data using the feature information extracted from the original image data. The feature information includes, for example, pixel color information, pixel frequency information, or pixel position information. In this way, image normalization may be performed independently of the characteristic information of the input image 101.
In another example, the controller 104 may obtain image preference information from a user, and cause the image processor 102 to generate the normalized image data corresponding to the image preference information. The image preference information includes information regarding a desired image quality of an output image, such as moiré suppression, improved color contrast, and improved color reproduction, for example. The image preference information may further include a desired size of an output image or a desired format of an output image. In this way, image normalization may be performed according to a user preference.
In another example, the controller 104 may obtain characteristic information from the image input 101, and cause the image processor 102 to generate normalized image data using the characteristic information. The characteristic information includes, for example, various image processing parameters, algorithms, or models, specific to an apparatus or an application program, which operates as the image input 101. In this way, image normalization may be performed, while considering the characteristic information of the image input 101.
The image processing system 100 of
The image processing system 200 of
In another exemplary operation, the scanner 203 reads an original image into original image data, and sends it to the server PC 201. The server PC 201 generates normalized image data from the original image data, and sends it to the printer 204, which prints the normalized image data as a printed image.
In yet another exemplary operation, the scanner 203 reads an original image into original image data, and sends it to the document server 202. The server PC 201 obtains the original image data from the document server 202, generates normalized image data from the original image data, and sends it to the mail server 206, which sends the normalized image data as email data.
As described above, the server PC 201 performs image processing of the present invention. The server PC 201 may be implemented as any kind of processor, for example, as shown in
The server PC 201 of
The CPU 302 controls an entire operation of the server PC 201. The ROM 304 includes any kind of involatile memory. The RAM 303 includes any kind of volatile memory, functioning as a work memory for the CPU 302. The HDD 306 includes any kind of storage device, capable of storing various kinds of programs including an image processing program of the present invention. The removable disc device 308 includes any kind of device capable of reading various data or programs from any kind of storage medium. The storage medium, which may store the image processing program of the present invention, includes any kind of optical disc such as CDs (compact discs) and DVDs (digital video discs), or any kind of disk such as a floppy disk and a magneto optical disk, for example. The display 310 includes any kind of display, such as a liquid crystal display, capable of displaying various data. The keyboard 311, and the pointing device 312 such as a mouse, allows a user to input instructions. The network interface 309 allows the server PC 201 to communicate with any other device via the network 207 of
In an exemplary operation, when the server PC 201 is activated, the CPU 302 loads the image processing program of the present invention from the HDD 306 onto the RAM 303, and starts operating according to the image processing program.
Alternatively, the CPU 302 may download the image processing program of the present invention from the storage medium using the removable disc device 208, or it may download the image processing program from the outside via the network interface 309.
Further, the image processing program of the present invention may be operated on a predetermined operating system (OS), or may be included as a part in a group of files implementing an application software program such as a word processing program or the OS.
Still referring to
In one exemplary embodiment, the server PC 201, the document server 202, the client PC 205, and the mail server 206 are incorporated into one processor operating based on the image processing program of the present invention. In another exemplary embodiment, all of the server PC 201, the document server 202, the client PC 205, and the mail server 206 are incorporated into one processor operating based on the image processing program of the present invention. In yet another exemplary embodiment, the scanner 203 and the printer 204 are incorporated into one device, such as a multifunctional printer. In yet another exemplary embodiment, all of the devices shown in
Further, the image processing system 200 of
Referring now to
Step S1 receives original image data from any one of the scanner 203, the document server 202, the client PC 205, and the mail server 206. Further, the original image data may be previously stored in the server PC 201.
Step S2 extracts destination information included in the user instruction received via the client PC 205. The destination information indicates a destination to which the image data is to be sent. In one example, the client PC 205 may instruct the server PC 201 to store the image data in the document server 202 as an image file for future use. In another example, the client PC 205 may instruct the server PC 201 to store the image data in the client PC 205 as an image file for image editing. In yet another example, the client PC 205 may instruct the server PC 201 to send the image data as email data to the mail server 206.
Step S3 determines whether image preference information is included in the user instruction. If the image preference information is available, the operation moves to Step S4 to extract feature information from the original image data based on the image preference information. If the image preference information is not available, the operation moves to Step S5 to extract feature information from the original image data based on default settings.
Step S6 applies image processing to the original image data, using the feature information extracted in Step S4 or S5. As a result, normalized image data is generated.
Step S7 stores or sends the normalized image data, according to the destination information obtained in Step S2, and the operation ends. In one example, the server PC 201 may store the normalized image data in one of the storage devices 305 and 307. In another example, the server PC 201 may send the normalized image data to the document server 202 for future use. In another example, the server PC 201 may send the normalized image data to the printer 204 for printing.
Referring now to
The halftone converter 401 converts the original image data from halftone to grayscale. The tone corrector 402 adjusts color contrast between character image data and background image data in the original image data. The character color corrector 403 adjusts character color in the original image data. The image processing controller 404 selects at least one of the halftone converter 401, the tone corrector 402, and the character color corrector 403, according to image preference information. If the image preference information is not available, the image processing controller 404 of
The halftone converter 401 extracts feature information of original image data, and applies image processing using the feature information, as illustrated in
Referring to
The tone corrector 402 extracts feature information of original image data, and applies image processing using the feature information, as illustrated in
Referring to
The character color corrector 403 extracts feature information of original image data, and applied image processing based on the feature information, as illustrated in
Step S708 of
Step S709 corrects a character color of each pixel according to the reference value calculated from the statistics of saturation. For example, the saturation of each of the character pixels is compared with the reference value, which is a threshold value set based on the statistics of saturation. If the saturation of a target pixel is equal to or below the reference value, color of the target pixel is adjusted to be a character color.
Referring now to
Step S13 determines whether characteristic information is available from the scanner 203. If the characteristic information is available, the operation moves to Step S14 to select image processing, such as image processing algorithms or parameters, corresponding to the characteristic information. If the characteristic information is not available, the operation moves to Step S5 to extract feature information from the original image data based on the default settings, and to Step S6 to generate normalized image data using the feature information.
Step S16 applies image processing, selected in Step S14, to the original image data. The processed image data is then stored or sent to the destination in Step S7.
The first halftone converter 801, the first tone corrector 803, and the first character color corrector 805 operate in a substantially similar manner to the halftone converter 401, the tone corrector 402, and the character color corrector 403, respectively.
The second halftone converter 802 converts original image data from halftone to grayscale, using the characteristic information of the scanner 203. For example, the second halftone converter 802 may perform an image processing operation described in the U.S. Pat. No. 5,825,937 filed on Mar. 27, 1997, the entire contents of which are hereby incorporated by reference. In another example, the second halftone converter 802 may perform an image processing operation using a frequency characteristic of the scanner 203.
The second tone corrector 804 adjusts color contrast between character image data and background image data in original image data, using characteristic information. For example, the second tone corrector 804 may perform scanner gamma correction using a gamma correction table of the scanner 203.
The second character color corrector 806 adjusts character color in original image data, using characteristic information. For example, the second character color corrector 806 may perform an image processing operation using a frequency characteristic of the scanner 203.
In an exemplary operation, the image processing controller 807 selects the first halftone converter 801 when the characteristic information is not known, while it selects the second halftone converter 802 when the characteristic information is known. Similarly, the image processing controller 807 may select at least one of the first tone correctors 803 and 804, or at least one of the first character color correctors 805 and 806, depending on the availability of the characteristic information.
In another exemplary operation, the image processing controller 807 selects at least one of the first and second halftone converters 801 and 802, at least one of the first tone correctors 803 and 804, or at least one of the first character color correctors 805 and 806, according to a user preference.
Numerous additional modifications and variations are possible in light of the above teachings. It is therefore to be understood that within the scope of the appended claims, the disclosure of this patent specification may be practiced otherwise than as specifically described herein.
For example, elements and/or features of different illustrative embodiments may be combined with each other and/or substituted for each other within the scope of this disclosure and appended claims.
Number | Date | Country | Kind |
---|---|---|---|
2004-059694 | Mar 2004 | JP | national |
Number | Name | Date | Kind |
---|---|---|---|
4977448 | Murata et al. | Dec 1990 | A |
5282046 | Yamaguchi | Jan 1994 | A |
5914793 | Suzuki et al. | Jun 1999 | A |
6172770 | Inoue | Jan 2001 | B1 |
6215561 | Kakutani | Apr 2001 | B1 |
6441915 | Imaizumi et al. | Aug 2002 | B1 |
6678405 | Kondo et al. | Jan 2004 | B1 |
6757083 | Nakamura | Jun 2004 | B1 |
6765206 | Sugiyama et al. | Jul 2004 | B2 |
6870637 | Watanabe | Mar 2005 | B1 |
7002706 | Sato | Feb 2006 | B2 |
7016072 | Yamazaki | Mar 2006 | B1 |
7123381 | Klassen | Oct 2006 | B2 |
7218421 | Tsuji | May 2007 | B2 |
7411694 | Nomizu | Aug 2008 | B2 |
7529003 | Fukawa | May 2009 | B2 |
20020075465 | Nakamura et al. | Jun 2002 | A1 |
20040070797 | Moritaku | Apr 2004 | A1 |
20050099643 | Sawada et al. | May 2005 | A1 |
20070041027 | Malik et al. | Feb 2007 | A1 |
20090185204 | Wu et al. | Jul 2009 | A1 |
Number | Date | Country |
---|---|---|
04-115669 | Apr 1992 | JP |
04-369971 | Dec 1992 | JP |
06-203202 | Jul 1994 | JP |
08-098016 | Apr 1996 | JP |
2001-341549 | Dec 2000 | JP |
2001-014458 | Jan 2001 | JP |
2001-197306 | Jul 2001 | JP |
2001-333237 | Nov 2001 | JP |
2001-153298 | May 2004 | JP |
2004-153298 | May 2004 | JP |
Number | Date | Country | |
---|---|---|---|
20050200911 A1 | Sep 2005 | US |