The present application claims priority under 35 U.S.C. § 119 to Japanese Patent Application No. 2023-148918, filed on Sep. 14, 2023. The above applications are hereby expressly incorporated by reference, in these entireties, into the present application.
The present invention is related to a document evaluation apparatus, a document evaluation method, and a document evaluation program for evaluating the comprehensibility of document data, as well as a document evaluation system.
The comprehensibility of information in a document is an ambiguous concept. As used herein, the term “comprehensibility” refers to the ease with which the contents of a document can be understood. It is difficult to objectively evaluate the comprehensibility of information in a document because the criteria for judging the comprehensibility of information in a document vary depending on the physical and ability characteristics, knowledge, skills, experience, and interests of a reader of the document, as well as the intended use and scenario of use of the document, and preferences.
Therefore, Non-Patent Document 1 (Universal Communication Design Association, “9 Items for Ease of Understanding”, Internet: <URL: https://ucda.jp/ninsho_mo kuteki/kijun.html #01>) classifies “information comprehensibility” into nine items from the viewpoint of information design for human communication, and defines evaluation criteria for each item. The nine items are “amount of information”, “task”, “text (clarity of meaning)”, “layout,”, “typography (legibility of text)”, “color design”, “marks and diagrams”, “entry (input) field”, and “problems in utilization”.
In addition, Patent Document 1 (Japanese Unexamined Patent Publication No. 2011-020387) proposes evaluation methods for the nine items disclosed in Non-Patent Document 1.
However, the evaluations conducted by the inventions disclosed in Non-Patent Document 1 and Patent Document 1 must be conducted by experts who have knowledge and evaluation skills in information design. Because there are individual differences in knowledge and skills even among experts, the human factor in conducting evaluation cannot be completely eliminated.
In addition, in Patent Document 1, dedicated evaluation methods, judging tools, and apparatuses are used for a plurality of types of evaluation items. Therefore, actual evaluation requires time and effort to use different evaluation methods, judgment tools, and apparatuses. Further, it is necessary for the experts to be familiar with the use of these evaluation methods, judgment tools, and apparatuses.
Therefore, there is demand for a device that collectively and automatically evaluates a plurality of evaluation items. However, documents to be evaluated contain various objects in addition to objects containing text, and in the case that a process that collectively evaluates a plurality of evaluation items is administered on such documents, the processing load is great, and problems arise such as a computing device having high processing power being required and time until evaluation results being obtained being required.
In addition, because some file formats such as PDF have complex structures, there is the problem of an excess burden being imposed if coordinate conversion processes such as object rotation are executed for each evaluation item, due to settings such as page rotation and page size for each page.
The present invention has been developed in view of the foregoing circumstances. It is an object of the present invention to provide a document evaluation apparatus, a document evaluation method, and a document evaluation program, as well as a document evaluation system that can reduce the burden and improve the efficiency of an evaluation process for evaluating a plurality of evaluation items related to the comprehensibility of a document.
A document evaluation apparatus of the present invention is equipped with: a document data acquisition unit that acquires document data including at least text objects; a preliminary processing unit that performs a preliminary process on the document data acquired by the document data acquisition unit; an information volume evaluation unit that evaluates the amount of information in the document data based on the preliminarily processed document data; a character evaluation unit that evaluates the legibility of text based on the preliminarily processed document data; and a color evaluation unit that evaluates the relationships among adjacent colors in the document data based on the preliminarily processed document data.
According to the document evaluation apparatus of the present invention, the preliminary process is performed on the document data, and based on the preliminarily processed document data, the amount of information in the document data, the legibility of text, and the relationship between adjacent colors are evaluated, thus reducing the burden and improving the efficiency of the evaluation process for a plurality of evaluation items related to the comprehensibility of the document.
Hereinafter, a document evaluation system that employs an embodiment of the document evaluation apparatus of the present invention will be described in detail with reference to the attached drawings.
The document evaluation system 1 of the present embodiment is equipped with a document evaluation apparatus 10 and a terminal device 20, as illustrated in
The document evaluation apparatus 10 acquires document data output from the terminal device 20 and evaluates the comprehensibility of the document data. The document evaluation apparatus 10 of the present embodiment collectively evaluates three evaluation items “amount of information”, “typography (legibility of text)”, and “color design” of the document data as evaluation items regarding comprehensibility of the document data, and outputs evaluation results to the terminal device 20 for display.
Specifically, the document evaluation apparatus 10 is equipped with a document data acquisition unit 11, a preliminary processing unit 12, an information volume evaluation unit 13, a character evaluation unit 14, a color evaluation unit 15, and an evaluation result output unit 16.
The document data acquisition unit 11 acquires document data output from the terminal device 20. The document data is document data that includes at least text objects, and may include vector graphic objects and bitmap image objects in addition to the text objects. Document data in PDF (Portable Document Format) format is output from the terminal device 20 and acquired by the document data acquisition unit 11 as the document data, for example.
The preliminary processing unit 12 administers a preliminary process on the document data acquired by the document data acquisition unit 11. The preliminary processing unit 12 of the present embodiment performs the preliminary process to remove objects that are not subjects of evaluation from the document data. Objects which are not subjects of evaluation are objects that are excluded from the three evaluation items described above.
Objects which are not subjects of evaluation include, for example, transparent objects, objects hidden in the background, and objects within a trimming range. These objects are visible by the creator of the document data when the document data is created on the terminal device 20, for example, but are not visible when the created document data is distributed to a person other than the creator and displayed by viewing software (such as Acrobat Reader). In other words, such objects are visible to the creator of the document data, but not to viewers of the document data after distribution.
Transparent objects include, for example, watermark character objects that the creator intentionally made transparent and inserted, or objects that the creator intentionally made transparent and inserted during a document data creation process, but forgot to delete when the document data was ultimately output from the terminal device 20. However, the transparent objects which are deleted are not limited to these, and other transparent objects are also deleted.
An object that is hidden in the background is, for example, a vector figure object that overlaps and is invisible by being hidden behind the overlapping portion.
Objects within a trimming range are objects that exist within a range of trimming applied by the trimming function of PDF editing software during a process of creating document data on the terminal device 20, for example. Objects within the trimming range are hidden when the document data is displayed or printed.
In addition, the preliminary processing unit 12 extracts metadata sets in PDF document data as the preliminary process.
Metadata is data related to the pages included in the document data, such as page rotation information, page information, and page size information. Metadata is employed during coordinate conversion of coordinates indicating the location of a target of alert, and in display of a report of evaluation results, to be described later. In the present embodiment, metadata is extracted before generating intermediate data to be described later, such that rotation processing is not performed when generating the intermediate data, and then rotation processing is performed using the metadata as necessary, thereby reducing the burden of the rotation processing and improving efficiency.
After the preliminary process to remove objects which are not subjects of evaluation and the preliminary process to extract metadata as described above, the preliminary processing unit 12 performs an intermediate data generation process to generate intermediate data based on the preliminarily processed document data. The intermediate data generation process generates intermediate data suitable for the evaluation of the three evaluation items described above from the preliminarily processed document data.
Specifically, the preliminary processing unit 12 generates bitmap image data and object information data as the intermediate data.
Bitmap image data is data generated by administering RIP (Raster Image Processer) processing to the preliminarily processed document data. The bitmap image data is used to evaluate the amount of information in the information volume evaluation unit 13 and to display color converted bitmap images in a color visibility simulation to be described later. Note that when the bitmap image data for evaluating the amount of information is generated, the preliminary processing unit 12 generates the bitmap image data for evaluating the amount of information after deleting figures and photographic objects that are not relevant to the evaluation of the amount of information in the document.
Object information data includes information such as object type, coordinates indicating the location of the object, object color, and the font, the size (number of points), and the line spacing of text included in the object. The object information data is employed to evaluate in the character evaluation unit 14 and to evaluate color design in the color evaluation unit 15.
In the present embodiment, the three evaluation items “amount of information”, “typography” and “color design” of the document data are evaluated employing the intermediate data generated based on the preliminarily processed document data by the preliminary processing unit 12 as described above.
The information volume evaluation unit 13 evaluates the amount of information in the document data. The information volume evaluation unit 13 of the present embodiment first administers an edge detection process to the bitmap image data for evaluating the amount of information generated by the intermediate data generation process. Then, the information volume evaluation unit 13 administers a binarization process to the bitmap image data in which edges have been detected, and evaluates the amount of information using the black-and-white bitmap image data obtained thereby. Specifically, the information volume evaluation unit 13 counts the number of black dots and the number of total dots in the black-and-white bitmap image data for each page of document data, and calculates a black dot ratio. The black dot ratio is the ratio of the number of black dots to the total number of dots in a page. The information volume evaluation unit 13 evaluates whether the black dot ratio is less than a reference value (19%, for example). In the case that the black dot ratio is less than the reference value, the document data is evaluated as not being a target of alert. In the case that the black dot ratio is greater than or equal to the reference value, the document data is evaluated as being a target of alert. This is because if the black dot ratio is above the reference value, the amount of information contained in a single page is too great to be easily understood.
In addition to the black dot ratio described above, the information volume evaluation unit 13 also counts the number of white dots and calculates a white dot ratio.
The character evaluation unit 14 evaluates the typography of the document data. The character evaluation unit 14 evaluates typography using the object information data generated by the intermediate data generation process as described above.
Specifically, the character evaluation unit 14 evaluates, based on the object information data, whether the size (number of points) of the characters included in each page of the document data is greater than or equal to a preset point threshold (8 points, for example). Then, the character evaluation unit 14 evaluates the size of the characters as not being a target of alert in the case that it is above the point threshold, and evaluates the size of the characters as being a target of alert in the case that it is below the point threshold. This is because when the character size is less than the point threshold, the characters are perceived as being too small and difficult to read.
The character evaluation unit 14 also evaluates whether the line spacing of the characters in each page of document data is greater than or equal to a preset line spacing threshold (1.5 lines, for example). The line spacing is an array spacing of each line, which is the sum of the height of the characters and a gap between each line and an adjacent line.
Then, the character evaluation unit 14 evaluates the line spacing as not being a target of alert in the case that the line spacing of the character is greater than or equal to the line spacing threshold, and evaluates the line spacing as being a target of alert in the case that the line spacing is below the line spacing threshold. This is because if the line spacing of a character is less than the line spacing threshold, the character spacing is perceived as being too narrow and difficult to read.
The character evaluation unit 14 evaluates whether the number of characters in each line of each page of document data is within a preset character count threshold (45 characters, for example). If the number of characters in each line is within the character count threshold, the character evaluation unit 14 evaluates it as not being a target of alert. If the number of characters in each line exceeds the character count threshold, the character evaluation unit 14 evaluates it as being a target of alert. This is because when the number of characters in a line exceeds the character count threshold, it is perceived that there are too many characters and difficult to read.
In addition, the character evaluation unit 14 also evaluates whether each character in the document data is deformed. Specifically, the character evaluation unit 14 evaluates whether the value of the aspect ratio, one of the font information for each character included in the document data, exceeds a certain range. In the PDF document data of the present embodiment, the aspect ratio described above is set for each character. If the character has no deformation, the aspect ratio is 100%, and if the character is stretched horizontally, the value of the aspect ratio is greater than 100%. In the present embodiment, this aspect ratio is taken from a character object, and in the case that the aspect ratio is within a certain range (e.g., 100%±25%), it is evaluated as not being a target of alert, and if it is outside the certain range, it is evaluated as being a target of alert. This is because if there is any deformation in characters, they are perceived as being difficult to read.
The character evaluation unit 14 may perform the evaluation process described above on a line by line basis.
The color evaluation unit 15 evaluates the color design of the document data. Specifically, the color evaluation unit 15 compares the colors of all adjacent objects for each page of document data and evaluates whether the colors are difficult to discriminate from each other. In the case that multiple colors are included in a single object, the color evaluation unit 15 divides the object by color, compares the colors of adjacent dots, and evaluates whether the colors are difficult to discriminate from each other.
The color evaluation unit 15 evaluates whether colors are difficult to discriminate from each other based on whether the difference in brightness of adjacent colors is greater than or equal to a preset brightness threshold value. In the case that the brightness difference between adjacent colors is greater than or equal to the preset brightness threshold, the color evaluation unit 15 evaluates the colors as not being a target of alert, and in the case that the brightness difference between adjacent colors is less than the preset brightness threshold, the color evaluation unit 15 evaluates the colors as being a target of alert.
When evaluating the color design as described above, the color evaluation unit 15 may create a color scheme list of adjacent objects and employ the color scheme list to evaluate the color design.
In addition to color evaluation for normal color vision, which uses the colors of the original document data as they are, the color evaluation unit 15 performs color evaluation for Protan color blindness (hereinafter referred to as “P type color vision”) (type 1 dichromatic perception), Deutan color blindness (hereinafter referred to as “D type color vision”) (type 2 dichromatic perception), and color evaluation for the elderly.
When performing color evaluation for P type color vision, the color evaluation unit 15 performs color conversion processing for P type color vision on the color of each object, and employs the converted colors to perform the color evaluation described above. When performing color evaluation for D type color vision, color conversion processing for D type color vision is applied to the color of each object, and the converted colors are employed to perform the color evaluation described above. When performing color evaluation for the elderly, color conversion processing for the elderly is applied to the color of each object, and the converted colors are employed to perform the color evaluation described above.
The color conversion process for P type color vision, the color conversion process for D type color vision, and the color conversion process for the elderly are each performed using preset color conversion tables.
The evaluation result output unit 16 outputs the evaluation results of the information volume evaluation unit 13, the character evaluation unit 14, and the color evaluation unit 15 to the terminal device 20 for display.
In the case that “Display Report” displayed in the dot count display field C2 is selected, the evaluation result output unit 16 creates a report for the document data that indicates whether the evaluation results for the amount of information is a target of alert or not for each page, and outputs it to the terminal device 20 for display.
The number of locations of targets of alert field T2 indicates the number of locations of targets of alert where the font size is below the point threshold, the number of lines of targets of alert where the line spacing is below the line spacing threshold, the number of lines of targets of alert where the number of characters in a line exceeds the character count threshold, and the number of locations of targets of alert where font deformation has been evaluated.
In the case that “Display Report” is selected in the number of locations of targets for alert field T2, the evaluation result output unit 16 generates a report for the document data that indicates whether the evaluation results for typography (font size, line spacing, number of characters in a line, and character deformation) are targets of alert or not for each page, and outputs the generated report to the terminal device 20 for display.
In addition, in the case that a specific “!” in the target of alert location display field S1 is selected, the evaluation result output unit 16 displays the RGB values of the original adjacent colors (first and second colors) of the selected location and the RGB values of the adjacent colors after being changed are displayed. The RGB values of the adjacent colors after being changed are color combinations for which the difference in brightness is greater than or equal to the brightness threshold, and which are recommended to be changed from the original color combinations. The recommended color combinations are obtained based on the RGB values of the original color combinations and are determined by a predefined table or function.
In the case that “Display Report” displayed in the number of target of alert locations display field S2 is selected, the evaluation result output unit 16 generates a report that indicate whether the color evaluation results for the document data are targets of alert or not for each page, and outputs it to the terminal device 20 for display.
In addition, the evaluation result output unit 16 is capable of outputting the original color document data, the document data after color conversion for P type color vision, the document data after color conversion for D type color vision, and the document data after color conversion for the elderly to the terminal device 20 for display. The evaluation result output unit 16 displays a display image selection screen as illustrated in
The document evaluation apparatus 10 is constituted by a computer and is equipped with a CPU (Central Processing Unit), semiconductor memory such as a ROM (Read Only Memory) and a RAM (Random Access Memory), storage such as a hard disk, and a communications interface. An embodiment of the document evaluation program of the present invention is installed in the storage of the document evaluation apparatus 10, and when this document evaluation program is executed by the CPU, each unit of the document evaluation apparatus 10 described above operates.
In the present embodiment, the functions of each part of the document evaluation apparatus 10 are realized by a document evaluation program, but the present invention is not limited to such a configuration. Some or all of the functions or control may be realized by hardware such as an ASIC (Application Specific Integrated Circuit), an FPGA (Field Programmable Gate Array), or other electrical circuits.
The terminal device 20 outputs document data in PDF format to the document evaluation apparatus 10 as described above, and displays the evaluation results regarding “amount of information”, “typography”, and “color design” of the document data.
The terminal device 20 is constituted by a computer and is equipped with a CPU, semiconductor memory such as a ROM and a RAM, storage such as a hard disk, and a communications interface.
A document evaluation application is installed in the storage of the terminal device 20, and the functions of the terminal device 20 described above are realized when this document evaluation application and the document evaluation program of the document evaluation apparatus 10 operate in conjunction.
The terminal device 20 is equipped with a control unit 21, an input unit 22, and a display unit 23. The control unit 21 is equipped with the CPU, semiconductor memory, storage and communications interface described above, and controls the entirety of the terminal device 20.
The input unit 22 is equipped with an input device such as a mouse and a keyboard and accepts various inputs from users. The display unit 23 is equipped with a display device, such as a liquid crystal display, and displays the evaluation results obtained by the document evaluation apparatus 10 described above.
Next, the flow of processes of the document evaluation system 1 of the present embodiment will be described with reference to
First, document data to be evaluated is created at the terminal device 20, and the document data to be evaluated is output to the document evaluation apparatus 10 (S10, PDF upload).
Next, a preliminary process is performed to remove objects from the document data which are not subjects of evaluation (S12). The document data D0 illustrated in
Next, after the objects which are not subjects of evaluation are removed from the document data, an intermediate data generation process is administered on preliminarily processed document data D1 to generate intermediate data (bitmap image data, object information data, and metadata) (S14).
Then, as illustrated in
In the information volume evaluation process, an edge detection process is performed on the bitmap image data generated in the intermediate data generation process for information volume evaluation, followed by a binarization process to generate black and white bitmap image data. Then, the amount of information on each page is calculated based on the black-and-white bitmap image data of each page, and the amount of information is evaluated. Next, display of a report on the amount of information is then performed. When displaying the report, the bitmap image data generated in the intermediate data generation process is displayed.
In the typography evaluation process, the size of characters, the spaces among lines of characters, the number of characters in each line, and the presence or absence of deformed characters are evaluated, based on the object information data generated in the intermediate data generation process described above. Then, the number of targets of alert and a list of the coordinates of the locations of the targets of alert on each page is obtained.
Here, there are cases in which page rotation (90° rotation, 180° rotation, etc., for example) is set for a page when the original document data was created, for example. In such a case, the appearance of the page is that after the page rotation, but the document data of the page itself is not rotated, and the typography evaluation process described above is also performed using the document data before the page rotation. Therefore, the list of coordinates of the locations of the targets of alert is a list of coordinates before page rotation. Accordingly, a coordinate conversion process is applied to the list of coordinate here such that the coordinate system thereof is the same as that of the page after page rotation. The coordinate conversion process is performed based on the number of pages, page size, and page rotation information included in the metadata.
The number of targets of alert on each page is then displayed as the result of the evaluation, and the locations of the targets of alert are highlighted based on the list of coordinates after the coordinate conversion. A typography evaluation report is then displayed. The bitmap image data generated by the intermediate data generation process is displayed in the displayed report.
In the color design evaluation process, whether it is difficult to discriminate adjacent colors is evaluated, based on the object information data generated by the intermediate data generation process described above. Then, a list of the coordinates of the locations of targets of alert is obtained for each page.
Here, in the color design evaluation process, in the case that page rotation is set for a page in the original document data, a coordinate conversion process is applied to the list of coordinates such that the coordinate system thereof is the same as the page after page rotation, in the same manner as in the typography evaluation process. This coordinate conversion process is performed based on the number of pages, page size, and page rotation information included in the metadata.
The number of targets of alert on each page is then displayed as the result of evaluation, and the locations of each target of alert based on the list of coordinates after coordinate conversion is displayed with a “!” mark. Then, a report of the color design evaluation is displayed. In the displayed report, the bitmap image data generated by the intermediate data generation process is displayed.
In the color vision simulation process, the bitmap image data generated by the intermediate data generation process is subjected to color conversion processes for P type color vision, D type color vision, and for elderly vision as described above, and the bitmap image data following these color conversion processes are generated and displayed, respectively. Then, a report of the bitmap image data after color conversion is displayed. When displaying the report, the bitmap image data before color conversion generated by the intermediate data generation process is also displayed.
In the document evaluation system 1 of the present embodiment, the preliminary process is performed to remove objects which are not subjects of evaluation from the document data, and based on the preliminarily processed document data, the amount of information in the document data, the legibility of text, and the relationships among adjacent colors are evaluated, thus reducing the burden and improving the efficiency of the evaluation process for a plurality of evaluation items related to the comprehensibility of a document.
In addition, the document evaluation system 1 of the present embodiment is equipped with the evaluation result output unit 16 that outputs the evaluation results of the information volume evaluation unit 13, the character evaluation unit 14, and the color evaluation unit 15. Therefore, users are enabled to immediately confirm the evaluation results on the terminal device 20.
Further, in the document evaluation system 1 of the present embodiment, document data may include vector graphic objects and bitmap image objects. Therefore, evaluation of various of document data is possible.
Still further, in the document evaluation system 1 of the present embodiment, document data may include font information for text. Therefore, the font information may be used in the typography evaluation process to evaluate whether characters are deformed.
Still yet further, in the document evaluation system 1 of the present embodiment, the evaluation result output unit 16 is capable of outputting information that indicates the locations of the targets of alert in the document data as evaluation results. Therefore, users are enabled to immediately understand the locations of the targets of alert in the document data.
In addition, in the document evaluation system 1 of the present embodiment, in the case that the document data includes information regarding page rotation, the evaluation result output unit 16 is capable of administering a coordinate conversion process according to the information regarding page rotation to the information that indicates the locations of targets of alert in the document data. Therefore, it is possible for the locations of targets of alert to be displayed accurately.
Next, another embodiment of the character evaluation process performed by the above character evaluation unit 14 will be described.
There are various types of typefaces for document data, including Japanese typefaces and European typefaces. In Japanese typefaces, characters are arranged such that they fit between a bottom reference line and a top reference line in the vertical direction, regardless of whether they are full width or half width characters. Meanwhile, European typefaces include letters that extend beyond a bottom reference, line, such as “g” and “j”.
Therefore, it is not desirable to use the same line spacing threshold for lines containing only Japanese typefaces and lines containing European typefaces when evaluating line spacing in the character evaluation process. Thus, different line spacing thresholds may be used for lines that do not contain European typefaces and lines that do contain European typefaces. Such a process will be described in detail below.
First, the character evaluation unit 14 determines whether the characters in the document data fall into the category of characters to be inspected or not to be inspected. Characters which are to be inspected are those whose meaning must be recognized by humans, and include both half width and full width characters in Japanese typefaces (hiragana, katakana, and kanji) and in European typefaces (English, French, German, etc.) that include both half width and full width characters. Characters which are not to be inspected include punctuation marks (“‘”, “°”, “,”, and “.”), symbols (“/”, “@”, and “¥”), spaces, etc.
Then, the character evaluation unit 14 determines whether a character which has been specified as that to be inspected is a European typeface. The typeface may be identified by character codes to determine whether it is a Japanese or European typeface, or by the type of font which is used.
Next, the character evaluation unit 14 determines whether the line contains a character to be inspected which has been determined to be a European typeface for each line in the document data.
As illustrated in
In contrast, as illustrated in
For this reason, when the character evaluation unit 14 evaluates the line spacing of a line which is determined not to include a European typeface, the line spacing is evaluated using a first line spacing threshold Th1. When evaluating the line spacing of a line which is determined to contain a European typeface, the character evaluation unit 14 uses a second line spacing threshold Th2, which is smaller than the first line spacing threshold Th1. By setting the second line spacing threshold Th2 to a value smaller than the first line spacing threshold Th1, it is possible to obtain a line spacing that improves the legibility of lines containing Japanese typefaces while maintaining the line spacing with a next line at the same value for the entire document in a sentence that includes a mixture of Japanese and European text.
In addition, when the character evaluation unit 14 evaluates the number of characters in a line, the characters not to be inspected described above may be excluded from the character count.
Incidentally, if the line spacing is set to be excessively wide in the case that the number of characters per line is small, it results in cohesiveness as a single sentence being difficult to achieve. In addition, the line spacing is the distance between each line of the sum of the height of the characters and the gap between each line, as described above, and is expressed as 1.5 lines (150%), etc., with the height of the characters as a reference. Thus, it can be said that the desired value for line spacing should be determined by the height of the characters and the number of characters per line.
Therefore, the document evaluation system 1 of the above embodiment may determine a recommended line spacing for text within a document based on the character size and number of characters or character size and width of the text in a direction perpendicular to the character arrangement direction. Thereby, it becomes possible to determine a line spacing which is easy to read.
Specifically, in the case that the number of characters in a line is 14 or less, the recommended line spacing is determined to be “character height×1.5”, that is, 1.5 lines.
Alternatively, in the case that the number of characters in a line is 15 or more and 39 or less, the recommended line spacing is determined to be “character height×(1.5+(number of characters-14)×0.02)”. That is, the recommended line spacing is determined to be 1.52 lines if the number of characters per line is 15 characters, 1.54 lines if the number of characters per line is 16 characters . . . 1.98 lines if the number of characters per line is 38 characters, and 2.0 lines if the number of characters per line is 39 characters.
In addition, if the number of characters per line is 40 or more and 45 or less, the recommended line spacing is determined to be “character height×2.0”, that is 2.0 lines.
Note that the method for determining a recommended line spacing described above is only one example, and the calculation formula and the maximum or minimum number of characters may be changed by user settings, for example. However, it is desirable for the line spacing to become greater the more characters there are per line (the wider the line) within a certain range of line spacing (1.5 to 2.0 lines, for example).
The recommended line spacing described above may then be used to evaluate the line spacing of the text.
In the document evaluation system 1 of the above embodiment, the portions of the document that are evaluated as targets of alert in the document evaluation are displayed in a highlighted manner. However, there are other manners of emphasized display, such as surrounding the targets of alert with lines, underlining the targets of alert, or placing an icon on the targets of alert, for example.
When additional information such as the lines that surround the targets of alert, the underlines, and the icons described above are superimposed on the document data, there may be cases in which it is difficult for an evaluator to distinguish between the additional information and the document data, depending on the combination of the color of the document data to be evaluated and the colors of the additional information.
Therefore, the evaluation result output unit 16 may adjust the color of the document data and the color of the additional information based on external input (input from the input unit 22 of the terminal device 20, for example) in order to improve the legibility of the document data to which the additional information has been added.
In the document evaluation system 1 of the above embodiment, in the case that portions of document data are evaluated as being targets of alert during the color design evaluation process, a recommended color combination after a change (proposed change) is determined and their RGB values are displayed. However, another embodiment of the method of determining the proposed change will be described below.
The evaluation result output unit 16 may determine a combination in which at least one of two colors is changed such that people with color blindness (colorblind people with P type color vision and D type color vision) would be able to distinguish the two colors from a color combination which is evaluated as a target of alert, and the change in at least one of the two colors is minimal for people with normal color vision may be determined as the proposed change in the color combination.
Specifically, the evaluation result output unit 16 first converts document data in a color space perceivable by people with normal color vision into image data in a color space perceivable by people with color blindness. The conversion from document data in the color space perceivable by people with normal color vision to image data in the color space perceivable by people with color blindness can be performed using known methods.
For example, the evaluation result output unit 16 may converts document data into document data in a linear sRGB color space according to Formulae (1) through (3) below.
Here, Rdevice, Gdevice, and Bdevice in Formulae (1) through (3) are the R, G, and B values in the document data, respectively. In addition, Rlinear, Glinear, and Blinear in Formulae (1) through (3) are the R, G, and B values in the linear sRGB color space, respectively.
Next, the evaluation result output unit 16 converts the image data in the linear sRGB color space to document data in an LMS color space which is perceivable by people with normal color vision according to Formula (4) below.
Continuing, the evaluation result output unit 16 converts the document data in the LMS color space which is perceivable by people with normal color vision into document data in an LMS color space which is perceivable by colorblind people with P type color vision, according to Formula (5) below.
The evaluation result output unit 16 converts the document data in the LMS color space which is perceivable by people with normal color vision into document data in an LMS color space which is perceivable by colorblind people with D type color vision according to Formula (6) below.
Then, the evaluation result output unit 16 converts the document data in the LMS color space which is perceivable by colorblind people with P type color vision into image data in an RGB color space which is perceivable by colorblind people with P type color vision. In addition, the evaluation result output unit 16 converts document data in the LMS color space which is perceivable by colorblind people with D type color vision into image data in an RGB color space which is perceivable by colorblind people with D type color vision. Here, the conversion from the LMS color space to the RGB color spaces may be performed using known methods.
Next, the evaluation result output unit 16 extracts combinations of two adjacent colors from the document data in the color spaces as RGB color spaces which are perceivable by people with color blindness.
Specifically, the evaluation result output unit 16 searches for combinations of two overlapping objects having different colors from each other in the document data in the color spaces which are perceivable by people with color blindness, using information indicating the position of each object. Then, the evaluation result output unit 16 extracts combinations of two colors within the combinations of two adjacent or overlapping objects having different colors from each other as combinations of two adjacent colors. Note that in the case that an object contains multiple colors, each portion having each of the colors is considered an object, and each two color combination of such portions is a combination of two adjacent colors.
Here, the evaluation result output unit 16 extracts combinations of two adjacent colors for each of the document data in the color space which is perceivable by colorblind people with P type color vision and the document data in the color space in the color space which is perceivable by colorblind people with D type color vision. Multiple combinations of two adjacent colors may be extracted from each type of document data.
Next, the evaluation result output unit 16 determines whether each of the combinations extracted in the manner described above is a combination is that which people with color blindness may easily confuse the two colors therein. Specifically, the evaluation result output unit 16 determines whether each combination is that which people with color blindness may easily confuse the two colors therein, based on the difference in brightness of the two colors which are perceived by people with color blindness.
The evaluation result output unit 16, for example, converts the RGB values of each of the two colors which are perceived by people with color blindness to L*a*b* values for each combination. The conversion from RGB values to L*a*b* values may be performed using known methods.
If the difference in the L* value, which is the difference in brightness between the two colors, is less than a predetermined threshold value, the evaluation result output unit 16 determines that the combination of the two colors is that which people with color blindness may easily confuse the two colors therein.
Here, the evaluation result output unit 16 determines whether each combination in the document data in the color space which is perceivable to colorblind people with P type color vision and the document data in the color space which is perceivable to colorblind people with D type color vision is a combination which colorblind people of each type of color vision may easily confuse the two colors therein.
Next, the evaluation result output unit 16 determines a proposed change for one of the colors of a combination of the two colors which is judged to be that which may be easily confused by people with color blindness, such that people with color blindness would be able to distinguish the two colors, and the change in at least one of the two colors is minimal for people with normal color vision.
Here, the L*a*b* values of the two colors which are perceived by people with normal color vision corresponding to the two colors of a combination which may be easily confused by people with color blindness are designated as [L1_n, a1_n, b1_n] and [L2_n, a2_n, b2_n], respectively. Between these, the color represented by [L2_n, a2_n, b2_n] is the target of change.
In addition, the L*a*b* values of the two colors above which are perceived by colorblind people with P type color vision are designated as [L1_p, a1_p, b1_p] and [L2_p, a2_p, b2_p], respectively. Further, the L*a*b* values of the two colors above which are perceived by colorblind people with D type color vision are designated as [L1_d, a1_d, b1_d] and [L2_d, a2_d, b2_d], respectively.
The evaluation result output unit 16 obtains [L3_n, a3_n, b3_n], which are the L*a*b* values of the color which is perceived by people with normal color vision after the color is changed, by solving a constrained optimization problem that minimizes the evaluation function J shown in Formula (7) below.
Here, min J (x)=f (x) is an equation that represents an optimization problem for obtaining a vector x that minimizes the evaluation function J. In addition, s. t. represents constraining conditions with respect to variables with respect to the evaluation function J.
The first line in Formula (7) means that a color with the smallest color difference for people with normal color vision from the color before the change is to be obtained as the color after the change.
In addition, const in Formula (7) is a preset threshold value for a difference in brightness at which people with color blindness can distinguish between two colors. L3_p and L3_d in Formula (7) are the L* values of the colors, which are respectively perceived by colorblind people with P type and D type color vision, for the colors [L3_n, a3_n, b3_n] described above.
The condition at the bottom line in Formula (7) is a condition that means that the L*a*b* value correspond to those which can be expressed in the RGB color space.
In the case that there are multiple solutions to the constrained optimization problem of Formula (7), the evaluation result output unit 16 selects the color with the smallest difference in brightness from the color to be changed as the color of the proposed change. That is, the evaluation result output unit 16 selects the color with the minimum value of | L2_n-L3_n| as the color of the proposed change.
The following is a specific example of how to solve the constrained optimization problem of Formula (7).
First, the L*a*b* color space which is perceived by people with normal color vision is quantized at predetermined intervals, and all L*a*b* values are compiled in a list.
Next, each of the L*a*b* values which are perceived by people with normal color vision in the above list is converted into L*a*b* values which are perceived by colorblind people with P type color vision, and each of the converted L*a*b* values is correlated and saved with each of the L*a*b* values which are perceived by people with normal color vision. In addition, each of the L*a*b* values which are perceived by people with normal color vision in the above list is converted into L*a*b* values which are perceived by colorblind people with D type color vision, and each of the converted L*a*b* values is correlated and saved with each of the L*a*b* values which are perceived by people with normal color vision.
Here, conversion of the L*a*b* values each of the L*a*b* values which are perceived by people with normal color vision to the L*a*b* values which are perceived by colorblind people with P type color vision, and conversion of the L*a*b* values each of the L*a*b* values which are perceived by people with normal color vision to the L*a*b* values which are perceived by colorblind people with D type color vision may be performed employing known methods.
For example, in the case that the L*a*b* values which are perceived by people with normal color vision are converted to the L*a*b* values which are perceived by colorblind people with P type color vision, first, the L*a*b* values which are perceived by people with normal color vision are converted to RGB values which are perceived by people with normal color vision. Next, the RGB values which are perceived by people with normal color vision are converted to RGB values which are perceived by colorblind people with P type color vision. This conversion may be performed by the method described above. Next, the RGB values which are perceived by colorblind people with P type color vision are converted to L*a*b* values which are perceived by colorblind people with P type color vision. The conversion from RGB values to L*a*b* values may be performed by employing known methods.
The conversion from L*a*b* values which are perceived by people with normal color vision to L*a*b* values which are perceived by colorblind people with D type color vision can be performed in the same manner as the conversion to L*a*b* values which are perceived by colorblind people with P type color vision described above.
Next, each of the L*a*b* value which are perceived by people with normal color vision that correspond to each of the L*a*b* values, the difference in brightness (L* value) of which is less than const in Formula (7) for [L1_p, a1_p, b1_p] in the list of L*a*b* values which are perceived by colorblind people with P type color vision described above, is deleted from the list of L*a*b* values which are perceived by people with normal color vision. In addition, each of the L*a*b* values each of the L*a*b* value which are perceived by people with normal color vision that correspond to each of the L*a*b* values, the difference in brightness (L* value) of which is less than const in Formula (7) for [L1_d, a1_d, b1_d] of each L*a*b* value which are perceived by colorblind people with D type color vision described above, is deleted from the list of L*a*b* values which are perceived by users having normal color vision.
The procedure above leaves L*a*b* values of multiple colors whose brightness difference is greater than const with respect to [L1_n, a1_n, b1_n], which are the L*a*b* values of the color that is not the target of change, in the list of L*a*b* values which are perceived by people with normal color vision described above. That is, multiple colors, of which the brightness difference from the color that is not a target of change is greater than const, are extracted from the L*a*b* color space which is perceived by people with normal color vision, as illustrated in
Next, L*a*b* values having the shortest distance from [L2_n, a2_n, b2_n], which are the L*a*b* values of the color to be changed on an a*b* plane are selected from among the L*a*b* values that remain in the list of L*a*b* values which are perceived by people with normal color vision, and are L*a*b* values of colors that can be expressed in an RGB color space.
Thereby, the color that can be expressed in the RGB color space and has the smallest color difference from the color to be changed is extracted from each color extracted as the color with a brightness difference greater than const from the color which is not to be changed, as illustrated in
Here, there may be multiple colors with the smallest color difference (distance on the a*b* plane) from the color to be changed, as illustrated in
The evaluation result output unit 16 determines the proposed change of one color by solving the constrained optimization problem of Formula (7) as described above for all combinations of two colors in which at least one of colorblind people with P type color vision and colorblind people with D type color vision is likely to confuse the two colors.
After the colors of the proposed changes are determined, the evaluation result output unit 16 transmits the proposed changes to the terminal device 20. Thereby, it becomes possible for the terminal device 20 to notify a user by displaying the proposed change of one of the colors in the combinations of two colors in which people with color blindness are likely to confuse the two colors on the display 23.
The present invention is not limited to the above embodiments, and may be embodied by modifying the constituent elements to an extent that it does not depart from the spirit and scope thereof at the implementation stage. In addition, various inventions may be realized by appropriate combinations of the plurality of constituent elements disclosed in the above embodiments. For example, all of the constituent elements disclosed in the embodiments may be combined as appropriate. It is, of course, possible for various modifications and applications to be realized within a scope that does not depart from the spirit and the scope of the present invention.
The additional items below are further disclosed with respect to the present invention.
The document evaluation apparatus of the present invention is equipped with: a document data acquisition unit that acquires document data including at least text objects; a preliminary processing unit that performs a preliminary process on the document data acquired by the document data acquisition unit; an information volume evaluation unit that evaluates an amount of information in the document data based on the preliminarily processed document data; a character evaluation unit that evaluates the legibility of text based on the preliminarily processed document data; and a color evaluation unit that evaluates the relationships among adjacent colors in the document data based on the preliminarily processed document data.
In the document evaluation apparatus of Item 1, the preliminary processing unit may perform a process that removes objects which are not subjects of evaluation from the document data as the preliminary process.
The document evaluation apparatus of Item 1 or 2 may be equipped with an evaluation result output unit that outputs the evaluation results of the information volume evaluation unit, the character evaluation unit, and the color evaluation unit.
In the document evaluation apparatus of any of Items 1 through 3, the document data may include vector graphic objects and bitmap image objects.
In the document evaluation apparatus of any of Items 1 through 4, the document data may include font information of text.
In the document evaluation apparatus of Item 3, the evaluation result output unit may output information indicating locations of targets of alert in the document data, as evaluation results.
In the document evaluation apparatus of Item 6, in the case that the document data includes page rotation information, the evaluation result output unit may administer a coordinate transformation process according to the page rotation to the information that indicates the locations of the targets of alert in the document data.
The document evaluation system of the present invention is equipped with a document evaluation apparatus of any of Items 1 through 7, and a terminal device that displays the evaluation results obtained by the document evaluation apparatus.
A method of evaluating a document of the present invention acquires document data that contains at least text objects, performs a preliminary process on the acquired document data, evaluates the amount of information in the document data based on the preliminarily processed document data, evaluates the legibility of text based on the preliminarily processed document data, and evaluates the relationship between adjacent colors in the document data based on the preliminarily processed document data.
A document evaluation program of the present invention causes a computer to execute the steps of: acquiring document data including at least text objects; performing preliminary process on the acquired document data; evaluating the amount of information in the document data based on the preliminarily processed document data; evaluating the legibility of text based on the preliminarily processed document data; and evaluating the relationship between adjacent colors in the document data based on the preliminarily processed document data.
Number | Date | Country | Kind |
---|---|---|---|
2023-148918 | Sep 2023 | JP | national |