The disclosure relates to an electronic apparatus and an operating method of the same. More particularly, the disclosure relates to an electronic apparatus for evaluating the quality of an image and an operating method of the electronic apparatus.
With the development and use of electronic apparatuses, a method of evaluating the quality of an image has been studied in an operation of displaying an image, or converting the quality, size, and shape of an image.
Image quality evaluation methods include a reference image quality evaluation method that refers to an evaluation reference image or a no-reference image quality evaluation method that performs evaluation without a reference image. Image quality evaluation may be determined based on an indicator such as Pearson's linear correlation coefficient (PLCC) or Spearman's rank order correlation coefficient (SRCC).
With recent developments in technology, deep learning-based no-reference image quality evaluation methods have been actively studied. In particular, with the development of transformer technology, the performance of no-reference image quality evaluation methods has also been improved.
Provided is an electronic apparatus for evaluating quality of an image.
According to an aspect of the disclosure, an electronic apparatus for evaluating quality of an image includes: at least one memory storing at least one instruction; and at least one processor configured to execute the at least one instruction, wherein the at least one instruction, when executed by the at least one processor, causes the electronic apparatus to: obtain the image, extract a feature map based on the image comprising a feature of the image, obtain, based on the feature map, a quality score for each reference region of the image, obtain, based on the feature map, an importance for each reference region of the image, and evaluate the quality of the image according to a final quality score of the image, wherein the final quality score is based on the quality score for each reference region and the importance for each reference region.
The at least one instruction, when executed by the at least one processor, may further cause the electronic apparatus to extract the feature map by inputting the image to a shifted window (Swin) transformer model.
The at least one instruction, when executed by the at least one processor, may further cause the electronic apparatus to extract the feature map by inputting the image to a revised-shifted window (Swin) transformer model, and the revised-Swin transformer model may be a model obtained by revising a Swin transformer model so that the feature map is extracted using a window having a same resolution regardless of a resolution of the image.
A resolution of a window used in the revised-Swin transformer model may be lower than a resolution of a window used in the Swin transformer model.
The reference region may include a region corresponding to a unit area of the feature map.
The Reference region may include a plurality of reference regions, and the at least one instruction, when executed by the at least one processor, may further cause the electronic apparatus to obtain the quality score for each of the plurality of reference regions by performing a convolution operation on the feature map using a first filter.
A resolution of the first filter may correspond to a resolution of each of the plurality of reference regions.
The one or more reference regions may include a plurality of reference regions, and the at least one instruction, when executed by the at least one processor, may further cause the electronic apparatus to obtain the importance for each of the plurality of reference regions by performing a convolution operation on the feature map using a second filter.
A resolution of the second filter may correspond to a resolution of two or more of the plurality of reference regions.
A resolution of the second filter may correspond to two or more reference regions spaced apart from each other from among the plurality of reference regions.
According to an aspect of the disclosure, a method of operating an electronic apparatus for evaluating quality of an image includes: obtaining the image; extracting a feature map based on the image comprising a feature of the image; obtaining, based on the feature map, a quality score for each reference region of the image; obtaining, based on the feature map, an importance for each reference region of the image; and evaluating the quality of the image according to a final quality score of the image, wherein the final quality score is based on the quality score for each reference region and the importance for each reference region.
The extracting the feature map may include extracting the feature map by inputting the image to a shifted window (Swin) transformer model.
The extracting the feature map may include extracting the feature map by inputting the image to a revised-shifted window (Swin) transformer model, and wherein the revised-Swin transformer model may be a model obtained by revising a Swin transformer model wherein the feature map is extracted using a window having a same resolution regardless of a resolution of the image.
A resolution of a window used in the revised-Swin transformer model is lower than a resolution of a window used in the Swin transformer model.
The reference region may include a plurality of reference regions. The obtaining the quality score for each of the plurality of reference regions may include performing a convolution operation on the feature map using a first filter.
According to an aspect of the disclosure, a non-transitory computer readable medium having instructions stored therein, which when executed by at least one processor, cause the at least one processor to execute a method of controlling an electronic apparatus for evaluating quality of an image, the method including: obtaining the image; extracting a feature map based on the image comprising a feature of the image; obtaining, based on the feature map, a quality score for each reference region of the image; obtaining, based on the feature map, an importance for each reference region of the image; and evaluating the quality of the image according to a final quality score of the image, wherein the final quality score is based on the quality score for each reference region and the importance for each reference region.
With regard to the method executed by the at last one processor based on the instructions stored on the non-transitory computer readable medium, the extracting the feature map may include extracting the feature map by inputting the image to a shifted window (Swin) transformer model.
With regard to the method executed by the at last one processor based on the instructions stored on the non-transitory computer readable medium, the extracting the feature map may include extracting the feature map by inputting the image to a revised-shifted window (Swin) transformer model, and the revised-Swin transformer model may be a model obtained by revising a Swin transformer model wherein the feature map is extracted using a window having a same resolution regardless of a resolution of the image.
With regard to the method executed by the at last one processor based on the instructions stored on the non-transitory computer readable medium, a resolution of a window used in the revised-Swin transformer model may be lower than a resolution of a window used in the Swin transformer model.
With regard to the method executed by the at last one processor based on the instructions stored on the non-transitory computer readable medium, the obtaining the quality score for each of the plurality of reference regions may include performing a convolution operation on the feature map using a first filter.
According to an aspect of the disclosure, an electronic apparatus includes: at least one memory storing at least one instruction; and at least one processor configured to execute the at least one instruction, wherein the at least one instruction, when executed by the at least one processor, causes the electronic apparatus to: obtain an image, extract a feature map of the image comprising at least one feature, identify one or more reference regions of the image, obtain a quality score for each of the one or more reference regions, obtain an importance for each of the one or more reference regions, and evaluate a quality level of the image according to a final quality score of the image, wherein the final quality score is based on the quality score for each of the one or more reference regions and the importance for each of the one or more reference regions.
The above and other aspects, features, and advantages of certain embodiments of the present disclosure will be more apparent from the following description taken in conjunction with the accompanying drawings, in which:
The terms used herein will be briefly described, and an embodiment of the disclosure will be described in detail.
The terms used herein are general terms used in the art in consideration of functions in an embodiment of the disclosure, but the terms may vary according to the intention of one of ordinary skill in the art, precedents, or new technology in the art. Also, some terms may be arbitrarily selected by the applicant, and in this case, the meaning of the selected terms will be described in detail in the detailed description of an embodiment of the disclosure. Accordingly, the specific terms used herein should be defined based on the unique meanings thereof and the whole context of the disclosure.
The singular forms “a,” “an,” and “the” are intended to include the plural forms as well, unless the context clearly indicates otherwise. The terms used herein, including technical or scientific terms, may have the same meaning as commonly understood by one of ordinary skill in the art described in the disclosure.
When a portion “includes” an element, another element may be further included, rather than excluding the existence of the other element, unless otherwise described. Also, the terms such as “ . . . unit” or “module” refer to units that perform at least one function or operation, and the units may be implemented as hardware or software or as a combination of hardware and software.
According to the situation, the expression “configured to” used herein may be used as, for example, the expression “suitable for”, “having the capacity to”, “designed to”, “adapted to”, “made to”, or “capable of”. The term “configured (or set) to” does not always mean only “specifically designed to” in hardware. Instead, the expression “a system configured to” may mean that the system is “capable of” operating together with another device or other parts. For example, “a processor configured (or set) to perform A, B, and C” may be a dedicated processor (e.g., an embedded processor) for performing a corresponding operation or a generic-purpose processor (such as a central processing unit (CPU) or an application processor) that may perform a corresponding operation by executing at least one software program stored in a memory.
Also, in the disclosure, it will be understood that when elements are “connected” or “coupled” to each other, the elements may be directly connected or coupled to each other, but may alternatively be connected or coupled to each other with an intervening element therebetween, unless specified otherwise.
As used herein, the expressions “at least one of a, b or c” and “at least one of a, b and c” indicate “only a,” “only b,” “only c,” “both a and b,” “both a and c,” “both b and c,” and “all of a, b, and c.”
An embodiment of the disclosure will now be described more fully with reference to the accompanying drawings for one of ordinary skill in the art to be able to perform the embodiment of the disclosure without any difficulty. However, an embodiment of the disclosure may be implemented in many different forms and is not limited to the embodiment described herein. Also, in the drawings, parts irrelevant to the description are omitted in order to clearly describe an embodiment of the disclosure, and like reference numerals denote like elements throughout the disclosure.
Hereinafter, embodiments of the disclosure will be described in detail with reference to the drawings.
Referring to
In an embodiment, the image quality evaluation apparatus 100 and an operating method of the image quality evaluation apparatus according to the disclosure may be an electronic apparatus for evaluating the quality of an image to be evaluated for an NR-IQA method and an operating method of the electronic apparatus. Hereinafter, for convenience of explanation, the image quality evaluation apparatus 100 is referred to as an electronic apparatus 100 for evaluating the quality of an image, and the operating method of the image quality evaluation apparatus 100 is referred to as an operating method of the electronic apparatus 100 for evaluating the quality of an image.
In an embodiment, the electronic apparatus 100 may include a memory 110, at least one processor 130, and a communication interface 140. However, not all of the components shown in
In an embodiment, instructions, data structures, and program code readable by the at least one processor 130 may be stored in the memory 110. In an embodiment, there may be one or more memories 110. In disclosed embodiments, operations performed by the at least one processor 130 may be implemented by executing instructions or code of a program stored in the memory 110.
In an embodiment, the memory 110 may include at least one of a flash memory type, a hard disk type, a multimedia card micro type, a card-type memory (e.g., SD or XD memory), a random-access memory (RAM), a static random-access memory (SRAM), a read-only memory (ROM), an electrically erasable programmable read-only memory (EEPROM), a programmable read-only memory (PROM), a mask ROM, a flash ROM, a hard disk drive (HDD), or a solid-state drive (SSD). Instructions or program code for performing functions or operations of the electronic apparatus 100 may be stored in the memory 110. Instructions, algorithms, data structures, program code, and application programs stored in the memory 110 may be implemented in a programming or scripting language such as C, C++, Java, or assembler.
In an embodiment, an image quality evaluation module 120 that may be used to obtain an image and evaluate the quality of the obtained image may be stored in the memory 110. In an embodiment, the image quality evaluation module 120 may include an image acquisition module 121, a feature map extraction module 122, and a final quality score calculation module 124. The image acquisition module 121, the feature map extraction module 122, and the final quality score calculation module 124 may be stored in the memory 110. However, not all of the modules shown in
In an embodiment, a “module” included in the memory 110 may refer to a unit that processes a function or an operation performed by the at least one processor 130. A “module” included in the memory 110 may be implemented as software such as instructions, algorithms, data structures, or program code.
In an embodiment, the image acquisition module 121 may include instructions or program code related to an operation or a function of obtaining an image. The image acquisition module 121 may include instructions or program code related to an operation or a function of receiving an image from an external server or peripheral electronic apparatuses. The image acquisition module 121 may include an image acquisition model. In an embodiment, the at least one processor 130 may obtain an image that is subjected to quality evaluation, by executing the instructions or the program code of the image acquisition module 121.
In an embodiment, the feature map extraction module 122 may include instructions or program code related to an operation or a function of extracting a feature map including a feature of an image based on the image. In an embodiment, the feature map extraction module 122 may include a revised-shifted window (Swin) transformer model 123. The feature map extraction module 122 may include instructions or program code related to an operation or a function of extracting a feature map from an image by using the revised-Swin transformer model 123. Although the feature map extraction module 122 includes the revised-Swin transformer model in
In an embodiment, the at least one processor 130 may extract a feature map from an image, by executing the instructions or the program code of the feature map extraction module 122. The at least one processor 130 may extract a feature map from an image by using the Swin transformer model or the revised-Swin transformer model included in the feature map extraction module 122. An operation of extracting a feature map from an image by using the Swin transformer model and the revised-Swin transformer model will be described below with reference to
In an embodiment, the final quality score calculation module 124 may include instructions or program code related to an operation or a function of obtaining a final quality score for evaluating the quality of an image based on an extracted feature map. In an embodiment, the at least one processor 130 may obtain a final quality score for evaluating the quality of an image based on an extracted feature map, by executing the instructions or the program code of the final quality score calculation module 124.
In an embodiment, the final quality score calculation module 124 may include a region quality calculation model 125, a region importance calculation model 126, and a final quality calculation model 127.
In an embodiment, the final quality score calculation module 124 may include instructions or program code related to an operation or a function of obtaining a quality score for each reference region of an image based on an extracted feature map, by using the region quality calculation model 125. The at least one processor 130 may obtain a quality score for each reference region of an image based on an extracted feature map, by using the region quality calculation model 125 included in the final quality score calculation module 124. An operation of obtaining a quality score for each reference region of an image based on an extracted feature map by using the region quality calculation model 125 will be described below with reference to
In an embodiment, the final quality score calculation module 124 may include instructions or program code related to an operation or a function of obtaining an importance for each reference region of an image based on an extracted feature map, by using the region importance calculation model 126. The at least one processor 130 may calculate an importance for each reference region of an image based on an extracted feature map, by using the region importance calculation model 126 included in the final quality score calculation module 124. An operation of calculating an importance for each reference region of an image based on an extracted feature map by using the region importance calculation model 126 will be described below with reference to
In an embodiment, the final quality score calculation module 124 may include instructions or program code related to an operation or a function of obtaining a final quality score of an image based on a calculated quality score for each reference region and a calculated importance for each reference region, by using the final quality calculation model 127. The at least one processor 130 may obtain a final quality score of an image based on a quality score for each reference region and an importance for each reference region of the image, by using the final quality calculation model 127 included in the final quality score calculation module 124. An operation of obtaining a final quality score of an image based on a quality score for each reference region and an importance for each reference region of the image by using the final quality calculation model 127 will be described below with reference to
In an embodiment, the image quality evaluation module 120 may include an image quality evaluation model for determining the quality of an image from a final quality score of the image. The image quality evaluation module 120 may include instructions or program code related to an operation or a function of evaluating the quality of an image from a final quality score, by using the image quality evaluation model. The at least one processor 130 may evaluate the quality of an image from a final quality score, by executing the instructions or the program code of the image quality evaluation module 120. The at least one processor 130 may evaluate the quality of an image from a final quality score by using the image quality evaluation model included in the image quality evaluation module 120.
However, the disclosure is not limited thereto, and the image quality evaluation model may be included in the final quality score calculation module 124 or may be included in the final quality calculation model 127.
In an embodiment, the at least one processor 130 may include at least one of, but not limited to, a central processing unit, a microprocessor, a graphics processing unit, an application processor (AP), an application-specific integrated circuit (ASIC), a digital signal processor (DSP), a digital signal processing device (DSPD), a programmable logic device (PLD), a field-programmable gate array (FPGA), a neural processing unit, or a dedicated artificial intelligence (AI) processor designed in a hardware structure specialized for training and processing an AI model.
In an embodiment of the disclosure, when one or more processors included in the at least one processor 130 are AI processors, the AI processors may each be designed to have a hardware structure specialized for processing a specific AI model.
In an embodiment of the disclosure, the at least one processor 130 may include circuitry such as a system on chip (SoC) or an integrated circuit (IC).
In an embodiment of the disclosure, the at least one processor 130 may execute various types of modules stored in the memory 110. The at least one processor 130 may execute at least one instruction constituting the various types of modules stored in the memory 110. The at least one processor 130 may process data according to predefined operation rules or an AI model, by executing the at least one instruction or program stored in the memory 110.
In an embodiment of the disclosure, the at least one processor 130 may execute at least one of the image acquisition module 121, the feature map extraction module 122, or the final quality score calculation module 124 stored in the memory 110.
In an embodiment of the disclosure, the at least one processor 130 may include a plurality of processors.
In an embodiment of the disclosure, at least one of the image acquisition module 121, the feature map extraction module 122, or the final quality score calculation module 124 stored in the memory 110 may be executed by any one of the plurality of processors.
In an embodiment, the communication interface 140 may perform data communication with an external server under the control of the at least one processor 130. Also, the communication interface 140 may perform data communication not only with the external server but also with other peripheral electronic apparatuses. The communication interface 140 may perform data communication with the server or other peripheral electronic apparatuses by using at least one of data communication methods including, for example, wired local area network (LAN), wireless LAN, Wi-Fi, Bluetooth, Zigbee, Wi-Fi Direct (WFD), infrared data association (IrDA), Bluetooth low energy (BLE), near-field communication (NFC), wireless broadband Internet (Wibro), world interoperability for microwave access (WiMAX), shared wireless access protocol (SWAP), wireless gigabit alliance (WiGig), and radio frequency (RF) communication.
In an embodiment, the at least one processor 130 may receive, from the external server or the peripheral electronic apparatuses, an image that is subjected to quality evaluation through the communication interface 140. In an embodiment, the at least one processor 130 may receive, from the external server or the peripheral electronic apparatuses, the Swin transformer model, the revised-Swin transformer model 123, the region quality calculation model 125, the region importance calculation model 126, and the final quality calculation model 127 through the communication interface 140.
Referring to
In an embodiment, the operating method of the electronic apparatus 100 for evaluating the quality of an image may include an operation S200 of extracting a feature map including a feature of the image based on the obtained image. In an embodiment, in the operation S200 of extracting a feature map, the at least one processor 130 may obtain an image by executing instructions or program code of the feature map extraction module 122. The least one processor 130 may extract a feature map including a feature of the image, by inputting the image to the Swin transformer model or the revised-Swin transformer model 123.
In an embodiment, the operating method of the electronic apparatus 100 for evaluating the quality of an image may include an operation S300 of calculating a quality score for each reference region of the image based on the extracted feature map. In an embodiment, in the operation S300 of obtaining a quality score for each reference region of the image, the at least one processor 130 may obtain an image by executing instructions or program code of the final quality score calculation module 124. The at least one processor 130 may calculate a quality score for each reference region of the image, by inputting the feature map to the region quality calculation model 125.
In an embodiment, the operating method of the electronic apparatus 100 for evaluating the quality of an image may include an operation S400 of obtaining an importance for each reference region of the image based on the extracted feature map. In an embodiment, in the step S400 of obtaining an importance for each reference region of the image, the at least one processor 130 may obtain an image by executing the instructions or the program code of the final quality score calculation module 124. The at least one processor 130 may obtain an importance for each reference region of the image, by inputting the feature map to the region importance calculation model 126.
In an embodiment, the operating method of the electronic apparatus 100 for evaluating the quality of an image may include an operation S500 of obtaining a final quality score of the image, based on the quality score for each reference region and the importance for each reference region. In an embodiment, in the step S500 of obtaining a final quality score of the image, the at least one processor 130 may obtain an image by executing the instructions or the program code of the final quality score calculation module 124. The at least one processor 130 may calculate a final quality score of the image, by inputting the quality score for each reference region and the importance for each reference region to the final quality calculation model 127.
In an embodiment, the operating method of the electronic apparatus 100 for evaluating the quality of an image may include an operation S600 of evaluating quality of the image according to the calculated final quality score. In an embodiment, in the operation S600 of evaluating quality of the image, the at least one processor 130 may obtain an image by executing the instructions or the program code of the final quality score calculation module 124. The at least one processor 130 may evaluate the quality of the image, by inputting the calculated final quality score to the image quality evaluation model.
Referring to
In an embodiment, the image 300 may have a resolution including N horizontal pixels and M vertical pixels. In an embodiment, M and N may each be a natural number. In an embodiment, the image 300 may be a color image or a black-and-white image. Also, the image 300 may be an image corresponding to a single frame among a video including a plurality of frames. In an embodiment, the electronic apparatus 100 and the operating method of the electronic apparatus 100 according to the disclosure may also be used to evaluate the quality of a video.
In an embodiment, in a revised-Swin transformer model 310 or a Swin transformer model, a patch including at least one pixel may be used as a unit area. In an embodiment, one patch may include 4 horizontal pixels and 4 vertical pixels. However, the disclosure is not limited thereto, and the number of pixels included in one patch is not limited thereto.
In an embodiment, in a process of extracting a one-dimensional (1D) feature tensor 320 from the image 300 by using the revised-Swin transformer model 310 or the Swin transformer model, the image 300 may be divided into a plurality of patches.
In an embodiment, the image 300 divided into the plurality of patches may have a resolution including P horizontal patches and Q vertical patches. In this case, P may be a value obtained by dividing M by a horizontal size of each patch. Q may be a value obtained by dividing N by a vertical size of each patch. In an embodiment, P and Q may each be a natural number.
In an embodiment, in the operation S200 of extracting a feature map based on the image, the 1 D feature tensor 320 may be extracted from the image 300 input to the revised-Swin transformer model 310. In an embodiment, at least one processor 130 may extract the 1 D feature tensor 320 by inputting the image 300 to the revised-Swin transformer model 310. The 1 D feature tensor 320 may have one axis and a size of
along the one axis. In this case, g may be the number of times patch merging is performed in the revised-Swin transformer model described below with reference to
Although the revised-Swin transformer model 310 is illustrated in
In an embodiment, in the operation S200 of extracting a feature map based on the image, the at least one processor 130 may reshape the 1 D feature tensor 320 extracted from the image 300 by the revised-Swin transformer model 310 into a three-dimensional (3D) feature tensor 330. In an embodiment, the at least one processor 130 may reshape the 1 D feature tensor 320 into the 3D feature tensor 330, in order to calculate a quality score for each reference region and an importance for each reference region of the image described below. In an embodiment, the 3D feature tensor 330 may have three axes, and may have sizes of
and C along the three axes. However, the disclosure is not limited thereto, and the 1D feature tensor 320 and the 3D feature tensor 330 may be extracted from the image 300 by using the Swin transformer model. Hereinafter, for convenience of explanation, the 3D feature tensor 330 may be referred to as a feature map 330.
In an embodiment, in the operation S300 of obtaining a quality score for each reference region of the image based on the feature map 330, a quality score for each reference region of the image may be obtained from the feature map 330 input to a region quality calculation model 340. In an embodiment, the term “quality score” may refer to a score calculated based on a resolution of pixels included in a reference region, a contrast ratio between adjacent pixels, a color, a structural similarity index, a peak signal-to-noise ratio, etc. However, the disclosure is not limited thereto, and a quality score may be set to increase the accuracy of a final evaluation score described below.
In an embodiment, the at least one processor 130 may obtain a quality score for each reference region of the image by inputting the feature map 330 into the region quality calculation model 340. In an embodiment, the obtained quality score for each reference region may be included in a quality score map 350. The at least one processor 130 may generate the quality score map 350 including a plurality of reference regions, by inputting the feature map 330 to the region quality calculation model 340. In an embodiment, the quality score for each reference region calculated by the region quality calculation model 340 may be included in each of the plurality of reference regions in the quality score map 350.
In an embodiment, the term “reference region” of an image may refer to a region corresponding to a unit area constituting the feature map 330. In an embodiment, the feature map 330 may be divided into a plurality of patches by the revised-Swin transformer model 310. Each of the plurality of patches may be a unit area constituting the feature map 330. In an embodiment, the term “reference region” of an image may refer to a region corresponding to each of a plurality of patches.
In an embodiment, when the quality score map 350 is divided into
horizontal reference regions and
vertical reference regions, a quality score for each reference region included in each of the reference regions may be u(i,j). In this case, j may be a natural number between 1 and
i may be a natural number between 1 and
An operation of obtaining a quality score for each reference region by using the region quality calculation model 340 and the quality score map 350 will be described below with reference to
In an embodiment, the region quality calculation model 340 may include an AI model. In an embodiment, the AI model included in the region quality calculation model 340 may include a machine learning model or a deep learning model. In an embodiment, the AI model included in the region quality calculation model 340 may be an AI model that receives the feature map 330 as an input and is trained to calculate a quality score for each reference region of the image and infer the quality score map 350.
In an embodiment, in the operation S400 of obtaining an importance for each reference region of the image based on the feature map 330, an importance for each reference region of the image may be obtained from the feature map 330 input to a region importance calculation model 360. In an embodiment, the term “importance” may refer to a degree of importance of a corresponding reference region in an image determined according to a preset criterion. In an embodiment, an importance of an object may be set to be greater than an importance of a background. In an embodiment, an importance of an object having a large area may be set to be greater than an importance of an object having a small area. In an embodiment, an importance of an area having a large gray scale may be set to be greater than an importance of an area having a small gray scale. However, the disclosure is not limited thereto, and an importance may be set to increase the accuracy of a final evaluation score described below.
In an embodiment, the at least one processor 130 may obtain an importance for each reference region of the image by inputting the feature map 330 to the region importance calculation model 360. In an embodiment, the obtained importance for each reference region may be included in an importance map 370. The at least one processor 130 may generate the importance map 370 including a plurality of reference regions, by inputting the feature map 330 to the region importance calculation model 360. In an embodiment, the importance for each reference region calculated by the region importance calculation model 360 may be included in each of the plurality of reference regions in the importance map 370.
In an embodiment, when the quality score map 350 is divided into
horizontal reference regions and
vertical reference regions, a quality score for each reference region included in each of the reference regions may be v(i,j). In this case, j may be a natural number between 1 and
i may be a natural number between 1 and
An operation of obtaining an importance for each reference region by using the region importance calculation model 360 and the importance map 370 will be described below with reference to
In an embodiment, the region importance calculation model 360 may include an AI model. In an embodiment, the AI model included in the region importance calculation model 360 may include a machine learning model or a deep learning model. In an embodiment, the AI model included in the region importance calculation model 360 may be an AI model that receives the feature map 330 as an input and is trained to calculate an importance for each reference region of the image and infer the importance map 370.
In an embodiment, examples of the AI model included in the region quality calculation model 340 and the AI model included in the region importance calculation model 360 may include a convolutional neural network (CNN), a deep neural network (DNN), a recurrent neural network (RNN), a restricted Boltzmann machine (RBM), a deep belief network (DBN), a bidirectional recurrent deep neural network (BRDNN), a deep Q-network, a generative adversarial network (GAN), a conditional generative adversarial network (CGAN), a deep convolutional generative adversarial network (DCGAN), and a variational auto encoder (VAE), and the AI model included in the region quality calculation model 340 and the AI model included in the region importance calculation model 360 are not limited to the above examples.
In an embodiment, the electronic apparatus 100 may train the AI model included in each of the revised-Swin transformer model 310, the region quality calculation model 340, and the region importance calculation model 360. In order to train the AI model included in each of the revised-Swin transformer model 310, the region quality calculation model 340, and the region importance calculation model 360, the electronic apparatus 100 may perform transfer learning using a pre-trained model. However, the disclosure is not limited thereto, and the electronic apparatus 100 may receive a trained AI model from an external server or peripheral electronic apparatuses through the communication interface 140. A method of training the AI model included in each of the revised-Swin transformer model 310, the region quality calculation model 340, and the region importance calculation model 360 will be described below with reference to
In an embodiment, in the operation S500 of obtaining a final quality score based on the quality score and the importance, a final quality score of the image may be obtained from the quality score map 350 and the importance map 370 input to a final quality calculation model 380. In an embodiment, the at least one processor 130 may obtain a final quality score of the image by inputting the quality score map 350 and the importance map 370 to the final quality calculation model 380. In an embodiment, the final quality calculation model 380 may calculate a final quality score of the image based on the quality score for each reference region included in each of the plurality of reference regions of the quality score map 350 and the importance for each reference region included in each of the plurality of reference regions of the importance map 370. The final quality calculation model 380 may obtain a final quality score of the image based on the quality score for each reference region and the importance for each reference region by using Equation 1.
In this case, S is a final quality score of the image, g is the number of times patch merging is performed in the revised-Swin transformer model or the Swin transformer model, P is a value obtained by dividing a horizontal resolution of the image by a horizontal size of one patch, Q is a value obtained by dividing a vertical resolution of the image by a vertical size of one patch, u is the quality score for each reference region, and v is the importance for each region.
In an embodiment, the at least one processor 130 may evaluate quality of the image based on the obtained final quality score. In an embodiment, the term “final quality score” may refer to a score calculated based on a resolution of an image, a contrast ratio of the image, a sharpness, a color, a structural similarity index, a peak signal-to-noise ratio, locations of an object and a background included in the image, a degree of distinction between the object and the background, etc. In an embodiment, the at least one processor 130 may evaluate that as the final quality score increases, the quality of the image increases. In an embodiment, when the final quality score is equal to or higher than a preset reference score, the at least one processor 130 may evaluate that the input image has a high quality suitable to be provided to a user or the quality of the converted image is high.
In an embodiment, the final quality score may be obtained based on the quality score obtained for each reference region of the image and the importance obtained for each reference region of the image.
In an embodiment, assuming that a final quality score is obtained through one operation of calculating an average of an entire image, in a case where there is a large quality score deviation for each region included in the image or there is a large importance deviation for each region, the final quality score may not appropriately reflect a change in the quality of the image according to the above case Also, because training is performed based only on the final quality score obtained according to a training image in a process of training the AI model included in each of the revised-Swin transformer model 310, the region quality calculation model 340, and the region importance calculation model 360, overfitting for the training image may occur and performance during an inference operation may be lowered.
In contrast, assuming that the final quality score is obtained based on the quality score obtained for each reference region of the image and the importance obtained for each reference region of the image according to the disclosure, in a case where there is a large quality score deviation for each region included in the image or there is a large importance deviation for each region in the image, the final quality score may be obtained by reflecting a change in the quality of the image according to the above case. Also, in a process of training the AI included in each of the revised-Swin transformer model 310, the region quality calculation model 340, and the region importance calculation model 360 described below, because training may be performed based on information about the quality score calculated for reach reference region of the image and the importance calculated for each reference region of the image at a low level calculated according to a training image, overfitting may be prevented.
In an embodiment, the electronic apparatus 100 for evaluating the quality of an image and the operating method of the electronic apparatus 100 may be used to evaluate the quality of an image in a display apparatus that provides an image to a user or an electronic apparatus that performs image conversion. In an embodiment, when the final quality score is lower than the reference score, the at least one processor 130 may evaluate that the input image has a low quality that is unsuitable for the user or the quality of the converted image is low. Embodiments of using the electronic apparatus 100 for evaluating the quality of an image and the operating method of the electronic apparatus 100 will be described below with reference to
The same elements as those described with reference to
Referring to
In an embodiment, the revised-Swin transformer model 310 may include a plurality of revised-Swin transformer blocks (e.g., 430, 450, 470, and 490). In an embodiment, the image 300 input the revised-Swin transformer model 310 may sequentially pass through the plurality of revised-Swin transformer blocks (e.g., 430, 450, 470, and 490). In this case, the same operation may be performed on an input tensor in each of the plurality of revised-Swin transformer blocks (e.g., 430, 450, 470, and 490). In
In an embodiment, linear embedding 420 may be performed on the partitioned patches. In an embodiment, a channel of a tensor including the partitioned patches may be converted through the linear embedding. In this case, the channel of the tensor converted through the linear embedding may vary according to a size of the revised-Swin transformer model 310.
Referring to
In an embodiment, layer normalization 433 may be performed on the tensor on which the linear embedding has been performed in the first block 431. In an embodiment, window multi-head self-attention (W-MSA) 434 may be performed on the tensor on which the layer normalization has been performed. Compared to multi-head self-attention (MSA) involving performing self-attention on the entire tensor in multiple heads, the W-MSA involves performing self-attention in units of windows 500, thereby reducing the complexity of calculation. The window 500 will be described below with reference to
In an embodiment, the tensor on which the W-MSA 434 has been performed may undergo layer normalization and multi-layer perception (MLP) 435. The tensor that is fully connected through the MLP 435 may be provided to the second block 432. In an embodiment, an activation function in the MLP 435 may be a Gaussian error linear unit (GELU).
In an embodiment, in the second block 432, the tensor on which the layer normalization has been performed again may undergo shifted window multi-head self-attention (SW-MSA) 436. The SW-MSA may involve cyclically shifting the window 500 and performing W-MSA so that adjacent windows have connectivity. Accordingly, self-attention may be performed through the SW-MSA even on patches included in a boundary between adjacent windows. Thereafter, the cyclically shifted window may be returned through reverse cyclic shift.
In an embodiment, in the second block 432, the tensor on which the SW-MSA has been performed may undergo layer normalization and then may undergo MLP. In an embodiment, residual connection may be applied after a module for performing W-MSA and a module for performing MLP included in the first block 431. Residual connection may be applied after a module for performing SW-MSA and a module for performing MLP of the second block 432.
In an embodiment, the tensor on which an operation has been performed by the first revised-Swin transformer block 430, through the first block 431 and the second block 432, may be output.
Referring back to
In an embodiment, an operation may be performed through the second revised-Swin transformer block 450 on the tensor on which the patch merging 440 has been performed. In an embodiment, the operation of the second revised-Swin transformer block 450 may be performed in the same manner as the operation in the first revised-Swin transformer block 430.
In an embodiment, patch merging 460 may be performed on a plurality of second patches included in the tensor passing through the second revised-Swin transformer block 450. In an embodiment, when a patch obtained by merging the second patches 511 is referred to as a third patch 521, the third patch 521 may be generated by combining a plurality of second patches. In an embodiment, the number of pixels included in the third patch 521 may be greater than the number of pixels included in the second patch 511. Horizontal and vertical sizes of the third patch 521 may be greater than horizontal and vertical sizes of the second patch 511.
In an embodiment, an operation may be performed through the third revised-Swin transformer block 470 on the tensor on which the patch merging 440 has been performed. In an embodiment, the operation of the third revised-Swin transformer block 470 may be performed in the same manner as the operation in the first revised-Swin transformer block 430.
In an embodiment, patch merging 480 may be performed on a plurality of third patches included in the tensor passing through the third revised-Swin transformer block 470. In an embodiment, when a patch obtained by merging the third patches 521 is a fourth patch, the fourth patch may be generated by combining a plurality of second patches. In an embodiment, the number of pixels included in the fourth patch may be greater than the number of pixels included in the third patch 521. Horizontal and vertical sizes of the fourth patch may be greater than horizontal and vertical sizes of the third patch 521.
In an embodiment, an operation may be performed through the fourth revised-Swin transformer block 490 on the tensor on which the patch merging 480 has been performed. In an embodiment, the operation of the fourth revised-Swin transformer block 490 may be performed in the same manner as the operation in the first revised-Swin transformer block 430.
In an embodiment, the 1 D feature tensor 320 may be extracted from the input image 300, through the first to fourth revised-Swin transformer blocks 430. In an embodiment, because patch merging is performed and then an operation is performed through each of revised-Swin transformer blocks, the 1 D feature tensor 320 may be extracted through the revised-Swin transformer model 310 having a hierarchical structure.
Referring to
In an embodiment, the revised-Swin transformer model may be a model for performing a Swin transformer operation by using the window 620 having a resolution less than a resolution of a window 600 used in the Swin transformer model. Hereinafter, an operation of the Swin transformer model will be described with reference to
Referring to
In an embodiment, windows used in the first to fourth revised-Swin transformer blocks 430, 450, 470, and 490 may have the same resolution. In an embodiment, a window used in the first revised-Swin transformer block 430 may be referred to as a first window 500, a window used in the second revised-Swin transformer block 450 may be referred to as a second window 510, a window used in the third revised-Swin transformer block 470 may be referred to as a third window 520, and a window used in the fourth revised-Swin transformer block 490 may be referred to as a fourth window. In an embodiment, the first window 500, the second window 510, the third window 520, and the fourth window may have the same resolution.
In an embodiment, as patch merging is performed after an operation is performed through each of the first to third revised-Swin transformer blocks 430, 450, and 470, a size of each patch constituting an image may increase and a resolution of the image including the patch may decrease.
In this case, a resolution of the image including the patch while an operation is performed in the revised-Swin transformer model 310 may be determined according to a resolution of the image 300 input to the revised-Swin transformer model 310.
In an embodiment, one patch may include 4 horizontal pixels and 4 vertical pixels, and during patch merging, four adjacent patches may be merged into one patch. In this case, when the image 300 having a resolution including 128 horizontal pixels and 128 vertical pixels is input, the image in the first revised-Swin transformer block 430 may have a resolution of 32 horizontal patches and 32 vertical patches. The image in the second revised-Swin transformer block 450 may have a resolution of 16 horizontal patches and 16 vertical patches. The image in the third revised-Swin transformer block 470 may have a resolution of 8 horizontal patches and 8 vertical patches. The image in the fourth revised-Swin transformer block 490 may have a resolution of 4 horizontal patches and 4 vertical patches.
In an embodiment, when the image 300 having a resolution of 224 horizontal pixels and 224 vertical pixels is input, the image in the first revised-Swin transformer block 430 may have a resolution of 56 horizontal patches and 56 vertical patches. The image in the second revised-Swin transformer block 450 may have a resolution of 28 horizontal patches and 28 vertical patches. The image in the third revised-Swin transformer block 470 may have a resolution of 14 horizontal patches and 14 vertical patches. The image in the fourth revised-Swin transformer block 490 may have a resolution of 7 horizontal patches and 7 vertical patches.
In an embodiment, the window 600 used in the Swin transformer model may have a resolution of 7 horizontal patches and 7 vertical patches. In an embodiment, the Swin transformer model may be a model trained to perform self-attention by using a window having a resolution of 7 horizontal patches and 7 vertical patches in each Swin transformer block in which patch merging is performed.
In an embodiment, when an image having a resolution of 224 horizontal pixels and 224 vertical pixels is input to the revised-Swin transformer model 310 and the window 600 used in the conventional Swin transformer model is used, because the image in the fourth revised-Swin transformer block 490 having a largest patch size has a resolution of 7 horizontal patches and 7 vertical patches through patch merging, the entire image may be included in one window. Accordingly, self-attention may be performed in 7 horizontal patches and 7 vertical patches included in one window, and a 1 D feature tensor may be inferred based on a weight of a trained model.
However, when an image having a resolution of 128 horizontal pixels and 128 vertical pixels is input to the revised-Swin transformer model 310, because the image in the fourth revised-Swin transformer block 490 has a resolution of 4 horizontal patches and 4 vertical patches, self-attention is performed in 4 horizontal patches and 4 vertical patches in one window. In this case, because self-attention is performed by using a window 610 having a resolution of 4 horizontal patches and 4 vertical patches unlike the trained model, an unintended 1 D feature tensor may be inferred.
To prevent this, in the case of the Swin transformer model, an input image may be resized or cropped to have a resolution corresponding to a resolution of an image input as training data during training.
The revised-Swin transformer model 310 of the disclosure may be a model trained to perform self-attention by using a window having a resolution of 4 horizontal patches and 4 vertical patches in each revised-Swin transformer block in which patch merging is performed. In this case, even when not only an image having a resolution of 224 horizontal pixels and 224 vertical pixels but also an image having a resolution of 128 horizontal pixels and 128 vertical pixels are input to the revised-Swin transformer model 310, self-attention may be performed by using the window 620 having a resolution of 4 horizontal patches and 4 vertical patches, like the trained model, to infer a 1 D feature tensor. Accordingly, in the case of the revised-Swin transformer model 310, an image having any of various resolutions may be provided as an input without having to resize or crop the input image to have a resolution corresponding to a resolution of an image input as training data during training.
In an embodiment, although the Swin transformer model performs self-attention by using a window having a resolution of 7 horizontal patches and 7 vertical patches and the revised-Swin transformer model performs self-attention by using a window having a resolution of 4 horizontal patches and 4 vertical patches, the disclosure is not limited thereto. The revised-Swin transformer model may be a model trained by using the window 620 having a sufficiently low resolution so that a feature map is extracted by the window 620 having the same resolution, regardless of a resolution of the input image 300. In this case, the sufficiently low resolution may be determined according to the purpose of use of the electronic apparatus 100, the target of use, etc. A resolution of a window may be determined based on a resolution of an image that is pre-predicted to be input to the electronic apparatus 100, the number of revised-Swin transformer blocks included in the revised-Swin transformer model, the number of times patch merging is performed, a patch size during patch partition, etc.
The electronic apparatus 100 of the disclosure may use a 1 D feature tensor extracted by using the Swin transformer model or the 1 D feature tensor 320 extracted by using the revised-Swin transformer model 310. For convenience of explanation, the following will be described assuming that the 1 D feature tensor 320 is extracted by using the revised-Swin transformer model 310.
Referring to
In an embodiment, the first filter may be a filter for extracting a feature indicating the quality of the image 300 through a convolution operation from the feature map 330. In an embodiment, a weight included in the first filter may be a weight trained to extract a feature indicating the quality of the image 300 through a convolution operation from the feature map 330. A weight included in the first filter may be a value updated according to a loss function based on a ground truth score obtained by evaluating the image 300 described below and a final quality score of the image 300 calculated through the electronic apparatus 100.
In an embodiment, as a convolution operation is performed by using the first filter having a resolution corresponding to each of the plurality of reference regions included in the image 300, a feature indicating the quality of each of the plurality of reference regions of the image 300 may be extracted. In an embodiment, the plurality of reference regions may respectively correspond to a plurality of patches divided in the revised-Swin transformer model 310. As a convolution operation is performed by using the first filter in the region quality calculation model 340, a feature indicating the quality corresponding to each of the plurality of patches may be extracted.
In an embodiment, a convolution operation may be performed in the first convolution layer 341 so that the extracted feature has one channel. Accordingly, the feature map 330 having C channels may be extracted as a feature map having one channel.
In an embodiment, in the region quality calculation model 340, an activation operation is performed in a first activation layer 342 on a tensor extracted by performing a convolution operation through the first convolution layer 341. In an embodiment, an activation operation using, for example, but not limited to, a Sigmoid function, a Sigmoid function and a linear normalization function, or a rectified linear unit (ReLU) function and a linear normalization function or a softmax normalization function may be performed in the first activation layer 342.
In an embodiment, the feature map 330 input to the region quality calculation model 340 may be extracted as the quality score map 350 through the first convolution layer 341 and the first activation layer 342. In an embodiment, at least one processor 130 may input the feature map 330 to the region quality calculation model 340 and may generate the quality score map 350 through the first convolution layer 341 and the first activation layer 342. In an embodiment, the quality score map 350 may be divided into a plurality of reference regions, and a quality score of the image 300 may be included in each of the plurality of reference regions.
Referring to
In an embodiment, a convolution operation may be performed on the feature map 330 by using a second filter in the second convolution layer 361. In an embodiment, a resolution of the second filter may correspond to two or more of a plurality of reference regions included in the image 300. A resolution of the second filter may correspond to two or more of a plurality of patches included in the feature map 330. A size of the second filter may correspond to a size of two or more of the plurality of patches included in the feature map 330. A convolution operation may be performed on two or more of the plurality of patches included in the feature map 330 by using the second filter in the second convolution layer 361.
In an embodiment, the second filter may be a filter for extracting a feature indicating an importance for each reference region of the image 300, through a convolution operation from the feature map 330. In an embodiment, a weight included in the second filter may be a weight trained to extract a feature indicating an importance for each reference region of the image 300 through a convolution operation from the feature map 330. A weight included in the second filter may be a value updated according to a loss function based on a ground truth score obtained by evaluating the image 300 described below and a final quality score of the image 300 calculated through the electronic apparatus 100.
In an embodiment, a convolution operation may be performed on the feature map 330 by using a first sub-filter in the first sub-convolution layer 362. A convolution operation may be performed on the feature map 330 by using a second sub-filter in the second sub-convolution layer 363. A convolution operation may be performed on the feature map 330 by using a third sub-filter in the third sub-convolution layer 364. In an embodiment, a weight included in each of the first sub-filter, the second sub-filter, and the third sub-filter may be a value updated according to a loss function based on a ground truth score obtained by evaluating the image 300 and a final quality score of the image 300 calculated through the electronic apparatus 100.
In an embodiment, as a convolution operation is performed by using the second filter having a resolution corresponding to two or more of the plurality of reference regions included in the image 300, a feature indicating an importance between the two or more reference regions corresponding to the second filter may be extracted. As a convolution operation is performed on the feature map 330 by using the second filter having a resolution corresponding to two or more reference regions, an importance of each reference region may be extracted from a relationship with other reference regions.
In an embodiment, the first sub-filter, the second sub-filter, and the third sub-filter may each be a filter for performing dilated convolution. In an embodiment, dilation rates of the first sub-filter, the second sub-filter, and the third sub-filter may be different from each other. In an embodiment, the term “dilation rate” may refer to an interval between reference regions having a weight included in each sub-filter. In an embodiment, a dilation rate of the second sub-filter may be greater than a dilation rate of the first sub-filter. A dilation rate of the third sub-filter may be greater than a dilation rate of the second sub-filter. In an embodiment, a dilation rate of the first sub-filter may be 4. A dilation rate of the second sub-filter may be 8. A dilation rate of the third sub-filter may be 16.
In an embodiment, referring to
In an embodiment, a receptive field of the second sub-filter may be wider than a receptive field of the first sub-filter. A receptive field of the second sub-filter may be wider than a receptive field of the first sub-filter. A receptive field of the third sub-filter may be wider than a receptive field of the second sub-filter.
In an embodiment, as a convolution operation is performed by using sub-filters having different receptive fields, an importance for each reference region considering adjacent reference regions from among the plurality of reference regions included in the image 300 and an importance for each reference region considering reference regions spaced apart from each other from among the plurality of reference regions may be extracted. In an embodiment, when the image 300 includes an object and a background, an importance for each reference region in the object and an importance for each reference region in the background may be extracted by performing a convolution operation by using a sub-filter having a small receptive field. Also, an importance for each reference region in an area including the background and the object may be extracted by performing a convolution operation by using a sub-filter having a wide receptive field.
In an embodiment, a convolution operation may be performed in the second convolution layer 361 so that the extracted feature has 1024 channels. Accordingly, the feature map 330 having C channels may be extracted as a feature map having 1024 channels. In an embodiment, a convolution operation may be performed in the first sub-convolution layer 362 so that the extracted feature has 1024 channels. A convolution operation may be performed in the second sub-convolution layer 363 so that the extracted feature has 1024 channels. A convolution operation may be performed in the third sub-convolution layer 364 so that the extracted feature has 1024 channels.
In an embodiment, a concatenation operation may be performed in a connection layer 365 on tensors extracted through the first sub-convolution layer 362, the second sub-convolution layer 363, and the third sub-convolution layer 364.
In an embodiment, a convolution operation may be performed by using a third filter in a third convolution layer 366 on the tensors connected through the connection layer 365. A resolution of the third filter may correspond to each of the plurality of reference regions included in the image 300. A resolution of the third filter may correspond to each of a plurality of patches included in the tensor on which a concatenation operation has been performed. In an embodiment, a size of the third filter may correspond to a size of each of the plurality of patches included in the feature map 330. A convolution operation may be performed on each of the plurality of patches included in the tensor on which a concatenation operation has been performed by using the third filter in the third convolution layer 366.
In an embodiment, a convolution operation may be performed in the third convolution layer 366 so that the extracted feature has one channel. Accordingly, the feature map 330 having 3072 channels may be extracted as a feature map having one channel.
In an embodiment, in the region importance calculation model 360, an activation operation may be performed in a second activation layer 367 on the tensor extracted by performing a convolution operation through the third convolution layer 366. In an embodiment, an activation operation using, for example, but not limited to, a Sigmoid function, a Sigmoid function and a linear normalization function, or a ReLU function and a linear normalization function or a softmax normalization function may be performed in the second activation layer 367.
In an embodiment, the feature map 330 input to the region importance calculation model 360 may be extracted as the importance map 370 through the second convolution layer 361, the connection layer 365, the third convolution layer 366, and the second activation layer 367. In an embodiment, at least one processor 130 may input the feature map 330 to the region importance calculation model 360 and may generate the importance map 370 through the second convolution layer 361, the connection layer 365, the third convolution layer 366, and the second activation layer 367. In an embodiment, the importance map 370 may be divided into a plurality of reference regions, and an importance of the image 300 may be included in each of the plurality of reference regions.
Referring to
In an embodiment, a weight of the AI model included in each of the revised-Swin transformer model 310, the region quality calculation model 340, and the region importance calculation model 360 may be updated by using a loss function based on a ground truth score 1010 pre-evaluated for the input image 300 and a final quality score calculated in the final quality calculation model 380. In this case, the pre-evaluated ground truth score 1010 may be a score obtained when image quality evaluation experts evaluate the quality of the input image 300 for training of the electronic apparatus 100 or may be a score obtained when the quality of the input image 300 is evaluated by using a pre-used evaluation method.
In an embodiment, a loss function used in the loss calculation model 1000 is as shown in Equation 2 and Equation 3 below.
In this case, x may be a final quality score calculated in the final quality calculation model 380, y may be a pre-evaluated ground truth score, LOSS may be a loss function, B may be the number of pieces of data included in a mini-batch where training is performed,
In an embodiment, as the loss calculation model 1000 uses the loss function using Equation 2 and 3, unnecessary training may be prevented in improving the accuracy and performance of image quality evaluation.
In an embodiment, the accuracy and performance of image quality evaluation may be determined based on an indicator such as a Pearson's linear correlation coefficient (PLCC) or a Spearman's rank order correlation coefficient (SRCC) In an embodiment, the electronic apparatus 100 of the disclosure may be trained to calculate and infer a final quality score by evaluating the quality of the input image 300, based on an evaluation indicator such as a PLCC or an SRCC.
In this case, when the AI model included in each of the revised Sin transformer model 310, the region quality calculation model 340, and the region importance calculation model 360 is trained with a conventional loss function simply according to a difference between a final quality score and a ground truth score, unnecessary training may be repeated to reduce a difference between an average or a standard deviation of the ground truth value and an average or a standard deviation of the final quality score, regardless of improving performance according to an evaluation indicator such as an PLCC or an SRCC.
In an embodiment, as the electronic apparatus 100 of the disclosure uses the loss function using Equation 2 and Equation 3, unnecessary training for reducing a difference between an average or a standard deviation of a ground truth and an average or a standard deviation of a final quality score may be prevented from being repeated. Accordingly, a time required to train the AI model included in each of the revised-Swin transformer model 310, the region quality calculation model 340, and the region importance calculation model 360 may be reduced.
Also, an operation of training the AI model included in each of the revised-Swin transformer model 310, the region quality calculation model 340, and the region importance calculation model 360 may be stabilized by including, which is the regularization coefficient, in the loss function. In an embodiment, a may be set to a value close to 1 and used.
In an embodiment, a weight of the AI model included in each of the revised-Swin transformer model 310, the region quality calculation model 340, and the region importance calculation model 360 may be updated to minimize the loss function. The electronic apparatus 100 may update a weight of the AI model included in each of the revised-Swin transformer model 310, the region quality calculation model 340, and the region importance calculation model 360 to minimize the loss function, by using a forward propagation algorithm and a back propagation algorithm.
Referring to
In an embodiment, the display apparatus 1100 may include a display 1110, a memory 1120, and at least one processor 1140. However, not all of the components of the display apparatus 1100 shown in
In an embodiment, the display 1110, the memory 1120, and the at least one processor 1140 may be electrically and/or physically connected to each other.
In an embodiment, the display 1110 may include any one of a liquid crystal display, a plasma display, an organic light-emitting diode display, and an inorganic light-emitting diode display. However, the disclosure is not limited thereto, and the display 1110 may include another type of display capable of providing an image to a user using the display apparatus 1100.
In an embodiment, an image quality evaluation module 1130 for evaluating the quality of an image may be stored in the memory 1120. In an embodiment, the image quality evaluation module 1130 may be a module described with reference to
In an embodiment, the at least one processor 1140 may control an operation of a display apparatus 1100. The at least one processor 1140 may evaluate the quality of an image used in the display apparatus 1100, by executing instructions or program code of the image quality evaluation module 1130.
In an embodiment, before an image received from an external server or an electronic apparatus is displayed on the display 1110, the at least one processor 1140 may evaluate the quality of the received image by using the image quality evaluation module 1130. When it is determined that the evaluated image has an appropriate quality to be provided to a user using the display apparatus 1100, the at least one processor 1140 may provide the image to the user through the display 1110.
In an embodiment, when it is determined that the image does not have an appropriate quality to be provided to the user using the display apparatus 1100, the at least one processor 1140 may provide the image to the user along with a notification indicating that the quality of the provided image is not appropriate (e.g., “the quality of the image is poor” or “a low-resolution image is displayed”). Also, the at least one processor 1140 may improve the quality of the provided image by using a module for improving the quality of an image such as up-scaling using interpolation or the like and then may provide the image to the user.
Referring to
In an embodiment, the electronic apparatus 1200 of
In an embodiment, the electronic apparatus 1200 may include a memory 1210, at least one processor 1240, and a communication interface 1250. However, not all of the components of the electronic apparatus 1200 shown in
In an embodiment, an image conversion module 1220 for converting the quality, size, and shape of an image and an image quality evaluation module 1230 for evaluating the quality of an image may be stored in the memory 1210. In an embodiment, the image quality evaluation module 1230 may be a module described with reference to
In an embodiment, the at least one processor 1240 may control an operation of the electronic apparatus 1200. The at least one processor 1240 may convert the quality, size, and shape of a received image, by executing instructions or program code of the image conversion module 1220.
The at least one processor 1240 may evaluate the quality of an image provided from the electronic apparatus 1200, by executing instructions or program code of the image quality evaluation module 1230.
In an embodiment, the at least one processor 1140 may convert the quality, size, and shape of an image received from an external server or an electronic apparatus through the communication interface 1250 by using the image conversion module 1220 and may evaluate the quality of the converted image by using the image quality evaluation module 1130. When it is determined that the converted image has an appropriate quality compared to a pre-set determination criterion, the at least one processor 1140 may provide the converted image to the external server or the other electronic apparatus through the communication interface 1250.
When it is determined that the converted image has an inappropriate quality compared to a preset determination criterion, the at least one processor 1140 may convert the quality, size, and shape of the converted image again by using the image conversion module 1220 or may provide information indicating that the image received from the external server or the other electronic apparatus has an inappropriate quality. Also, the at least one processor 1140 may provide the converted image along with information indicating that the converted image has an inappropriate quality to the external server or the other electronic apparatus.
According to an embodiment of the disclosure, there may be provided a computer-readable recording medium having recorded thereon a program for performing at least one of embodiments of the operating method on a computer.
A program executed by an electronic apparatus described in the disclosure may be implemented as a hardware component, a software component, and/or a combination of a hardware component and a software component. The program may be executed by any system capable of executing computer-readable instructions.
The software may include a computer program, code, instructions, or a combination of one or more thereof, and may configure a processing device to operate as desired or instruct the processing device independently or collectively.
The software may be implemented as a computer program including instructions stored in a computer-readable storage medium. Examples of the computer-readable recording medium include a magnetic storage medium (e.g., a read-only memory (ROM), a random-access memory (RAM), a floppy disk, or a hard disk) and an optical recording medium (e.g., a compact disc ROM (CD-ROM), or a digital versatile disc (DVD)). The computer-readable recording medium may be distributed in computer systems connected in a network so that computer-readable code is stored and executed in a distributed fashion. The recording medium may be computer-readable, may be stored in a memory, and may be executed by a processor.
The computer-readable storage medium may be provided in the form of a non-transitory storage medium. Here, “non-transitory” means that the storage medium does not include a signal (e.g., an electromagnetic wave) and is tangible, but does not distinguish whether data is stored semi-permanently or temporarily in the storage medium. For example, the “non-transitory storage medium” may include a buffer in which data is temporarily stored.
Also, a program according to embodiments of the disclosure may be provided in a computer program product. The computer program product is a product purchasable between a seller and a purchaser.
The computer program product may include a software program and a computer-readable storage medium in which the software program is stored. For example, the computer program product may include a product (e.g., a downloadable application) that is electronically distributed as a software program through an electronic market (e.g., Samsung Galaxy Store) or a manufacturer of an electronic apparatus. For electronic distribution, at least a part of the software program may be stored in a storage medium or may be temporarily generated. In this case, the storage medium may be a storage medium of a server of the manufacturer of the electronic apparatus, a server of the electronic market, or a relay server that temporarily stores the software program.
Although the embodiments have been described by the limited embodiments and the drawings as described above, various modifications and variations are possible by one of ordinary skill in the art from the above description. For example, the described techniques may be performed in a different order from the described method, and/or the described elements such as a computer system and a module may be combined or integrated in a different form from the described method, or may be replaced or substituted by other elements or equivalents to achieve appropriate results.
Number | Date | Country | Kind |
---|---|---|---|
10-2023-0030046 | Mar 2023 | KR | national |
10-2023-0082127 | Jun 2023 | KR | national |
This application is a by-pass continuation of International Application No. PCT/KR2024/095318, filed on Feb. 16, 2024, which is based on and claims priority to Korean Patent Application No. 10-2023-0030046, filed on Mar. 7, 2023, in the Korean Intellectual Property Office, and Korean Patent Application No. 10-2023-0082127, filed on Jun. 26, 2023, in the Korean Intellectual Property Office, the disclosures of which are incorporated by reference herein in their entireties.
Number | Date | Country | |
---|---|---|---|
Parent | PCT/KR2024/095318 | Feb 2024 | WO |
Child | 18781512 | US |