The disclosure relates in general to a neural network model fusion method and an electronic device using the same.
Image recognition technology has a wide range of application. Particularly, application fields, such as medical image analysis, information security control and crime investigation, require the use of image recognition technology with high accuracy.
Take medical image analysis for example. Currently, the interpretation of medical images employs a severity classification model, but the accuracy is usually not high enough. It would be a great benefit to the patients if the image recognition technology can provide higher accuracy. Therefore, the research personnel in both the medical and the engineering fields are dedicated to the above regard.
The disclosure is directed to a neural network model fusion method and an electronic device using the same.
According to one embodiment, a neural network model fusion method is provided. The neural network model fusion method includes the following steps. An image is received. The image is analyzed through several neural network models. The neural network models include at least two of a degree classification model, a multi-class identification model and an object detection model. Several analysis results are obtained according to the neural network models. These analysis results are converted into a number of conversion factors. The conversion factors are inputted into a fusion model to obtain a fusion result.
According to another embodiment, an electronic device is provided. The electronic device includes a processor configured to perform a neural network model fusion method. The neural network model fusion method includes the following steps. An image is received. The image is analyzed through several neural network models. The neural network models include at least two of a degree classification model, a multi-class identification model and an object detection model. Several analysis results are obtained according to the neural network models. These analysis results are converted into a number of conversion factors. The conversion factors are inputted into a fusion model to obtain a fusion result.
The above and other aspects of the invention will become better understood with regard to the following detailed description of the preferred but non-limiting embodiment(s). The following description is made with reference to the accompanying drawings.
In the following detailed description, for purposes of explanation, numerous specific details are set forth in order to provide a thorough understanding of the disclosed embodiments. It will be apparent, however, that one or more embodiments may be practiced without these specific details. In other instances, well-known structures and devices are schematically shown in order to simplify the drawing.
Various embodiments of the neural network model fusion method are disclosed below. The neural network models for different purposes, such as the degree classification model (for example, configured to classify disease severity), the multi-class identification model (for example, configured to perform multi-lesion classification for various types of lesions) and the object detection model (for example, configured to detect lesion locations), are fused through machine learning to effectively increase the performance in image recognition. Particularly, when it comes to the interpretation of medical images, the neural network model fusion method of the present disclosure effectively increases the accuracy in the classification of disease severity and the identification of lesions.
Referring to
To put it in greater details, the analysis results AR obtained by the degree classification model MD1 from the analysis of the image block B0 of the image P0 are presented as a distribution probability matrix MX11 of levels L0 to L4, and the present invention is not limited thereto. Refer to
As indicated in
As indicated in
As indicated in
As indicated in
As indicated in
As indicated in
Refer to
As indicated in
As indicated in
As indicated in
As indicated in
As indicated in
As indicated in
As indicated in
As indicated in
As indicated in
As indicated in
The degree classification model MD1, the multi-class identification model MD2 and the object detection model MD3 perform different types of analysis to the image P0 to obtain several analysis results AR. The neural network model fusion method of the present disclosure fuses the degree classification model MD1, the multi-class identification model MD2 and the object detection model MD3 to effectively increase the accuracy in the identification of lesions. Based on the experimental results, the analysis results AR obtained by the degree classification model MD1 alone show that the image is classified as level L1. On the other hand, when the degree classification model MD1, the multi-class identification model MD2 and the object detection model MD3 are fused together, the fusion result RS1 obtained through the fusion of the above three models shows that the image is classified as level L2 instead of L1. Since level L2 is the correct result, the interpretation accuracy is therefore increased.
Refer to
In step S110, an image P0 is provided, wherein the image P0 is transmitted to the transmission interface 110 through a network 900.
Then, the method proceeds to step S120, whether the size of the image P0 is smaller than a predetermined size is determined by the inference module 121. If the size of the image P0 is smaller than predetermined size, then the method proceeds to step S130.
Then, the method proceeds to step S130, the size of the image P0 is adjusted by the inference module 121 to match the predetermined size.
Then, the method proceeds to step S140, the image P0 is analyzed by the inference module 121 through the degree classification model MD1, the multi-class identification model MD2 and the object detection model MD3. Step S140 includes: analyzing the image P0 by the inference module 121 through the degree classification model MD; analyzing the image P0 by the inference module 121 through the multi-class identification model MD2; and analyzing the image P0 by the inference module 121 through the object detection model MD3.
Then, the method proceeds to step S150, several analysis results AR are obtained by the inference module 121 according to the degree classification model MD1, the multi-class identification model MD2 and the object detection model MD3.
Then, the method proceeds to step S160, the analysis results AR are converted into a number of conversion factors TD by the decision module 122. Refer to Table 4. The analysis results AR obtained by the degree classification model MD1 are presented as a distribution probability matrix MX11. The distribution probability matrix MX11 is element-wise multiplied by a weight matrix (the weight matrix is exemplified by [1, 2, 3, 4, 5]; however, this exemplification is not for limiting the scope of the present disclosure) to obtain a weighted probability matrix MX12 used as the conversion factors TD of the degree classification model MD11. The element-wise multiplication of the distribution probability matrix MX11 and the weight matrix is a product of corresponding elements between the distribution probability matrix MX11 and the weight matrix. The weight matrix can suitably emphasize the importance of each of the levels L0 to L4. Based on experience, the content of the weight matrix can be emphasized according to the application to serve the needs in different scenario.
Refer to Table 5. The analysis results AR obtained by the multi-class identification model MD2 are presented as an individual probability matrix MX21. The individual probability matrix MX21 is converted into a distribution probability matrix MX22 of levels L0 to L4 through a conversion model or a correspondence table. The distribution probability matrix MX22 is used as the conversion factors TD of the multi-class identification model MD2. The distribution probability matrix MX22 of the multi-class identification model MD2 has 5 levels L0 to L4. The distribution probability matrix MX11 of the degree classification model MD1 also has 5 levels L0 to L4.
Refer to Table 6. The analysis results AR obtained by the object detection model MD3 are presented as an individual region description matrix MX31. The individual region description matrix MX31 is converted into an all-object region description matrix MX32 of objects such as microaneurysm (MA), hemorrhage (H), hard exudate (HE), soft exudate (SE) and neovascularization (NEO). In the all-object region description matrix MX32, A represents an object area, and N represents an object quantity.
Then, the method proceeds to step S170, the conversion factors TD is inputted to the fusion model ML to obtain a fusion result RS1 by the decision module 122.
As disclosed above, the analysis results AR respectively obtained by the degree classification model MD1, the multi-class identification model MD2 and the object detection model MD3 can be fused through machine learning to obtain a fusion result RS1. The fusion result RS1 combines the advantages of the neural network models used for different purposes, such that the performance in image recognition can be effectively improved. Particularly when it comes to the interpretation of medical images, the neural network model fusion method of the present disclosure can effectively increase the accuracy in the classification of disease severity and the identification of lesions.
Apart from the embodiments disclosed above, the neural network model fusion method of the present disclosure can also be used in the fusion of different models. Referring to
It will be apparent to those skilled in the art that various modifications and variations can be made to the disclosed embodiments. It is intended that the specification and examples be considered as exemplary only, with a true scope of the disclosure being indicated by the following claims and their equivalents.
| Number | Name | Date | Kind |
|---|---|---|---|
| 7676076 | Spahn | Mar 2010 | B2 |
| 8879813 | Solanki et al. | Nov 2014 | B1 |
| 9445716 | Liu et al. | Sep 2016 | B2 |
| 9524450 | Ravindran et al. | Dec 2016 | B2 |
| 9589374 | Gao et al. | Mar 2017 | B1 |
| 9922272 | Cheng et al. | Mar 2018 | B2 |
| 10019654 | Pisoni | Jul 2018 | B1 |
| 10068171 | Wshah | Sep 2018 | B2 |
| 10198655 | Hotson | Feb 2019 | B2 |
| 10297070 | Zhu | May 2019 | B1 |
| 10657364 | El-Khamy | May 2020 | B2 |
| 20030225719 | Juang | Dec 2003 | A1 |
| 20040061777 | Sadok | Apr 2004 | A1 |
| 20040252878 | Okuda | Dec 2004 | A1 |
| 20060245631 | Levenson | Nov 2006 | A1 |
| 20080292194 | Schmidt et al. | Nov 2008 | A1 |
| 20130156320 | Fredembach | Jun 2013 | A1 |
| 20150010218 | Bayer | Jan 2015 | A1 |
| 20150170002 | Szegedy | Jun 2015 | A1 |
| 20150363643 | Huber | Dec 2015 | A1 |
| 20160259994 | Ravindran et al. | Sep 2016 | A1 |
| 20160292856 | Niemeijer et al. | Oct 2016 | A1 |
| 20170112372 | Chakravorty et al. | Apr 2017 | A1 |
| 20170140253 | Wshah | May 2017 | A1 |
| 20170270653 | Garnavi | Sep 2017 | A1 |
| 20170270674 | Shrivastava | Sep 2017 | A1 |
| 20180012107 | Xu | Jan 2018 | A1 |
| 20180047159 | Schlegl et al. | Feb 2018 | A1 |
| 20180089505 | El-Khamy | Mar 2018 | A1 |
| 20180122068 | Garnavi et al. | May 2018 | A1 |
| 20180130203 | Abedini | May 2018 | A1 |
| 20180165809 | Stanitsas et al. | Jun 2018 | A1 |
| 20180165810 | Jia et al. | Jun 2018 | A1 |
| 20180189613 | Wolf | Jul 2018 | A1 |
| 20180204111 | Zadeh et al. | Jul 2018 | A1 |
| 20180211403 | Hotson | Jul 2018 | A1 |
| 20180214087 | Balaji et al. | Aug 2018 | A1 |
| 20190019318 | Cinnamon | Jan 2019 | A1 |
| 20190122071 | Jin | Apr 2019 | A1 |
| 20190130216 | Tomioka | May 2019 | A1 |
| 20190164290 | Wang | May 2019 | A1 |
| 20190236411 | Zhu | Aug 2019 | A1 |
| 20190392263 | Ferrari | Dec 2019 | A1 |
| 20190392318 | Ghafoor | Dec 2019 | A1 |
| 20200074238 | Umeno | Mar 2020 | A1 |
| 20200074243 | Gu | Mar 2020 | A1 |
| 20200097756 | Hashimoto | Mar 2020 | A1 |
| 20200117954 | Jiang | Apr 2020 | A1 |
| 20200117959 | Sargent | Apr 2020 | A1 |
| 20200125923 | Jones | Apr 2020 | A1 |
| 20200134380 | Xu | Apr 2020 | A1 |
| 20200143160 | Liu | May 2020 | A1 |
| 20200193207 | Kim | Jun 2020 | A1 |
| Number | Date | Country |
|---|---|---|
| 101068498 | Nov 2007 | CN |
| 104392444 | Mar 2015 | CN |
| 106408562 | Feb 2017 | CN |
| 106793992 | May 2017 | CN |
| 107203778 | Sep 2017 | CN |
| 107256550 | Oct 2017 | CN |
| 107292887 | Oct 2017 | CN |
| 107330449 | Nov 2017 | CN |
| 107330876 | Nov 2017 | CN |
| 104603630 | Dec 2017 | CN |
| 107437092 | Dec 2017 | CN |
| 107704886 | Feb 2018 | CN |
| 107967688 | Apr 2018 | CN |
| 108021916 | May 2018 | CN |
| 108198620 | Jun 2018 | CN |
| 107358606 | Jul 2018 | CN |
| 108363704 | Aug 2018 | CN |
| 108363714 | Aug 2018 | CN |
| 2285266 | Apr 2014 | EP |
| 3121789 | Jan 2018 | EP |
| 6177314 | Aug 2017 | JP |
| 514513 | Dec 2002 | TW |
| 201503106 | Jan 2015 | TW |
| M527991 | Sep 2016 | TW |
| WO 0014668 | Mar 2000 | WO |
| WO 2016040317 | Mar 2016 | WO |
| WO 2016132152 | Aug 2016 | WO |
| WO 2016134211 | Aug 2016 | WO |
| WO 2017092615 | Jun 2017 | WO |
| WO 2017147418 | Aug 2017 | WO |
| WO 2017210690 | Dec 2017 | WO |
| WO 2017222849 | Dec 2017 | WO |
| WO 2018022523 | Feb 2018 | WO |
| WO 2018044872 | Mar 2018 | WO |
| WO 2018045269 | Mar 2018 | WO |
| WO 2018093865 | May 2018 | WO |
| WO 2018118373 | Jun 2018 | WO |
| WO 2018120644 | Jul 2018 | WO |
| WO 2018132336 | Jul 2018 | WO |
| Entry |
|---|
| Akilan et al., “A Late Fusion Approach for Harnessing Multi-CNN Model High-level Features”, 2017 IEEE International Conference on Systems, Man, and Cybernetics (SMC), Banff Center, Banff, Canada, Oct. 5-8, 2017, pp. 566-571 (6 pages). |
| Akram et al., “Detection and classification of retinal lesions for grading of diabetic retinopathy”, Computers in Biology and Medicine, vol. 45, 2014, pp. 161-171 (11 pages). |
| American Academy of Opthalmology Retina/Vitreous Panel, Preferred Practice Pattern® Guidelines, “Diabetic Retinopathy”, American Academy of Opthalmology, San Francisco, CA, 2016, 63 pages. |
| Lavinia et al., “Fusion Based Deep CNN for Improved Large-Scale Image Action Recognition”, 2016 IEEE International Symposium on Multimedia, IEEE Computer Society, 2016, pp. 609-614 (6 pages). |
| Number | Date | Country | |
|---|---|---|---|
| 20200193244 A1 | Jun 2020 | US |