Intense interest in applying classifiers (such as convolutional neural networks (CNNs)) in biomedical image analysis is wide spread. For example, CNNs can be used to suggest to an authorized professional whether one or more biomedical images are likely to have one or more given characteristics (which can be represented by one or more of |Y| possible labels) so that the professional can diagnose a medical condition of a patient.
In order for a CNN to perform this function, the CNN needs to be trained using annotated biomedical training images that indicate whether the training images have one or more of the |Y| possible labels. For example, for the CNN to be able to spot a condition in an image, many training images annotated as showing the condition and many training images annotated as not showing the condition can be used to train the CNN. The better trained the CNN is, the less likely the CNN is to misclassify an image.
The success of CNNs for this purpose, however, is impeded by the lack of large annotated datasets in biomedical imaging. Annotating biomedical images is not only tedious and time consuming, but also demanding of costly, specialty-oriented knowledge and skills, which are not easily accessible.
Accordingly, new mechanisms for reducing the burden of annotating biomedical images are desirable.
In accordance with some embodiments, systems, methods, and media, for selecting candidates for annotation for use in training a classifier are provided.
In some embodiments, systems for selecting candidates for labelling and use in training a convolutional neural network (CNN) are provided, the systems comprising: a memory device; and at least one hardware processor configured to: receive a plurality of input candidates, wherein each candidate includes a plurality of identically labelled patches; and for each of the plurality of candidates: determine a plurality of probabilities, each of the plurality of probabilities being a probability that a unique patch of the plurality of identically labelled patches of the candidate corresponds to a label using a pre-trained CNN; identify a subset of candidates of the plurality of input candidates, wherein the subset does not include all of the plurality of candidates, based on the determined probabilities; query an external source to label the subset of candidates to produce labelled candidates; and train the pre-trained CNN using the labelled candidates.
In some embodiments, methods for selecting candidates for labelling and use in training a convolutional neural network (CNN) are provided, the methods comprising: receiving a plurality of input candidates at a hardware processor, wherein each candidate includes a plurality of identically labelled patches; and for each of the plurality of candidates: determining a plurality of probabilities, each of the plurality of probabilities being a probability that a unique patch of the plurality of identically labelled patches of the candidate corresponds to a label using a pre-trained CNN; identifying a subset of candidates of the plurality of input candidates, wherein the subset does not include all of the plurality of candidates, based on the determined probabilities; querying an external source to label the subset of candidates to produce labelled candidates; and training the pre-trained CNN using the labelled candidates.
The research underlying the various embodiments described herein was partly funded by the National Institutes of Health (NIH) under grant R01 HL128785. The applicants thank the NIH for its support of certain aspects of this work.
In accordance with some embodiments, mechanisms, which can include systems, methods, and/or media, for selecting candidates for annotation for use in training a classifier are provided. In some embodiments, the mechanisms can be used in connection with computer aided diagnosis (CAD) in biomedical imaging. More particularly, for example, the mechanisms can be used to select images for annotation, the images can be annotated (in any suitable manner), the annotated images can be used to fine-tune a classifier, and that classifier can be used to perform computer aided diagnosis on biomedical images used, for example, to help an authorized professional (e.g., a medical doctor) to diagnose a medical condition of a patient.
Turning to
As shown, the process receives as an input a set U of n candidates Ci, where i∈[1, n]. The candidates can be received from any suitable source. For example, in some embodiments, a CAD system can include a candidate generator which can produce the set of candidates. Each of the candidates Ci can be labelled with one of more of |Y| possible labels (e.g., informative or non-informative). Some of the candidates can be correctly labelled (true positives) and some of the candidates can be incorrectly labelled (false positives). These candidates can be produced in any suitable manner. For example, in some embodiments, candidate images can be extracted from video.
As also shown in
Data augmentation can be performed in any suitable manner, and any suitable amount of data augmentation can be performed in some embodiments. For example, in some embodiments, an image that is a single frame of a colonoscopy video and that has a size of 712 pixels by 480 pixels can be received and used to form a candidate. The whole image can be labeled as informative or non-informative. The image can then be cropped into 21 patches (e.g., images that are 50 pixels by 50 pixels) from the image by translating the image by ten (or any other suitable numbers, such as twenty) percent of a resized bounding box in vertical and horizontal directions. Each resulting patch can be rotated eight times by mirroring and flipping. All 21 patches can then share the same label and the group of these patches is named as one candidate.
In some embodiments, a factor f (e.g., where factor f∈{1.0, 1.2, 1.5}) can be used to enlarge a patch (e.g., to realize an augmented data set of the original size, 1.2 times larger, and 1.5 times larger) and then crop it back to the original size. For example, if one patch is sized at 10 pixels by 10 pixels, it can be enlarged by a factor f equal to 1.2 to produce a patch of 12 pixels by 12 pixels, and then the patch can be crop to the center 10 pixels by 10 pixels as new a patch after data augmentation.
The manner of performing data augmentation can be based on the application. For example, for colonoscopy frame classification, translation data augmentation can be applied by ten percent of a resized bounding box in vertical and horizontal directions. As another example, for polyp detection, rotation data augmentation can be applied at the center of polyp location. As still another example, for pulmonary embolism detection, scale plus rotation data augmentation can be applied—e.g., by extracting three different physical sizes, e.g., 10 mm, 15 mm, 20 mm wide, by rotating the longitudinal and cross-sectional vessel planes around the vessel axis.
The patches generated from the same candidate can be given the same label(s).
As further shown in
As still further shown in
The outputs of the process shown in
As shown in
Between lines 2 and 10, the process can loop through each candidate Ci in set U, where i∈[1, n].
At line 3, the process can determine the probabilities of each patch xij, where j∈[1, m], in candidate Ci corresponding to the |Y| labels by applying CNN Mt-1 to the patches.
Next, at line 4, the process can determine the average of the probabilities determined at line 3, and, if the average is greater than 0.5, the process can assign to a set Si′ the top α percent of the m patches (i.e., the α percent of the m patches having the highest probabilities) of Ci at line 5. Otherwise, if the average is less than or equal to 0.5, the process can assign to set Si′ the bottom α percent of the m patches (i.e., the α percent of the m patches having the lowest probabilities) of Ci at line 7.
Then, at line 9, the process can calculate a number Ri for candidate Ci using the patches in Si′ using the following equation:
Ri=λ1ei+λ2di
where:
In some embodiments, Ri can additionally or alternatively be calculated based on other characteristics of the patches of a candidate, such as variance, Gaussian distance, standard deviation, and divergence.
After the process has looped through all of the candidates Ci in U at lines 2-10, the process sorts the candidates Ci in U according to the corresponding values Ri at line 11.
Next, the process queries for labels for the top b candidates of Ci in U at line 12 to produce a set of labelled candidates Q. This can be performed in any suitable manner. For example, the top b candidates of Ci in U can be presented to an authorized professional who can manually label each of the top b candidates.
Then, at line 13, the candidates in Q can be added to the set L and removed from the set U, and t can be incremented.
Finally, at line 14, the CNN Mt-1 can be fine-tuned using set L to produce new CNN Mt. The CNN can be fine-tuned in any suitable manner.
Intense interest in applying convolutional neural networks (CNNs) in biomedical image analysis is wide spread, but its success is impeded by the lack of large annotated datasets in biomedical imaging. Annotating biomedical images is not only tedious and time consuming, but also demanding of costly, specialty-oriented knowledge and skills, which are not easily accessible. The process illustrated in
For example, in some embodiments, an AlexNet can be fine-tuned for different example applications using the learning parameters shown in
As mentioned above, in lines 1-15, the process loops until the classification performance of Mt is satisfactory. What is satisfactory can be defined in any suitable manner. For example, the classification performance of Mt can be determined to be satisfactory when newly annotated samples are mostly predicted by current model correctly.
The CAD mechanism described herein can be used for any suitable application. For example, in some embodiments, it can be used for colonoscopy frame classification, polyp detection, pulmonary embolism (PE) detection, and carotid intima-media thickness (CIMT) image classification.
In some embodiments, the mechanisms described herein can be implemented in any suitable platform. For example, in some embodiments, the mechanisms can be implemented in the Caffe framework (described at http://caffe.berkeleyvision.org/, which is hereby incorporated by reference herein in its entirety) based on the pre-trained AlexNet model.
In some embodiments, values of λ1, λ2, and a can be as set forth in the following table for six example configurations:
Turning to
As illustrated in the left column of the table in
In some embodiments, it may be observed that:
In some embodiments, multiple methods may be used to select a particular pattern: for example, entropy, Gaussian distance, and standard deviation would seek Pattern A, while diversity, variance, and divergence look for Pattern C. See
As mentioned above, in some embodiments, the mechanisms described herein can be used for carotid intima-media thickness (CIMT) image classification.
Turning to
The objects can be received from any suitable source. For example, in some embodiments, the objects xij can be generated automatically for each AU Ci through data augmentation.
Data augmentation can be performed in any suitable manner, and any suitable amount of data augmentation can be performed in some embodiments. For example, in some embodiments, an image that is a single frame of a colonoscopy video and that has a size of 712 pixels by 480 pixels can be received and used to form an AU. The whole image can be labeled as informative or non-informative. The image can then be cropped into 21 patches (e.g., images that are 50 pixels by 50 pixels) from the image by translating the image by ten (or any other suitable numbers, such as twenty) percent of a resized bounding box in vertical and horizontal directions. Each resulting patch can be rotated eight times by mirroring and flipping. All 21 patches can then share the same label and the group of these patches is named as one AU.
In some embodiments, a factor f (e.g., where factor f∈{1.0, 1.2, 1.5}) can be used to enlarge a patch (e.g., to realize an augmented data set of the original size, 1.2 times larger, and 1.5 times larger) and then crop it back to the original size. For example, if one patch is sized at 10 pixels by 10 pixels, it can be enlarged by a factor f equal to 1.2 to produce a patch of 12 pixels by 12 pixels, and then the patch can be crop to the center 10 pixels by 10 pixels as new a patch after data augmentation.
The manner of performing data augmentation can be based on the application. For example, for colonoscopy frame classification, translation data augmentation can be applied by ten percent of a resized bounding box in vertical and horizontal directions. As another example, for polyp detection, rotation data augmentation can be applied at the center of polyp location. As still another example, for pulmonary embolism detection, scale plus rotation data augmentation can be applied—e.g., by extracting three different physical sizes, e.g., 10 mm, 15 mm, 20 mm wide, by rotating the longitudinal and cross-sectional vessel planes around the vessel axis.
Each of the AUs Ci can be labelled with one of more of |Y| possible labels. Some of the AUs can be correctly labelled (true positives) and some of the AUs can be incorrectly labelled (false positives). These AUs can be produced in any suitable manner. For example, in some embodiments, AUs can be images extracted from video.
The patches generated from the same AU can be given the same label(s).
As further shown in
As still further shown in
Like the process of
As shown in the figure, in lines 1-10, the process loops until the classification performance of Mt is satisfactory. What is satisfactory can be defined in any suitable manner. For example, the classification performance of Mt can be determined to be satisfactory when newly annotated samples are mostly predicted by current model correctly.
Between lines 2 and 5, the process can loop through each AU Ci in set U, where i∈[1, n].
At line 3, the process can determine the probabilities of each object xij, where j∈[1, m], in AU Ci corresponding to the |Y| labels by applying CNN Mt-1 to the objects.
Next, at line 4, the process can determine the entropy of the AU using the following formula:
After the process has looped through all of the AUs Ci in U at lines 2-5, the process sorts the AUs Ci in U according to the corresponding values entropy at line 6 and stores the sorted AUs in U′.
Next, the process queries for labels for the top b AUs of Ci in U′ at line 7 to produce a set of labelled candidates Q. This can be performed in any suitable manner. For example, the AUs in U′ can be annotated as described below in connections with
Then, at line 8, the candidates in Q can be added to the set L and removed from the set U, and t can be incremented.
Finally, at line 9, the CNN Mt-1 can be fine-tuned using set L to produce new CNN Mt. The CNN can be fine-tuned in any suitable manner, for example as described above in connection with
As mentioned above, in lines 1-10, the process loops until the classification performance of Mt is satisfactory. What is satisfactory can be defined in any suitable manner. For example, the classification performance of Mt can be determined to be satisfactory when newly annotated samples are mostly predicted by current model correctly.
The process of
Annotating CIMT is not only tedious, laborious, and time consuming, but also demanding of costly, specialty-oriented knowledge and skills, which are not easily accessible. As described herein, some embodiments dramatically reduce the cost of expert annotation in CIMT by providing: (1) a new mechanism which simplifies the entire CIMT annotation process to six simple mouse clicks; and (2) a new process (described in connection with
Turning to
Turning to
As discussed below in connection with
where M(p) denotes the confidence map prediction of pixel p being in the ROI, C* is the largest connected component in M that is nearest to the carotid bulb, and I(p) is an indicator function for pixel p=[px, py] that is defined as:
An example of a determined ROI is shown as the rectangle in
Basically, the indicator function excludes the pixels located farther than 1 cm from the carotid bulb location. This choice of the distance threshold is motivated by the fact that the ROI is located within 1 cm to the right of the carotid bulb. Any other distance can be used in some embodiments.
To automatically measure intima-media thickness, the lumen-intima and media-adventitia interfaces of the carotid artery must be detected within the ROI. This interface segmentation problem can be treated as a three-class classification task with the goal to classify each pixel within the ROI into one of three categories: 1) a pixel on the lumen-intima interface, 2) a pixel on the media-adventitia interface, and 3) a background pixel.
During operation, the trained CNN can be applied to a given suspected ROI in a convolutional manner, generating two confidence maps with the same size as the ROI. The first confidence map shows the probability of a pixel being on the lumen-intima interface; the second confidence map shows the probability of a pixel being on the media-adventitia interface. A relatively thick high-probability band is apparent along each interface, which hinders the accurate measurement of intima-media thickness. To thin the detected interfaces, the confidence map can be scanned column by column, searching for the rows with the maximum response for each of the two interfaces. By doing so, a one-pixel-thick boundary with a step-like shape around each interface can be obtained. To further refine the boundaries, two active contour models (a.k.a., snakes), one for the lumen-intima interface and one for the media-adventitia interface, can be used. The open snakes can be initialized with the current step-like boundaries and then deformed solely based on the probability maps generated by the CNN rather than the original image content.
Turning to
As shown, the process receives as an input a set D of n annotation units (AUs) Ui, where i∈[1, n]. The AUs can be received from any suitable source. For example, in some embodiments, a CAD system can include an AU generator which can produce the set of AUs. Each of the AUs Ui can be labelled with one of more of |Y| possible labels. Some of the AUs can be correctly labelled (true positives) and some of the AUs can be incorrectly labelled (false positives). These AUs can be produced in any suitable manner. For example, in some embodiments, AU images can extracted from video.
As also shown in
Data augmentation can be performed in any suitable manner, and any suitable amount of data augmentation can be performed in some embodiments. For example, in some embodiments, an image that is a single frame of a colonoscopy video and that has a size of 712 pixels by 480 pixels can be received and used to form an AU. The whole image can be labeled as informative or non-informative. The image can then be cropped into 21 patches (e.g., images that are 50 pixels by 50 pixels) from the image by translating the image by ten (or any other suitable numbers, such as twenty) percent of a resized bounding box in vertical and horizontal directions. Each resulting patch can be rotated eight times by mirroring and flipping. All 21 patches can then share the same label and the group of these patches is named as one AU.
In some embodiments, a factor f (e.g., where factor f∈{1.0, 1.2, 1.5}) can be used to enlarge a patch (e.g., to realize an augmented data set of the original size, 1.2 times larger, and 1.5 times larger) and then crop it back to the original size. For example, if one patch is sized at 10 pixels by 10 pixels, it can be enlarged by a factor f equal to 1.2 to produce a patch of 12 pixels by 12 pixels, and then the patch can be cropped to the center 10 pixels by 10 pixels as new a patch after data augmentation.
The manner of performing data augmentation can be based on the application. For example, for colonoscopy frame classification, translation data augmentation can be applied by ten percent of a resized bounding box in vertical and horizontal directions. As another example, for polyp detection, rotation data augmentation can be applied at the center of polyp location. As still another example, for pulmonary embolism detection, scale plus rotation data augmentation can be applied—e.g., by extracting three different physical sizes, e.g., 10 mm, 15 mm, 20 mm wide, by rotating the longitudinal and cross-sectional vessel planes around the vessel axis.
The patches generated from the same AU can be given the same label(s).
As further shown in
As still further shown in
The outputs of the process shown in
As shown in the figure, in lines 2-19, the process loops until the classification performance of Mτ is satisfactory. What is satisfactory can be defined in any suitable manner. For example, the classification performance of Mτ can be determined to be satisfactory when newly annotated samples are mostly predicted by current model correctly.
Between lines 3 and 11, the process can loop through each AU Ui in set D, where i∈[1, n].
A line 4, the process can determine the probabilities of each patch xij, where j∈[1, m], in AU Ui corresponding to the |Y| labels by applying CNN Mτ-1 to the patches.
Next, at line 5, the process can determine the mean of the probabilities determined at line 4, and, if the mean is greater than 0.5, the process can assign to a set Ui∝ the top α percent of the m patches (i.e., the α percent of the m patches having the highest probabilities) of Ui at line 6. Otherwise, if the mean is less than or equal to 0.5, the process can assign to set Uiα′ the bottom α percent of the m patches (i.e., the α percent of the m patches having the lowest probabilities) of Ui at line 8.
Then, at line 5, the process can build a quota Ri for AU Ui for Uiα using the following equation:
Ri=λ1ei+λ2di
where:
In some embodiments, Ri can additionally or alternatively be calculated based on other characteristics of the patches of an AU, such as variance, Gaussian distance, standard deviation, and divergence.
After the process has looped through all of the AUs Ui in D at lines 3-11, the process sorts the AUs Ui in D according to the corresponding values Ri at line 12 to produce D′, a sorted list of Ui, and R′, a sorted list of Ri.
At line 13, the process first normalizes the top w*b entries in R′ so that they have values between zero and one. Then, at line 13, the process converts the normalized values to sampling probabilities having values between zero and one and adding up to one.
Next, the process queries for labels for the top b candidates of Ui in D′ based on the corresponding values in Rsel at line 14 to produce a set of labelled candidates Q. This can be performed in any suitable manner. For example, the top b candidates of Ui in D′ can be presented to an authorized professional who can manually label each of the top b candidates.
Then, at line 15, the current model Mτ-1 can be used to test labeled data L and get predictions p.
At line 16, the predictions p can next be compared with the labels assigned at line 14, the misclassified AUs can be assigned to set H.
Then, at line 17, the CNN Mτ-1 can be fine-tuned using the union of sets H and Q to produce new CNN Mτ. The CNN can be fine-tuned in any suitable manner. For example, in some embodiments, an AlexNet CNN can be fine-tuned for different example applications using the learning parameters shown in
Then, at line 18, the AUs in Q can be added to the set L and removed from the set D, and τ can be incremented.
As mentioned above, in lines 2-19, the process loops until the classification performance of Mτ is satisfactory. What is satisfactory can be defined in any suitable manner. For example, the classification performance of Mτ can be determined to be satisfactory when newly annotated samples are mostly predicted by current model correctly.
The splendid success of convolutional neural networks (CNNs) in computer vision is largely attributed to the availability of large annotated datasets, but in biomedical imaging it is very challenging to create such large datasets, as annotating biomedical images is not only tedious, laborious, and time consuming, but also demanding of costly, specialty-oriented skills, which are not easily accessible. As described herein, various embodiments dramatically reduce annotation cost by integrating active learning and transfer learning into a single framework by providing: (1) annotation units (AUs) which strike a balance between annotation efficiency and label integrity; (2) a comprehensive analysis of the CNN prediction patterns associated with AUs; (3) four active selection strategies to identify the AUs most effective in boosting the CNN performance; (4) a process (as described in connection with
In some embodiment, alternatively to performing the process in
Turning to
The medical imaging device and/or the computer can each include any suitable components. For example, in some embodiments, each can include one or more hardware processors (e.g., microprocessor(s), microcontroller(s), digital signal processor(s), etc.), one or more memory devices (e.g., RAM, ROM, EPROM, FLASH, static memory, dynamic memory, solid state drive, hard disk, etc.), one or more computer network interfaces (e.g., NIC card), one or more input devices (e.g., mouse, keyboard, light-pen, touch screen, wand sensor, etc.), one or more output devices (e.g., display, speaker, printer, etc.), and/or any other suitable computer device.
Any of the processes described herein can be programmed into any suitable memory devices in the medical imaging device and/or computer and be executed by a hardware processor in the medical imaging device and/or computer.
It should be understood that at least some of the above described steps of the process of
The processes disclosed herein may be implemented by a computer program product. The computer program product may include computer code arranged to instruct a computer to perform the functions of one or more of the various processes described above. The computer program product and/or the code for performing such methods may be provided to an apparatus, such as a computer, on computer-readable media. In some implementations, any suitable computer-readable media can be used for storing instructions for performing the functions and/or processes described herein. For example, in some implementations, computer-readable media can be transitory or non-transitory. For example, non-transitory computer-readable media can include media such as non-transitory forms of magnetic media (such as hard disks, floppy disks, etc.), non-transitory forms of optical media (such as compact discs, digital video discs, Blu-ray discs, etc.), non-transitory forms of semiconductor media (such as flash memory, electrically programmable read only memory (EPROM), electrically erasable programmable read only memory (EEPROM), etc.), any suitable media that is not fleeting or devoid of any semblance of permanence during transmission, and/or any suitable tangible media. As another example, transitory computer-readable media can include signals on networks, in wires, conductors, optical fibers, circuits, any suitable media that is fleeting and devoid of any semblance of permanence during transmission, and/or any suitable intangible media.
An apparatus, such as a computer, may be configured in accordance with such code to perform one or more processes in accordance with the various methods discussed herein.
Such an apparatus may take the form of a data processing system. Such a data processing system may be a distributed system. For example, such a data processing system may be distributed across a network.
In another arrangement, a computer-readable medium comprising instructions executable by a processor to perform any one of the various methods disclosed herein is provided.
Although the invention has been described and illustrated in the foregoing illustrative embodiments, it is understood that the present disclosure has been made only by way of example, and that numerous changes in the details of implementation of the invention can be made without departing from the spirit and scope of the invention, which is limited only by the claims that follow. Features of the disclosed embodiments can be combined and rearranged in various ways.
This application claims the benefit of U.S. Provisional Patent Application No. 62/491,069, filed Apr. 27, 2017, and U.S. Provisional Patent Application No. 62/663,931, filed Apr. 27, 2018, each of which is hereby incorporated by reference herein its entirety.
This invention was made with government support under R01 HL128785 awarded by the National Institutes of Health. The government has certain rights in the invention.
Number | Name | Date | Kind |
---|---|---|---|
9152926 | Liang et al. | Oct 2015 | B2 |
9330336 | Tajbakhsh et al. | May 2016 | B2 |
9449381 | Liang | Sep 2016 | B2 |
9603554 | Liang et al. | Mar 2017 | B2 |
9684957 | Wu et al. | Jun 2017 | B2 |
9700213 | Tajbakhsh et al. | Jul 2017 | B2 |
9741116 | Liang et al. | Aug 2017 | B2 |
9747687 | Tajbakhsh et al. | Aug 2017 | B2 |
9924927 | Shin et al. | Mar 2018 | B2 |
9959615 | Liang et al. | May 2018 | B2 |
9978142 | Chi et al. | May 2018 | B2 |
10052027 | Tajbakhsh et al. | Aug 2018 | B2 |
10055843 | Tajbakhsh et al. | Aug 2018 | B2 |
10120980 | Liang | Nov 2018 | B2 |
10157045 | Venkataramani et al. | Dec 2018 | B2 |
10157467 | Dincer et al. | Dec 2018 | B2 |
10328282 | An et al. | Jun 2019 | B2 |
10610203 | Liang et al. | Apr 2020 | B2 |
10861151 | Liang et al. | Dec 2020 | B2 |
10956785 | Liang et al. | Mar 2021 | B2 |
11436725 | Taher et al. | Sep 2022 | B2 |
20080082468 | Long et al. | Apr 2008 | A1 |
20140135627 | Liang et al. | May 2014 | A1 |
20170083792 | Rodriguez-Serrano et al. | Mar 2017 | A1 |
20170099200 | Ellenbogen et al. | Apr 2017 | A1 |
20170124701 | Liang et al. | May 2017 | A1 |
20170337682 | Liao et al. | Nov 2017 | A1 |
20180144241 | Liu et al. | May 2018 | A1 |
20180144243 | Hsieh et al. | May 2018 | A1 |
20180189229 | Desoli et al. | Jul 2018 | A1 |
20180225820 | Liang et al. | Aug 2018 | A1 |
20180225823 | Zhou et al. | Aug 2018 | A1 |
20180253866 | Jain et al. | Sep 2018 | A1 |
20180260957 | Yang et al. | Sep 2018 | A1 |
20180293496 | Vogels et al. | Oct 2018 | A1 |
20180293734 | Lim et al. | Oct 2018 | A1 |
20180307947 | Choi et al. | Oct 2018 | A1 |
20180314943 | Liang et al. | Nov 2018 | A1 |
20190021677 | Grbic et al. | Jan 2019 | A1 |
20190065908 | Lee et al. | Feb 2019 | A1 |
20190130110 | Lee et al. | May 2019 | A1 |
20190223738 | Böscke et al. | Jul 2019 | A1 |
20190244348 | Buckler et al. | Aug 2019 | A1 |
20190325621 | Wang et al. | Oct 2019 | A1 |
20190332896 | Liang et al. | Oct 2019 | A1 |
20200074271 | Liang et al. | Mar 2020 | A1 |
20200074701 | Liang et al. | Mar 2020 | A1 |
20200134461 | Chai et al. | Apr 2020 | A1 |
20200175352 | Cha et al. | Jun 2020 | A1 |
20200272864 | Faust et al. | Aug 2020 | A1 |
20200327162 | Pevtsov et al. | Oct 2020 | A1 |
20200364477 | Siddiquee et al. | Nov 2020 | A1 |
20200380695 | Zhou et al. | Dec 2020 | A1 |
20210150710 | Taher et al. | May 2021 | A1 |
Number | Date | Country |
---|---|---|
2012109670 | Aug 2012 | WO |
2012109676 | Aug 2012 | WO |
2013116865 | Aug 2013 | WO |
2013116867 | Aug 2013 | WO |
2015031641 | Mar 2015 | WO |
2015113014 | Jul 2015 | WO |
2015142808 | Sep 2015 | WO |
2015164724 | Oct 2015 | WO |
2015164768 | Oct 2015 | WO |
2016161115 | Oct 2016 | WO |
2017027475 | Feb 2017 | WO |
2018232388 | Dec 2018 | WO |
2019019199 | Jan 2019 | WO |
Entry |
---|
Yan (“Multi-Instance Deep Learning: Discover Discriminative Local Anatomies for Bodypart Recognition”) IEEE Transactions on Medical Imaging, vol. 35, No. 5, May 2016 (Year: 2016). |
Maggiori (“Convolutional Neural Networks for Large-Scale Remote-Sensing Image Classification”) IEEE Transactions on Geoscience and Remote Sensing, vol. 55, No. 2, Feb. 2017 (Year: 2017). |
Chakraborty (“Active Batch Selection via Convex Relaxations with Guaranteed Solution Bounds”) IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 37, No. 10, Oct. 2015 (Year: 2015). |
Montenegro (“A framework for interactive training of automatic image analysis models based on learned image representations, active learning and visualization techniques”). Universidad Nacional de Colombia Engineering School, Systems and Industrial Engineering Department Bogotá D.C., Colombia (Year: 2016). |
Grinsven (“Fast Convolutional Neural Network Training Using Selective Data Sampling: Application to Hemorrhage Detection in Color Fundus Images”) IEEE Transactions on Medical Imaging, vol. 35, No. 5, May 2016 (Year: 2016). |
Oliveira (“A Data Augmentation Methodology to Improve Age Estimation Using Convolutional Neural Networks”) 2016 29th SIBGRAPI Conference on Graphics, Patterns and Images (Year: 2016). |
Liu (“Fully Convolutional Attention Networks for Fine-Grained Recognition”) arXiv:1603.06765v4 [cs.CV] Mar. 21, 2017 (Year: 2017). |
Tajbakhsh (“Convolutional Neural Networks for Medical Image Analysis: Full Training or Fine Tuning?”) IEEE Transactions on Medical Imaging, vol. 35, No. 5, May 2016 (Year: 2016). |
Li (“A high accuracy automatic timing method for photo finish systems”) 2014 IEEE International Conference on Progress in Informatics and Computing (Year: 2014). |
Shin (“Automating Carotid Intima-Media Thickness Video Interpretation with Convolutional Neural Networks”) 2016 IEEE Conference on Computer Vision and Pattern Recognition (Year: 2016). |
Settles, B., “Active Learning Literature Survey”, Computer Sciences Technical Report: University of Wisconsin-Madison, 2010, vol. 52, No. 55-56, 67 pages. |
Shannon, C., “A Mathematical Theory of Communication”, The Bell System Technical Journal, Oct. 1948, vol. 27, No. 3, pp. 379-423. |
Shelhamer, E. et al., “Fully Convolutional Networks for Semantic Segmentation”, IEEE Transactions on Pattern Analysis and Machine Intelligence, Apr. 2017 (Date of Publication: May 24, 2016), vol. 39, No. 4, pp. 640-651 <DOI:10.1109/TPAMI.2016.2572683>. |
Shin, H-C. et al., “Deep Convolutional Neural Networks for Computer-Aided Detection: CNN Architectures, Dataset Characteristics and Transfer Learning”, IEEE Transactions on Medical Imaging, May 2016 (date of publication Feb. 2016), vol. 35, No. 5, pp. 1285-1298 <DOI:10.1109/TMI.2016.2528162>. |
Shin, H-C. et al., “Interleaved text/image Deep Mining on a large-scale radiology database”, IEEE Conference on Computer Vision and Pattern Recognition (Boston, MA, Jun. 7-12, 2015), 2015, pp. 1090-1099 <DOI:10.1109/CVPR.2015.7298712>. |
Shin, J. et al., “Automating Carotid Intima-Media Thickness Video Interpretation with Convolutional Neural Networks”, IEEE Conference on Computer Vision and Pattern Recognition (Las Vegas, NV, Jun. 27-30, 2016), 2016 (date added to IEEE Xplore Dec. 2016), pp. 2526-2535 <DOI:10.1109/CVPR.2016.277>. |
Simonyan, K. et al., “Very deep convolutional networks for large-scale image recognition”, arXiv, Published as a conference paper at ICLR 2015, 2014, 14 pages. |
Stark, F. et al., “CAPTCHA recognition with active deep learning”, Citeseer, Sep. 2015, 8 pages. |
Stein, J. et al., “Use of Carotid Ultrasound to Identify Subclinical Vascular Disease and Evaluate Cardiovascular Disease Risk: A Consensus Statement from the American Society of Echocardiography Carotid Intima-Media Thickness Task Force Endorsed by the Society for Vascular Medicine”, Journal of the American Society of Echocardiography, 2008, vol. 21, No. 2, pp. 93-111 <DOI:10.1016/j.echo.2007.11.011>. |
Szegedy, C. et al., “Going deeper with convolutions”, IEEE Conference on Computer Vision and Pattern Recognition (Boston, MA, Jun. 7-12, 2015), Date Added to IEEE Xplore: Oct. 2015 <DOI:10.1109/CVPR.2015.7298594>. |
Tajbakhsh, N. et al., “Chapter 5—Automatic Interpretation of Carotid Intima-Media Thickness Videos Using Convolutional Neural Networks”, Deep Learning for Medical Image Analysis, Jan. 2017, pp. 105-131 <DOI:10.1016/B978-0-12-810408-8.00007-9>. |
Tajbakhsh, N. et al., “Computer-aided pulmonary embolism detection using a novel vessel-aligned multi-planar image representation and convolutional neural networks”, International Conference on Medical Image Computing and Computer-Assisted Intervention, Springer, 2015, pp. 62-69 <DOI:10.1007/978-3-319-24571-3_8>. |
Tajbakhsh, N. et al., “Convolutional neural networks for medical image analysis: Full training or fine tuning?”, May 2016 (Date of Publication: Mar. 7, 2016), vol. 35, No. 5, pp. 1299-1312 <DOI:10.1109/TMI.2016.2535302>. |
Van Rijn, J. et al., “Polyp miss rate determined by tandem colonoscopy: a systematic review”, American Journal of Gastroenterology, 2006, vol. 101, No. 2, pp. 343-350 <DOI:10.1111/j.1572-0241.2006.00390.x>. |
Wang, D. et al., “A new active labeling method for deep learning”, International Joint Conference on Neural Networks (Beijing, China, Jul. 6-11, 2014), 2014, pp. 112-119 <DOI:10.1109/IJCNN.2014.6889457>. |
Wang, H. et al., “Comparison of machine learning methods for classifying mediastinal lymph node metastasis of non-small cell lung cancer from 18F-FDG PET/CT images”, EJNMMI Research, Jan. 2017, vol. 7, No. 11, 11 pages <DOI:10.1186/s13550-017-0260-9>. |
Weiss, K. et al., “A survey of transfer learning”, Journal of Big Data, May 2016, vol. 3, No. 9, 40 pages <DOI:10.1186/s40537-016-0043-6>. |
World Health Organization (WHO)., “Global Atlas on Cardiovascular Disease Prevention and Control”, Sep. 2011, 166 pages, available online <URL:https://www.who.int/cardiovascular_diseases/publications/atlas_cvd/en/>. |
Yang, L. et al., “Suggestive annotation: A deep active learning framework for biomedical image segmentation”, arXiv, Jun. 2017, pp. 399-407. |
Yuan, X-T. et al., “Truncated power method for sparse eigenvalue problems”, The Journal of Machine Learning Research, Jan. 2013, vol. 14, No. 1, pp. 899-925. |
Zhang, J. et al., “Transfer Learning for Cross-Dataset Recognition: A Survey”, 2017, accessed from: https://arxiv.org/abs/1705.04396. |
Zhou, B. et al., “Places: A 10 million image database for scene recognition”, IEEE Transactions on Pattern Analysis and Machine Intelligence, Jul. 2017, vol. 40, No. 6, pp. 1452-1464 <DOI:10.1109/TPAMI.2017.2723009>. |
Zhou, B. et al., “Places: An Image Database for Deep Scene Understanding”, arXiv preprint, Oct. 2016, 12 pages, accessed from: arXiv:1610.02055. |
Zhou, Z. et al., “AFT*: Integrating Active Learning and Transfer Learning to Reduce Annotation Efforts”, arXiv:1802.00912, 2018, 14 pages. |
Zhou, Z. et al., “Fine-Tuning Convolutional Neural Networks for Biomedical Image Analysis: Actively and Incrementally”, IEEE Conference on Computer Vision and Pattern Recognition (Honolulu, HI, Jul. 21-26, 2017), 2017 (Date Added to IEEE Xplore: Nov. 9, 2017), pp. 4761-4772 <DOI:10.1109/CVPR.2017.506>. |
Zhou, Z. et al., “Integrating Active Learning and Transfer Learning for Carotid Intima-Media Thickness Video Interpretation”, Journal of Digital Imaging, 2019 (available online Nov. 2018), vol. 32, pp. 290-299 <DOI:10.1007/s10278-018-0143-2>. |
Huang, S., “Cost-Effective Training of Deep CNNs with Active Model Adaptation”, in: Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, ACM. Aug. 10-23, 2018pp. 1580-1588. |
Lin, L., “Active Self-Paced Learning for Cost-Effective and Progressive Face Identification”, IEEE Transactions on Pattern Analysis and Machine Intelligence 40, Jan. 7-19, 2018, pp. 1-13. |
Huang, S-J. et al., “Active learning by querying informative and representative examples”, International Conference on Neural Information Processing (Vancouver, Canada, Dec. 6-9, 2010), 2010, vol. 1, pp. 9 pages. |
Liu, B., “Active learning for human pose estimation”, in: Proceedings of the IEEE International Conference on Computer Vision, Oct. 22-29, 2017, pp. 4363-4372. |
Xie, S., “Aggregated residual transformations for deep neural networks”, in: Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 5987-5995. |
Al Rahhal, M. et al., “Deep learning approach for active classification of electrocardiogram signals”, Information Sciences, Jun. 2016, vol. 345, pp. 340-354 <DOI:10.1016/j.ins.2016.01.082>. |
Borisov, A. et al., “Active Batch Learning with Stochastic Query-by-Forest (SQBF)”, JMLR: Workshop and Conference Proceedings, 2010 (published 2011), vol. 16, pp. 59-69. |
Calder, K. et al., “The mortality of untreated pulmonary embolism in emergency department patients”, Annals of Emergency Medicine, Mar. 2005, vol. 45, No. 3, pp. 302-310 <DOI:10.1016/j.annemergmed.2004.10.001>. |
Carneiro, G. et al., “Unregistered multiview mammogram analysis with pre-trained deep learning models”, Medical Image Computing and Computer-Assisted Intervention—MICCAI (Munich, Germany, Oct. 5-9, 2015), 2015, vol. 9351, pp. 652-660 <DOI:10.1007/978-3-319-24574-4 78>. |
Chakraborty, S. et al., “Active Batch Selection via Convex Relaxations with Guaranteed Solution Bounds”, IEEE Transactions on Pattern Analysis and Machine Intelligence, Oct. 2015 (Date of Publication: Jan. 12, 2015), vol. 37, No. 10, pp. 1945-1958 <DOI:10.1109/TPAMI.2015.2389848>. |
Chattopadhyay, R. et al., “Joint transfer and batch-mode active learning”, International Conference on Machine Learning (Atlanta, GA, Jun. 16-21, 2013), 2013, pp. 253-261. |
Chen, H. et al., “Automatic fetal ultrasound standard plane detection using knowledge transferred recurrent neural hetworks”, Medical Image Computing and Computer-Assisted Intervention—MICCAI (Munich, Germany, Oct. 5-9, 2015), 2015, pp. 507-514. |
Chen, H. et al., “Standard plane localization in fetal ultrasound via domain transferred deep neural networks”, IEEE Journal of Biomedical and Health Informatics, Sep. 2015 (Date of Publication: Apr. 21, 2015), vol. 19, No. 5, pp. 1627-1636 <DOI:10.1109/JBHI.2015.2425041>. |
D'Agostino, R. et al., “General Cardiovascular Risk Profile for use in Primary Care”, Circulation, Feb. 2008 (available online Jan. 2008), vol. 117, No. 6, pp. 743-753 <DOI:10.1161/CIRCULATIONAHA.107.699579>. |
Delsanto, S. et al., “Characterization of a Completely User-Independent Algorithm for Carotid Artery Segmentation in 2-D Ultrasound Images”, IEEE Transactions on Instrumentation and Measurement, Aug. 2007 (date of publication Jul. 2007), vol. 56, No. 4, pp. 1265-1274 <DOI:10.1109/TIM.2007.900433>. |
Deng, J. et al., “Imagenet: A large-scale hierarchical image database”, IEEE Conference on Computer Vision and Pattern Recognition (Miami, FL, Jun. 20-25, 2009), 2009, pp. 248-255 <DOI:10.1109/CVPR.2009.5206848>. |
Feige, U., “A Threshold of In n for Approximating Set Cover”, Journal of the ACM, Jul. 1998, vol. 45, No. 4, pp. 634-652 <DOI:10.1145/285055.285059>. |
Gao, M. et al., “Holistic classification of ct attenuation patterns for interstitial lung diseases via deep convolutional neural networks”, Computer Methods in Biomechamics and Biomedical Engineering. Imaging & Visualization, 2015 (available online Jun. 2016), vol. 6, No. 1, pp. 1-6 <DOI:10.1080/21681163.2015.1124249>. |
Gepner, A. et al., “Comparison of Coronary Artery Calcium Presence, Carotid Plaque Presence, and Carotid Intima-Media Thickness for Cardiovascular Disease Prediction in the Multi-Ethnic Study of Atherosclerosis”, Circulation, Jan. 2015, vol. 8, No. 1, pp. 1-8 <DOI:10.1161/CIRCIMAGING.114.002262>. |
Greenspan, H. et al., “Guest Editorial Deep Learning in Medical Imaging: Overview and Future Promise of an Exciting New Technique”, May 2016 (Date of Publication: Apr. 29, 2016), vol. 35, No. 5, pp. 1153-1159 <DOI:10.1109/TMI.2016.2553401>. |
Gunn, S. et al., “Support Vector Machines for Classification and Regression”, ISIS Technical Report, May 1998, vol. 14, pp. 85-86. |
Guyon, I. et al., “Results of the Active Learning Challenge”, JMLR: Workshoup and Conference Proceedings, 2011, vol. 16, pp. 19-45. |
He, K. et al., “Deep Residual Learning for Image Recognition”, IEEE Conference on Computer Vision and Pattern Recognition, Jun. 2016, pp. 770-778 <DOI:10.1109/CVPR.2016.90>. |
Heresbach, D. et al., “Miss rate for colorectal neoplastic polyps: a prospective multicenter study of back-to-back video colonoscopies”, Endoscopy, Apr. 2008, vol. 40, No. 4, pp. 284-290 <DOI:10.1055/s-2007-995618>. |
Juang, G. et al., “Densely connected convolutional networks”, arXiv preprint, 2016, 9 pages, <accessed from: arXiv:1608.06993>. |
Juang, S-J. et al., “Active learning by querying informative and representative examples”, International Conference on Neural Information Processing (Vancouver, Canada, Dec. 6-9, 2010), 2010, vol. 1, pp. 9 pages. |
Hurst, T. et al., “Incidence of Subclinical Atherosclerosis as a Marker of Cardiovascular Risk in Retired Professional Football Players”, The American Journal of Cardiology, Apr. 2010 (available online Feb. 2010), vol. 105, No. 8, pp. 1107-1111 <DOI:10.1016/j.amjcard.2009.12.012>. |
Jia, Y. et al., “Caffe: Convolutional Architecture for Fast Feature Embedding”, arXiv preprint, 2014, 4 pages, accessed from: arXiv:1408.5093. |
Kass, M. et al., “Snakes: Active Contour Models”, International Journal of Computer Vision, 1988, vol. 1, No. 4, pp. 321-331. |
Kim, D. et al., “CT Colonography versus Colonoscopy for the Detection of Advanced Neoplasia”, The New England Journal of Medicine, Oct. 2007, vol. 357, No. 14, pp. 1403-1412 <DOI:10.1056/NEJMoa070543>. |
Konyushkova, K. et al., “Learning Active Learning from Data”, Advances in Neural Information Processing Systems, 2017, pp. 4226-4236. |
Krizhevsky, A. et al., “ImageNet Classification with Deep Convolutional Neural Networks”, Advances in Neural Information Processing Systems, 2012, pp. 1097-1105. |
Kukar, M. et al., “Transductive reliability estimation for medical diagnosis”, Artificial Intelligence in Medicine, Sep.-Oct. 2003, vol. 29, No. 1-2, pp. 81-106 <DOI:10.1016/S0933-3657(03)00043-5>. |
Lecun, Y. et al., “Deep learning”, Nature, May 2015, vol. 521, No. 7553, pp. 436-444 <DOI:10.1038/nature14539>. |
Lecun, Y. et al., “Gradient-based learning applied to document recognition”, Proceedings of the IEEE, Nov. 1998, vol. 86, No. 11, pp. 2278-2324 <DOI:10.1109/5.726791>. |
Leufkens, A. et al., “Factors influencing the miss rate of polyps in a back-to-back colonoscopy study”, Endoscopy, May 2012, vol. 44, No. 5, pp. 470-475 <DOI:10.1055/s-0031-1291666>. |
Li, J., “Active learning for hyperspectral image classification with a stacked autoencoders based neural network”, IEEE International Conference on Image Processing (Phoenix, AZ, Sep. 25-28, 2016), 2016, pp. 1062-1065 <DOI:10.1109/ICIP.2016.7532520>. |
Li, M. et al., “Confidence-based active learning”, IEEE Transactions on Pattern Analysis and Machine Intelligence, Aug. 2006 (Date of Publication: Jun. 19, 2006), vol. 28, No. 8, pp. 1251-1261 <DOI:10.1109/TPAMI.2006.156>. |
Liang, J. et al., “Computer aided detection of pulmonary embolism with tobogganing and mutiple instance classification in CT pulmonary angiography”, International Conference of Information Processing in Medical Imaging (Kerkrade, Netherlands, Jul. 2007), 2007, pp. 630-641. |
Liang, J. et al., “United Snakes”, Medical Image Analysis, Apr. 2006 (available online Nov. 2005), vol. 10, No. 2, pp. 215-233 <DOI:10.1016/j.media.2005.09.002>. |
Liang, Q. et al., “A multiscale dynamic programming procedure for boundary detection in ultrasonic artery images”, IEEE Transactions on Medical Imaging, Feb. 2000, vol. 19, No. 2, pp. 127-142 <DOI:10.1109/42.836372>. |
Litjens, G. et al., “A survey on deep learning in medical image analysis”, Medical Image Analysis, Dec. 2017, vol. 42, pp. 60-88 <DOI:10.1016/j.media.2017.07.005>. |
Loizou, C. et al., “An integrated system for the segmentation of atherosclerotic carotid plaque”, IEEE Transactions on Information Technology in Biomedicine, Nov. 2007, vol. 11, No. 6, pp. 661-667 <DOI:10.1109/TITB.2006.890019>. |
Long, J. et al., “Fully Convolutional Networks for Semantic Segmentation”, IEEE Conference on Computer Vision and Pattern Recognition, 2015, pp. 3431-3440. |
Lorenz, M. et al., “Prediction of Clinical Cardiovascular Events With Carotid Intima-Media Thickness”, Circulation, Jan. 2007, vol. 115, No. 4, pp. 459-467 <DOI:10.1161/CIRCULATIONAHA.106.628875>. |
Margeta, J. et al., “Fine-tuned convolutional neural nets for cardiac mri acquisition plane recognition”, Computer Methods in Biomechanics and Biomedical Engineering. Imaging & Visualization, Aug. 2015, pp. 1-11 <DOI:10.1080/21681163.2015.1061448>. |
Menchon-Lara, R-M. et al., Automatic Evaluation of Carotid Intima-Media Thickness in Ultrasounds Using Machine Learning, International Work-Conference on the Interplay Between Natural and Artifical Computation, 2013, vol. 7931, pp. 241-249 <DOI:10.1007/978-3-642-38622-0_25>. |
Menchon-Lara, R-M. et al., “Fully automatic segmentation of ultrasound common carotid artery images based on machine learning”, Neurocomputing, Mar. 2015 (available online Oct. 2014), vol. 151, No. 1, pp. 161-167 <DOI:10.1016/j.neucom.2014.09.066>. |
Mosinska, J. et al., “Active learning and proofreading for delineation of curvilinear structures”, Medical Image Computing and Computer Assisted Intervention—MICCAI (Quebec City, Canada, Sep. 11-13, 2017), 2017, pp. 165-173. |
Mozaffarian, D. et al., “Heart Disease and Stroke Statistics—2015 Update”, Circulation, Jan. 2015 (available online Dec. 2014), vol. 131, No. 4, pp. e29-e322 <DOI:10.1161/CIR.0000000000000152>. |
Pabby, A. et al., “Analysis of colorectal cancer occurrence during surveillance colonoscopy in the dietary Polyp Prevention Trial”, Gastrointestinal Endoscopy, Mar. 2005, vol. 61, No. 3, pp. 385-391 <DOI:10.1016/S0016-5107(04)02765-8>. |
Pan, S. et al., “A survey on transfer learning”, IEEE Transactions on Knowledge and Data Engineering, Oct. 2010 (Date of Publication: Oct. 16, 2009), vol. 22, No. 10, pp. 1345-1359 <DOI:10.1109/TKDE.2009.191>. |
Rabeneck, H. et al., “Outcomes of colorectal cancer in the united states: no change in survival (1986-1997).”, The American Journal of Gastroenterology, Feb. 2003, vol. 98, No. 2, pp. 471-477 <DOI:10.1111/i.1572-0241.2003.07260.x>. |
Sadigh, G. et al., “Challenges, controversies, and hot topics in pulmonary embolism imaging”, American Journal of Roentgenology, Mar. 2011, vol. 196, No. 3, pp. 497-515 <DOI:10.2214/AJR.10.5830>. |
Schlegl, J. et al., “Unsupervised pre-training across image domains improves lung tissue classification”, Medical Computer Vision: Algorithms for Big Data International Workshop (Cambridge, MA, Sep. 18, 2014), 2014, pp. 82-93. |
Alex, V., et al., “Generative adversarial networks for brain lesion detection,” in Medical Imaging 2017: Image Processing, vol. 10133, p. 101330G, International Society for Optics and Photonics, 2017. |
Anders, C. et al., “Understanding Patch-Based Learning by Explaining Predictions”, arXiv.org, submitted Jun. 2018, 7 pages, arXiv:1806.06926v1. |
Arjovsky, M., et al., “Wasserstein generative adversarial networks,” in International Conference on Machine Learning, pp. 214-223, 2017. |
Bai, W., et al., “Semi-supervised learning for network-based cardiac mr image segmentation,” in International Conference on Medical Image Computing and Computer-Assisted Intervention, pp. 253-260. Springer, 2017. |
Baumgartner, C.F., et al., “Visual feature attribution using wasserstein gans,” in Proc IEEE Comput Soc Conf Comput Vis Pattern Recognit, 2017. |
Cai, J., et al., “Iterative attention mining for weakly supervised thoracic disease pattern localization in chest x-rays,” in International Conference on Medical Image Computing and Computer-Assisted Intervention, pp. 589-598. Springer, 2018. |
Chang, H., et al., “Pairedcyclegan: Asymmetric style transfer for applying and removing makeup,” in 2018 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2018. |
Chen, X., et al., “Unsupervised detection of lesions in brain mri using constrained adversarial autoencoders,” arXiv preprint arXiv:1806.04972, 2018. |
Choi, Y., et al., “Stargan: Unified generative adversarial networks for multidomain image-to-image translation,” in the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Jun. 2018,https://github.com/yunjey/StarGAN. |
Goodfellow, I., et al., “Generative adversarial nets,” in Advances in neural information processing systems, pp. 2672-2680, 2014. |
Gulrajani, I., et al., “Improved training of wasserstein gans,” in Advances in Neural Information Processing Systems, pp. 5767-5777, 2017. |
Hwang, S., et al., “Self-transfer learning for weakly supervised lesion localization,” in Medical Image Computing and Computer-Assisted Intervention—MICCAI 2016, pp. 239-246, Springer International Publishing. |
Isola, P., et al., “Image-to-image translation with conditional adversarial networks,” arXiv preprint arXiv:1611.07004, 2016. |
Kim, T., et al., “Learning to discover cross-domain relations with generative adversarial networks,” in International Conference on Machine Learning, pp. 1857-1865, 2017. |
Kistler, M., et al., “The virtual skeleton database: An open access repository for biomedical research and collaboration,” J Med Internet Res, 15(11):e245, Nov. 2013. |
Ledig, C., et al., “Photo-realistic single image super-resolution using a generative adversarial network,” in CVPR, vol. 2, p. 4, 2017. |
Li, Z., et al., “Thoracic disease identification and localization with limited supervision,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 8290-8299, 2018. |
Liu, M.-Y., et al., “Unsupervised image-to-image translation networks,” in Advances in Neural Information Processing Systems, pp. 700-708, 2017. |
Liu, Z., et al., “Deep learning face attributes in the wild,” in Proceedings of the IEEE International Conference on Computer Vision, pp. 3730-3738, 2015. |
Menze, B. H., et al., “The multimodal brain tumor image segmentation benchmark (brats),” IEEE transactions on medical imaging, 34(10):1993, 2015. |
Non-final Office Action for U.S. Appl. No. 16/397,990, dated Sep. 15, 2020, 67 pages. |
Notice of Allowance for U.S. Appl. No. 16/397,990, dated Dec. 1, 2020, 11 pages. |
Notice of Allowance for U.S. Appl. No. 16/397,990, dated Jan. 14, 2021, 16 pages. |
Oquab, M., et al., “Is object localization for free?-weakly-supervised learning with convolutional neural networks,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 685-694, 2015. |
Pinheiro, P. O., et al., “From image-level to pixellevel labeling with convolutional networks,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1713-1721, 2015. |
Schlegl, T., et al., “f-anogan: Fast unsupervised anomaly detection with generative adversarial networks,” Medical Image Analysis, 2019, https://github.com/tSchiegl/f-AnoGAN. |
Selvaraju, R. R., et al., “Grad-cam: Visual explanations from deep networks via gradient-based localization,” in Proceedings of the IEEE International Conference on Computer Vision, pp. 618-626, 2017. |
Shin, S. Y., et al., “Joint weakly and semi-supervised deep learning for localization and classification of masses in breast ultrasound images,” IEEE transactions on medical imaging, 2018. |
Simonyan, K., et al., “Deep inside convolutional networks: Visualising image classification models and saliency maps,” arXiv preprint arXiv:1312.6034, 2013. |
Singh, K. K., et al., “Hide-and-seek: Forcing a network to be meticulous for weakly-supervised object and action localization,” in 2017 IEEE International Conference on Computer Vision (ICCV), pp. 3544-3553, IEEE, 2017. |
Tang, Y., et al., “Attention-guided curriculum learning for weakly supervised classification and localization of thoracic diseases on chest radiographs,” in International Workshop on Machine Learning in Medical Imaging, pp. 249-258, Springer, 2018. |
Wang, K., et al., “Cost-effective active learning for deep image classification” (2017), IEEE Transactions on Circuits and Systems for Video Technology 27, pp. 2591-2600. |
Wang, X., et al., “Chestx-ray8: Hospital-scale chest x-ray database and benchmarks on weakly-supervised classification and localization of common thorax diseases,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2097-2106, 2017. |
Wolterink, J. M., et al., “Deep mr to ct synthesis using unpaired data,” in International Workshop on Simulation and Synthesis in Medical Imaging, pp. 14-23, Springer, 2017. |
Yi, Z., “Dualgan: Unsupervised dual leaming for image-to-image translation,” in ICCV, pp. 2868-2876, 2017. |
Zhang, X., “Adversarial complementary learning for weakly supervised object localization,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1325-1334, 2018. |
Zhou, B., et al., “Learning deep features for discriminative localization,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2921-2929, 2016. |
Zhu, J.-Y., et al., “Toward multimodal image-to-image translation,” in Advances in Neural Information Processing Systems, pp. 465-476, 2017. |
Zhu, J.-Y., et al., “Unpaired image-to-image translation using cycle-consistent adversarial networks,” arXiv preprint, 2017. |
Number | Date | Country | |
---|---|---|---|
20180314943 A1 | Nov 2018 | US |
Number | Date | Country | |
---|---|---|---|
62663931 | Apr 2018 | US | |
62491069 | Apr 2017 | US |