Automated detection of controls in computer applications with region based detectors

Information

  • Patent Grant
  • 11775814
  • Patent Number
    11,775,814
  • Date Filed
    Wednesday, July 31, 2019
    5 years ago
  • Date Issued
    Tuesday, October 3, 2023
    a year ago
Abstract
Controls within images of a user interface of a computer application are detected by way of region-based R-FCN and Faster R-CNN engines. Datasets comprising images containing application control, wherein the application controls include images of application where width is greater than height, width is equal to height and height is greater than width are retrieved. Each of the datasets is processed with the R-FCN and Faster R-CNN engines to generate a software configured to recognize, from an input image, application controls wherein the application controls are characterized by dimensions where width is greater than height, where width is substantially equal to height, and where height is greater than width.
Description
FIELD OF THE DISCLOSURE

This disclosure relates generally to the field of data processing systems and more particularly to computerized image recognition.


BACKGROUND

Accurate identification and extraction of data from business documents is an important aspect of computerized processing of business documents. In particular, as the demand for automation increases, it is imperative to recognize controls in legacy application programs which do not provide programmatic access in order to automate usage of such applications. For websites, the code is available in one form or another so detection of controls and their type on the website is relatively straightforward. However, many licensed applications do not allow permit access to their code. Automated detection of controls on such applications for automation is a challenge. Detection of such controls could reduce monotonous tasks such as data entry and form filling, thereby potentially saving thousands of man-hours and leading to better accuracy.


Many attempts have been made previously to solve this problem using computer vision and general edge detection techniques. However, these techniques are not generalized and heavily rely on the completeness of edges. Hence, controls with broken edges cannot be detected by this approach. Moreover, visually appealing controls on modern webpages such as a textbox with just an underline are not detected by these techniques.


SUMMARY

While there have been significant advancements in object detection with state-of-the-art networks performing well on various domains, use of such networks for control detection on computer application screens is a novel approach. Disclosed herein is a computerized method, system and computer program product for automated detection of application level controls displayed by a computer application. A first dataset comprising images containing a first type of application control is retrieved, wherein each application control of the first type of application control is characterized by dimensions where width is greater than height. A second dataset comprising images containing a second type of application control is retrieved, wherein each application control of the second type of application control is characterized by dimensions where width is substantially equal to height. A third dataset comprising images containing a third type of application control is retrieved, wherein each application control of the third type of application control is characterized by dimensions where height is greater than width. The first, second and third datasets are each processed with a region-based R-FCN engine to generate a first trained region-based R-FCN engine that is trained to recognize application controls characterized by dimensions where width is greater than height, a second trained region-based R-FCN engine that is trained to recognize application controls characterized by dimensions where width is substantially equal to height, and a third trained region-based R-FCN engine that is trained to recognize application controls characterized by dimensions where height is greater than width. The first, second and third trained region-based R-FCN engines are combined to generate a software robot configured to recognize, from an input image, application controls wherein the application controls are characterized by dimensions where width is greater than height, where width is substantially equal to height, and where height is greater than width.


In other embodiments, the region-based R-FCN engines may be replaced with or be supplemented by faster R-CNN engines.


These and additional aspects related to the invention will be set forth in part in the description which follows, and in part will be apparent to those skilled in the art from the description or may be learned by practice of the invention. Aspects of the invention may be realized and attained by means of the elements and combinations of various elements and aspects particularly pointed out in the following detailed description and the appended claims.


It is to be understood that both the foregoing and the following descriptions are exemplary and explanatory only and are not intended to limit the claimed invention or application thereof in any manner whatsoever.





BRIEF DESCRIPTION OF THE DRAWINGS

The accompanying drawings, which are incorporated in and constitute a part of this specification exemplify the embodiments of the present invention and, together with the description, serve to explain and illustrate principles of the inventive techniques disclosed herein. Specifically:



FIG. 1 is high-level block diagram of a computerized system that automatically detects controls in application user interfaces.



FIGS. 2A, 2B, 2C and 2D are images of various controls employed by various sample application program user interfaces.



FIGS. 3 and 4 are block diagrams illustrating operation of preferred machine learning models.



FIGS. 5A, 5B, 5C and 5D are block diagrams illustrating operation of two alternative embodiments of bot generation employed by the system of FIG. 1.



FIG. 6 is a graph comparing performance of two preferred machine learning models (Faster R-CNN and R-CNN).



FIG. 7 is a graph showing class-wise performance using Faster R-CNN.



FIG. 8 is a graph showing log-average miss rate for various controls using Faster R-CNN.



FIG. 9 is a graph showing class-wise mean average precision (mAP) using R-FCN.



FIG. 10 is a graph showing log average missrate using R-FCN.



FIG. 11 is a graph showing predicted object information.



FIGS. 12A, 12B, 12C, 12D, and 12E are graphs showing P-R curves for various controls using Faster R-CNN.



FIGS. 13A, 13B, 13C, 13D, and 13E are graphs showing P-R curves for various controls using R-FCN.



FIG. 14 illustrates a block diagram of hardware that may be employed in an implementation of the systems disclosed herein.





DETAILED DESCRIPTION

In the following detailed description, reference will be made to the accompanying drawings, in which identical functional elements are designated with like numerals. Elements designated with reference numbers ending in a suffix such as .1, .2, .3 are referred to collectively by employing the main reference number without the suffix. For example, 100 refers to topics 100.1, 100.2, 100.3 generally and collectively. The aforementioned accompanying drawings show by way of illustration, and not by way of limitation, specific embodiments and implementations consistent with principles of the present invention. These implementations are described in sufficient detail to enable those skilled in the art to practice the invention and it is to be understood that other implementations may be utilized and that structural changes and/or substitutions of various elements may be made without departing from the scope and spirit of present invention. The following detailed description is, therefore, not to be construed in a limited sense.


Explained below in connection with the disclosed embodiments is an overview of various approaches that have been employed in object detection. Preferred embodiments for detecting controls on a computer screen using state-of-the-art object detection models are then described. The disclosed object detection-based approach is also able to detect controls without definite boundaries like trees, which was not possible using any of the previous approaches.


In the following description, to improve readability, reference is made to the following publications by the accompanying reference numbers below:

  • [1] Wei Liu, Dragomir Anguelov, Dumitru Erhan, Christian Szegedy, Scott Reed, Cheng-Yang Fu, and Alexander C Berg, Ssd: Single shot multibox detector, European conference on computer vision, pages 21-37. Springer, 2016.
  • [2] Shaoqing Ren, Kaiming He, Ross Girshick, and Jian Sun, Faster r-cnn: Towards real-time object detection with region proposal networks, Advances in neural information processing systems, pages 91-99, 2015.
  • [3] Jifeng Dai, Yi Li, Kaiming He, and Jian Sun, R-fcn: Object detection via region-based fully convolutional networks, Advances in neural information processing systems, pages 379-387, 2016.
  • [4] Navneet Dalal and Bill Triggs, Histograms of oriented gradients for human detection, International Conference on computer vision & Pattern Recognition (CVPR'05), volume 1, pages 886-893. IEEE Computer Society, 2005.
  • [5] M. A. Hearst, S. T. Dumais, E. Osuna, J. Platt, and B. Scholkopf, Support vector machines, IEEE Intelligent Systems and their Applications, 13(4):18-28, July 1998.
  • [6] J. Canny, A computational approach to edge detection, IEEE Transactions on Pattern Analysis and Machine Intelligence, PAMI-8(6):679-698, November 1986.
  • [7] Alex Krizhevsky, Ilya Sutskever, and Geoffrey E Hinton, Imagenet classification with deep convolutional neural networks, Advances in neural information processing systems, pages 1097-1105, 2012.
  • [8] Ross Girshick, Jeff Donahue, Trevor Darrell, and Jitendra Malik, Rich feature hierarchies for accurate object detection and semantic segmentation, Proceedings of the IEEE conference on computer vision and pattern recognition, pages 580-587, 2014.
  • [9] Jasper R R Uijlings, Koen E A Van De Sande, Theo Gevers, and Arnold W M Smeulders, Selective search for object recognition, International journal of computer vision, 104(2):154-171, 2013.
  • [10] Joseph Redmon, Santosh Divvala, Ross Girshick, and Ali Farhadi, You only look once: Unified, real-time object detection, Proceedings of the IEEE conference on computer vision and pattern recognition, pages 779-788, 2016.
  • [11] Ross Girshick, Fast r-cnn, Proceedings of the IEEE international conference on computer vision, pages 1440-1448, 2015.
  • [12] Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun, Deep residual learning for image recognition, Proceedings of the IEEE conference on computer vision and pattern recognition, pages 770-778, 2016.
  • [13] Christian Szegedy, Wei Liu, Yangqing Jia, Pierre Sermanet, Scott Reed, Dragomir Anguelov, Dumitru Erhan, Vincent Vanhoucke, and Andrew Rabinovich, Going deeper with convolutions, Proceedings of the IEEE conference on computer vision and pattern recognition, pages 1-9, 2015.
  • [14] Tsung-Yi Lin, Michael Maire, Serge Belongie, James Hays, Pietro Perona, Deva Ramanan, Piotr Dollár, and C Lawrence Zitnick, Microsoft coco: Common objects in context, European conference on computer vision, pages 740-755. Springer, 2014.
  • [15] Andrew G Howard, Menglong Zhu, Bo Chen, Dmitry Kalenichenko, Weijun Wang, Tobias Weyand, Marco Andreetto, and Hartwig Adam, Mobilenets: Efficient convolutional neural networks for mobile vision applications, arXiv preprint arXiv:1704.04861, 2017.
  • [16] Tsung-Yi Lin, Priya Goyal, Ross Girshick, Kaiming He, and Piotr Dollár, Focal loss for dense object detection, Proceedings of the IEEE international conference on computer vision, pages 2980-2988, 2017.


In developing the disclosed embodiments, experiments were performed with various state-of-the-art object detection models including Single Shot object detectors like SSD [1] and Region based detectors like Faster R-CNN [2] and R-FCN [3]. A comparative analysis of the results of the foregoing models was performed, taking into consideration mean Average Precision (mAP), loss and accuracy. While there have been many advances in the domain of object detection, region-based detectors have the advantage of a separate region proposal network and a classification network. This allows manipulation of these networks according to needs. Disclosed herein is an implementation of two region-based detectors namely faster region based Convolutional Neural Network (faster R-CNN) and Region Based Fully Convolutional Network (R-FCN). A comparison of results generated by both embodiments is also provided.


Detection of on-screen controls was performed using HOG [4] and SVM [5] but those approaches were only successful with proper region proposals. The reduced runtime of inferencing using this approach comes at the cost of accuracy. Since the region proposal algorithm is based on finding edges and then joining all the adjacent points having the same color or intensity, it is heavily dependent on the edge detection algorithm [6]. In many cases this algorithm fails to propose proper Regions Of Interest (ROI). There have been many approaches to solve the object detection problem using a variety of architectures, as Deep Neural Networks (DNNs) have shown to outperform traditional feature-based extraction methods.


The performance in object detection has significantly improved with the introduction of Convolutional Neural Networks (CNN). These CNN models have drastically improved the accuracy on image classification tasks as well. Since the introduction of AlexNet [7], many different CNN models have been developed to solve the problem of image classification and object detection. Current state-of-the-art object detectors are divided into 2 types: One stage detectors (Single Shot Detectors) and two stage detectors. The most common detectors are two stage detectors where the first stage is responsible for generating Regions of Interests (ROIs) and the second stage is responsible for the classification and regression task. This architecture was first proposed by Regions with CNN features (R-CNN) [8] which used Selective Search [9] for proposing regions and then classifying these regions using CNN. As a lot of time was consumed in proposing regions using the Selective Search algorithm, this stage was replaced with a Region Proposal Network (RPN) so that it can learn the features of the regions which are to be detected. Faster R-CNN [2] thus performed better and faster than its predecessors as the convolutional layers are shared and the weights of the networks are updated in an alternating fashion.


One stage detectors have been built to detect objects in real-time. Single Shot Detectors (SSD) [1] were introduced to create a scale invariant detector which can use multiple layers of multiple scales to perform classification. Redmon et al. [10] introduced a straightforward method to learn the bounding box coordinates of the object and the class score of the object in a region in one pass over an image.


In developing the disclosed embodiments, three datasets for three different models were prepared. The models are distinguished based on the features of the controls such as height, width, aspect ratios and scales. The first model, seen in FIG. 1 at 106 consists of the controls that share the feature of width being greater than the height. This includes textboxes, dropdown, pagetabs, buttons and horizontal scrollbars. The second model seen at 107 comprises of checkboxes, radio buttons and image-buttons, sharing the feature of maintaining aspect ratio of 1 (height=width) and the third model 108 consists of table, tree and vertical scrollbars, all of which have the height generally greater than the width. For each model, a training set, validation set and test set were prepared (109, 110, 111). The validation set is preferably prepared by randomly taking some of the images from the dataset, which are then omitted from the training portion. A large number of images were collected for these controls taking into consideration different resolutions, sizes and styles of each control. A considerable number of images were collected from legacy applications like SAP, Powerbuilder and Oracle E-Business Suite, in addition to web forms.


Examples of the various controls classified as 106, 107 or 108 may be seen in the screenshots shown in FIGS. 2A, 2B, 2C and 2D. Reference number 202 shows examples of buttons, reference number 203 shows examples of dropdown boxes, reference number 204 shows examples of textboxes, reference number 205 shows examples of pagetab and reference number 206 shows examples of scrollbars.


The images were then annotated using Labellmg, a graphical image annotation tool and label object bounding boxes in images. Further information regarding Labellmg may be found at Github.com, for example at: https://github.com/tzutalin/labellmg. The annotations were preferably performed in a manner such that features of every control such as border and color are retained. After consideration of all these scenarios, a total of 6764 controls were annotated. The class-wise bifurcation of these annotations is summarized in Table 1 below.









TABLE 1







Number of annotations for each class










Class
Annotations














Button
1655



Dropdown
1043



Pagetab
169



Scrollbars
436










Labellmg stores these annotations in XML files in Pascal-VOC Format. These files store important information about the bounding boxes such as xmin, ymin, xmax, ymax, width, height and class-name. These XML files are then converted to CSV files to carry out the data analysis of all these controls as discussed below and to generate a TensorFlow Record, which is then fed to the network. In addition to the training and validation set, a set of unseen images has been chosen for the test set. The test set is also preferably distributed in such a way that all the features and variations of controls are included. This ensures that the results are a generalized representation and are not biased.


Turning to FIG. 1, an application control detection system 10 employs the three models 106, 107 and 108. Separate training, validation and data sets are generated (109, 110, 111) are generated for each model 106, 107, 108, respectively, as described above. In the embodiment of FIG. 1, two convolutional networks (R-FCN 112 and Faster R-CNN 113) are applied to the data sets 109, 110, 111.


Region based detectors have a separate Region Proposal Network (RPN), that proposes the probable regions that may contain the object and their objectness score, and an object detection network that is responsible for regression of bounding boxes and classification. Since region-based networks have a separate region proposal and classification network, extra time has to be devoted for proposing regions. Hence, its main bottleneck is the time required in region proposals.


R-CNN and Fast R-CNN [11] use selective search to propose regions, which is slow with 2 seconds per proposal on CPU. Hence, the system 10 employs an implementation of Faster R-CNN, an improvement over R-CNN and Fast R-CNN. In contrast to these, faster R-CNN shares the same convolutional layer for region proposal and detection phases. Hence the major computational time required for region proposal in the previous two models is eliminated in faster R-CNN, giving near real time results. Moreover, since selective search is purely based on computer vision and shows no progressive learning with time, it is likely that faster R-CNN which learns to propose regions, given training examples will show better accuracy than its predecessors. Faster R-CNN also shows a unique training pattern. Faster R-CNN employs a Region Proposal Network (RPN) and a detection network that are first trained independently with an Imagenet pretrained model. The training then follows an alternating pattern, where the shared convolutions are fixed and the RPN and detection networks finetune their weights in an alternate fashion. The architecture of faster R-CNN is as shown in FIG. 3. This architecture is divided into three parts: 1) Feature network 302, 2) Region proposal network 304, and 3) Detection network 306. The Feature network 302 is responsible for generating features 303 from the input image. This usually consists of a pretrained classification network. In a preferred implementation, the feature extractor is a pretrained ResNet Network [12]. The reason for using residual networks is to prevent training loss from saturating. The region proposal network 304 employs a base network described in InceptionNet [13] that is pretrained on the MS COCO dataset [14]. This is the network responsible for generating areas of the image that have a high probability of containing object, called “Region of Interests” (ROI) 305. The reason why InceptionNet is used for region proposals is that it allows capture of objects of varying sizes due to its wider network. Hence the size of the control will not affect the region proposals. This network consists of one common layer which serves as input to two other layers, for classification and regression. The image first passes through the first stage of the network, during which the possible ROIs 305 will be generated. The region proposal step is separated from the classification step, which means the loss in generating the ROIs will be backpropagated through the region proposal network 304 only. The region proposal network (RPN) 304 will generate the compressed feature maps as output, which are then passed as input to the second stage. The detection network 306 takes input from both feature network 302 and RPN 304 and is responsible for final classification of identified objects and bounding box regression of the identified objects.


Model Hyper-Parameters. In the Faster R-CNN based implementation, the images have been rescaled to 600×1024. In one embodiment, the momentum used is 0.9. In this implementation, the initial learning rate is set to 0.0003. The learning rate is reduced to 0.00003 after 900000 steps, which is further reduced to 0.000003 after 1200000 steps. Also, an analysis of the control dimensions was carried out to select appropriate scale and aspect ratios for each model. The mean, median, minimum and maximum height and width for every control was programmatically calculated. A heuristic analysis was again performed on these values, finally resulting in appropriate scales and aspect ratios for each of the three models. Using properly selected values compared to the default values improved the learning rate and accuracy significantly, hence enhancing performance.


The system 10 also employs a Region Based Fully Convolutional Network (R-FCN) 112 to improve the performance of the pipeline. As R-FCN is a fully convolutional network, it takes less time for inference as compared to Faster R-CNN. Also, R-FCN uses region-based feature maps. These feature maps are independent of ROIs and hence they can be computed outside each ROI. R-FCN calculates the score map of each ROI with the ground truth to determine the objectness and class score. As R-FCN does not have a fully connected (FC) layer after the ROI pooling operation, it performs much faster than Faster R-CNN. By using R-FCN the inference time on a CPU was decreased by 2× as compared to the Faster R-CNN implementation. R-FCN also performed better in terms of accuracy than Faster R-CNN as showed in the graphs. The architecture of R-FCN is shown in FIG. 4. As shown in FIG. 4, the image first passes through a pretrained CNN 310 to generate feature maps 311. These feature maps 311 then pass through two convolutional layers 312 and 313 to generate ROIs 314 and position sensitive score maps 315. Position sensitive score maps 315 are the sub region specific feature maps of the object. The combined results of generated ROIs 314 and position sensitive score maps 315, are mapped into a vote array 318. This mapping is termed as position sensitive ROI pooling. The values of this vote array are averaged at 320 to generate a class score 322. The method employed for bounding box regression is almost the same. Other 4×k×k convolutional filters are generated from the same feature maps (the object is assumed to be divided into k×k subregions). Single shot detectors like SSD were tried but did have not depicted satisfactory performance for the said use case. A comparative analysis of all the approaches is presented in further herein.


Model Hyper-Parameters. All of the hyperparameters of R-FCN based implementation are the same as that of Faster R-CNN based implementation except the method of determining appropriate scales and aspect ratios and introduction of data augmentation, which are discussed in detail below.


Scales and Aspect Ratios. A K-means clustering algorithm is employed to determine proper values of scales and aspect ratios with K=3 since 3 scales and aspect ratios have been used in a default implementation of R-FCN. Using properly selected values compared to the default values improved the learning rate and accuracy significantly, hence enhancing performance.


Data Augmentation. To solve data imbalance among classes, various data augmentation methods were applied, namely random crop, random adjust brightness, random adjust contrast and RGB to grayscale. These augmentation methods solved data imbalance to a certain extent, increasing performance.


Results of the engines 112 and 113 are employed by bot generation engine 116 to generate one or more cognitive software robots (bots), seen generally at 102. The cognitive bots 102 may then be employed by a user 120 to process one or more documents 105 to generate mapped controls 122 where controls in a given application level user interface, such as shown in FIGS. 2A, 2B, 2C, and 2D are identified to enable automated usage by a software robot.



FIGS. 5A, 5B, 5C and 5D are block diagrams illustrating operation of four alternative embodiments of bot generation module 116 employed by the system of FIG. 1. In the embodiment of FIG. 5A, a bot 102 is configured to employ each R-FCN engine 112 to process an image 105. This permits recognition of the various controls for which each of the R-FCN engines (112.1, 112.2, 112.3) is optimized. The order of processing by the engines 112 shown in FIG. 5A is not critical and may be changed. Moreover, the processing need not be sequential but instead may be performed concurrently as seen in FIG. 5C. In the embodiment of FIG. 5B, a bot 102 is configured to employ each Faster R-CNN engine 113 to process an image 105. This permits recognition of the various controls for which each of the Faster R-CNN engines (113.1, 113.2, 113.3) is optimized. The order of processing by the engines 113 shown in FIG. 5B is not critical and may be changed. Moreover, the processing need not be sequential but instead may be performed concurrently as seen in FIG. 5D.


The bots 102 may be employed in a Robotic Process Automation (RPA) system such as available from Automation Anywhere, Inc. Such an RPA system implements a bot creator that may be used by a RPA user, to create one or more bots that are used to automate various business processes executed by one or more computer applications, such as the applications that generate the user interfaces seen in FIGS. 2A, 2B, 2C and 2D. The term “bot” as used herein refers to a set of instructions that cause a computing resource to interact with one or more user level computer applications to perform tasks provided by the one or more user level computer applications. Once created, the bot may be employed to perform the tasks as encoded by the instructions to interact with one or more user level computer applications.


In certain environments, the information provided by an application may contain sensitive information, the distribution or viewing of which may be subject to various regulatory or other restrictions. In such an environment, as described in U.S. patent application “DETECTION AND DEFINITION OF VIRTUAL OBJECTS IN REMOTE SCREENS”, Ser. No. 15/957,030, filed on Apr. 19, 2018, which application is hereby incorporated by reference in its entirety, an automation controller, resident on a computer system operates in conjunction with an RPA system to interact with another, remote, computer system. The RPA system sends automation commands and queries to the automation controller, while respecting the security compliance protocols of the remote computer system. As described, a compliance boundary may be implemented in connection with a remote access module. The compliance boundary represents a logical boundary, across which, any transfer of data or other information is controlled by agreements between parties. In certain embodiments, the remote access module may operate to prevent the RPA user from performing certain tasks on the remote system, by way of example and not limitation, copying files, loading cookies, or transmitting data from the remote computer system, through or beyond the compliance boundary via the internet or via any other output device that would violate the security protocols established by the remote computer system. The remote access module may take the form of remote desktop products available from Citrix or Microsoft, which permit connection to a remote computer, to establish a communication link between the user's system and the remote system to permit apps, files, and network resources to be made available. The system 10 described herein may be employed in the above described environment to permit recognition of the application controls provided by the application accessed across the aforementioned compliance boundary.


RESULTS. The results suggest that as the number of instances of the object increase in the dataset, the model performs pretty well for that class. This is evident from FIG. 8 wherein we can see that as the number of instances of horizontal scrollbar is the lowest its average miss rate is the highest at 0.44. Whereas the textbox class has the highest number of annotations and thus its average miss rate is the lowest at 0.05. This can be solved if we increase the dataset to incorporate more instances of minority class.


The models were evaluated against a validation set consisting of 28 images and carried out a comparative analysis of their mean average precision (mAP) for model-1 (textboxes, buttons, dropdowns, pagetabs and horizontal scrollbars). The set consisted of images of various applications previously unseen by the model.


Faster R-CNN Results. The Faster R-CNN implementation achieved a mAP of 75.37. The Intersection Over Union (IOU), mAP, classwise predictions and log average miss rate (lamr) are shown in Table 2 below, and FIGS. 6, 7, and 8 respectively.









TABLE 2







Intersection over Union (IoU) for each class using Faster R-CNN










Class
IoU







Button
0.8426



Dropdown
0.8274



Pagetab
0.8095



Scrollbar_horizontal
0.8556



Textbox
0.8334



Mean IoU
0.8368










The Precision-Recall (P-R) curve for each of the five classes of model-1 using faster R-CNN are shown in FIGS. 12A, 12B, 12C, 12D, and 12E which are graphs showing P-R curves for various controls using Faster R-CNN, and in FIGS. 13A, 13B, 13C, 13D, and 13E which are graphs showing P-R curves for various controls using R-FCN.


R-FCN Results. The R-FCN implementation achieved mAP of 88.93 using the R-. The Intersection Over Union (IOU), mAP, log average miss rate (lamr), and classwise predictions using R-FCN are shown in Table 3 below, FIG. 9, FIG. 10 and FIG. 11, respectively.









TABLE 3







Intersection over Union (IoU) for each class using Faster R-FCN








Class
IoU





Button
0.8941


Dropdown
0.8821


Pagetab
0.8213


Scrollbar_horizontal
0.8384


Textbox
0.8814


Mean IoU
0.8834









The Precision-Recall (P-R) curve for each of the five classes of model-1 using R-FCN are shown in FIGS. 13A, 13B, 13C, 13D and 13E.


The current network is significantly heavy and takes about 1-2 days for training on our dataset. In an alternative embodiment, a shallow network that retains the present accuracy and is computationally inexpensive compared to R-FCN may be employed. Also, training on different models such as SSD ResNet, MobileNet [15] or RetinaNet [16] can be employed. Additional embodiments may be specifically focused on additional controls like sliders, image buttons, tables, and toggle buttons. Moreover, since the dataset is a critical factor influencing the performance, the addition of more data will lead to enhanced performance.



FIG. 14 illustrates a block diagram of hardware that may be employed in an implementation of the RPA system as disclosed herein. FIG. 14 depicts a generalized example of a suitable general-purpose computing system 1400 in which the described innovations may be implemented in order to improve the processing speed and efficiency with which the computing system 1400 operates to perform the functions disclosed herein. With reference to FIG. 14 the computing system 1400 includes one or more processing units 1402, 1404 and memory 1406, 1408. The processing units 1402, 1406 execute computer-executable instructions. A processing unit can be a general-purpose central processing unit (CPU), processor in an application-specific integrated circuit (ASIC) or any other type of processor. The tangible memory 1406, 1408 may be volatile memory (e.g., registers, cache, RAM), non-volatile memory (e.g., ROM, EEPROM, flash memory, etc.), or some combination of the two, accessible by the processing unit(s). The hardware components in FIG. 14 may be standard hardware components, or alternatively, some embodiments may employ specialized hardware components to further increase the operating efficiency and speed with which the system 100 operates. The various components of computing system 1400 may be rearranged in various embodiments, and some embodiments may not require nor include all of the above components, while other embodiments may include additional components, such as specialized processors and additional memory.


Computing system 1400 may have additional features such as for example, storage 1410, one or more input devices 1414, one or more output devices 1412, and one or more communication connections 1416. An interconnection mechanism (not shown) such as a bus, controller, or network interconnects the components of the computing system 1400. Typically, operating system software (not shown) provides an operating system for other software executing in the computing system 1400, and coordinates activities of the components of the computing system 1400.


The tangible storage 1410 may be removable or non-removable, and includes magnetic disks, magnetic tapes or cassettes, CD-ROMs, DVDs, or any other medium which can be used to store information in a non-transitory way, and which can be accessed within the computing system 1400. The storage 1410 stores instructions for the software implementing one or more innovations described herein.


The input device(s) 1414 may be a touch input device such as a keyboard, mouse, pen, or trackball, a voice input device, a scanning device, or another device that provides input to the computing system 1400. For video encoding, the input device(s) 1414 may be a camera, video card, TV tuner card, or similar device that accepts video input in analog or digital form, or a CD-ROM or CD-RW that reads video samples into the computing system 1400. The output device(s) 1412 may be a display, printer, speaker, CD-writer, or another device that provides output from the computing system 1400.


The communication connection(s) 1416 enable communication over a communication medium to another computing entity. The communication medium conveys information such as computer-executable instructions, audio or video input or output, or other data in a modulated data signal. A modulated data signal is a signal that has one or more of its characteristics set or changed in such a manner as to encode information in the signal. By way of example, and not limitation, communication media can use an electrical, optical, RF, or other carrier.


The terms “system” and “computing device” are used interchangeably herein. Unless the context clearly indicates otherwise, neither term implies any limitation on a type of computing system or computing device. In general, a computing system or computing device can be local or distributed and can include any combination of special-purpose hardware and/or general-purpose hardware with software implementing the functionality described herein.


While the invention has been described in connection with a preferred embodiment, it is not intended to limit the scope of the invention to the particular form set forth, but on the contrary, it is intended to cover such alternatives, modifications, and equivalents as may be within the spirit and scope of the invention as defined by the appended claims.

Claims
  • 1. A computerized method for automated detection of application level controls displayed by a computer application, comprising: retrieving a first dataset comprising images containing a first type of application control, wherein each first type of application control of a plurality of first type of application controls in each image in the first dataset is characterized by dimensions where width of the first type of application control is greater than height of the first type of application control;retrieving a second dataset comprising images containing a second type of application control, wherein each second type of application control of a plurality of second type of application controls in each image in the second dataset is characterized by dimensions where width of the second type of application control is substantially equal to height of the second type of application control;retrieving a third dataset comprising images containing a third type of application control, wherein each third type of application control of a plurality of third type of application controls in each image in the third dataset is characterized by dimensions where height of the third type of application control is greater than width of the third type of application control;processing each of the first, second and third datasets with a region-based R-FCN engine to generate: a first trained region-based R-FCN engine that is trained to recognize application controls characterized by dimensions where width is greater than height,a second trained region-based R-FCN engine that is trained to recognize application controls characterized by dimensions where width is substantially equal to height, anda third trained region-based R-FCN engine that is trained to recognize application controls characterized by dimensions where height is greater than width; andcombining the first, second and third trained region-based R-FCN engines to generate a software robot configured to recognize, from an input image generated from a screen of an application program, application controls wherein the application controls are characterized by dimensions where width is greater than height, where width is substantially equal to height, or where height is greater than width.
  • 2. The computerized method of claim 1 wherein combining the first, second and third trained region-based R-FCN engines to generate a software robot comprises: combining the first, second and third trained region-based R-FCN engines to cause sequential processing by the first, second and third trained region-based R-FCN engines.
  • 3. The computerized method of claim 1 wherein combining the first, second and third trained region-based R-FCN engines to generate a software robot comprises: combining the first, second and third trained region-based R-FCN engines to cause concurrent processing by the first, second and third trained region-based R-FCN engines.
  • 4. The computerized method of claim 1 wherein each of the first, second and third trained region-based R-FCN engines comprises: a pretrained convolutional neural network that generates from an input image: a plurality of feature maps, anda second and a third convolutional neural network that each receive the plurality of feature maps, wherein the second convolutional neural network generates regions of interest from the feature maps and wherein the third convolutional neural network generates a plurality of position sensitive score maps;combining the regions of interest and the position sensitive score maps to generate a vote array that comprises a position sensitive regions of interest pooling; andaveraging values of the vote array to generate a class score.
  • 5. A computerized method for automated detection of application level controls displayed by a computer application, comprising: retrieving a first dataset comprising images containing a first type of application control, wherein each first type of application control of a plurality of first type of application controls in each image in the first dataset is characterized by dimensions where width of the first type of application control is greater than height of the first type of application control;retrieving a second dataset comprising images containing a second type of application control, wherein each second type of application control of a plurality of second type of application controls in each image in the second dataset is characterized by dimensions where width of the second type of application control is substantially equal to height of the second type of application control;retrieving a third dataset comprising images containing a third type of application control, wherein each third type of application control of a plurality of third type of application controls in each image in the third dataset is characterized by dimensions where height of the third type of application control is greater than width of the third type of application control;processing each of the first, second and third datasets with a region-based faster R-CNN engine to generate: a first trained region-based faster R-CNN engine that is trained to recognize application controls characterized by dimensions where width is greater than height,a second trained region-based faster R-CNN engine that is trained to recognize application controls characterized by dimensions where width is substantially equal to height, anda third trained region-based faster R-CNN engine that is trained to recognize application controls characterized by dimensions where height is greater than width, andcombining the first, second and third trained region-based faster R-CNN engines to generate a software robot configured to recognize, from an input image generated from a screen of an application program, application controls characterized by dimensions where width is greater than height, where width is substantially equal to height, or where height is greater than width.
  • 6. The computerized method of claim 5 wherein combining the first, second and third trained region-based faster R-CNN engines to generate a software robot comprises: combining the first, second and third trained region-based faster R-CNN engines to cause sequential processing by the first, second and third trained region-based faster R-CNN engines.
  • 7. The computerized method of claim 5 wherein combining the first, second and third trained region-based faster R-CNN engines to generate a software robot comprises: combining the first, second and third trained region-based faster R-CNN engines to cause concurrent processing by the first, second and third trained region-based faster R-CNN engines.
  • 8. The computerized method of claim 5 wherein each of the first, second and third trained region-based faster R-CNN engines comprises: a feature network that generates features from the input image,a region proposal network that generates from the features, regions of interest, wherein the regions of interest comprise areas of the input image that have a high probability of containing an object, anda detection network that generates from the regions of interest, classifications of identified objects and bounding box regressions of the identified objects.
  • 9. A robotic process automation system comprising: data storage for storing a plurality of images of application user interfaces, wherein each application user interface of the application user interfaces comprises one or more application controls usable by a human user to interact with an application that generates the application user interface;a server processor operatively coupled to the data storage and configured to execute instructions that when executed cause the server processor to recognize one or more of the application controls, by: retrieving a first dataset comprising images containing a first type of application control, wherein each first type of application control of a plurality of first type of application controls in each image in the first dataset is characterized by dimensions where width of the first type of application control is greater than height of the first type of application control;retrieving a second dataset comprising images containing a second type of application control, wherein each second type of application control of a plurality of second type of application controls in each image in the second dataset is characterized by dimensions where width of the second type of application control is substantially equal to height of the second type of application control;retrieving a third dataset comprising images containing a third type of application control, wherein each third type of application control of a plurality of third type of application controls in each image in the third dataset is characterized by dimensions where height of the third type of application control is greater than width of the third type of application control;processing each of the first, second and third datasets with a region-based R-FCN engine to generate: a first trained region-based R-FCN engine that is trained to recognize application controls characterized by dimensions where width is greater than height,a second trained region-based R-FCN engine that is trained to recognize application controls characterized by dimensions where width is substantially equal to height,a third trained region-based R-FCN engine that is trained to recognize application controls characterized by dimensions where height is greater than width; andcombining the first, second and third trained region-based R-FCN engines to generate a first software robot configured to recognize, from an input image, application controls wherein the application controls are characterized by dimensions where width is greater than height, where width is substantially equal to height, or where height is greater than width.
  • 10. The robotic process automation system of claim 9 wherein combining the first, second and third trained region-based R-FCN engines to generate a software robot comprises: combining the first, second and third trained region-based R-FCN engines to cause sequential processing by the first, second and third trained region-based R-FCN engines.
  • 11. The robotic process automation system of claim 9 wherein combining the first, second and third trained region-based R-FCN engines to generate a software robot comprises: combining the first, second and third trained region-based R-FCN engines to cause concurrent processing by the first, second and third trained region-based R-FCN engines.
  • 12. The robotic process automation system of claim 9 wherein each of the first, second and third region-based R-FCN engines comprises: a pretrained convolutional neural network that generates from an input image: a plurality of feature maps, anda second and a third convolutional neural network that each receive the plurality of feature maps, wherein the second convolutional neural network generates regions of interest from the feature maps and wherein the third convolutional neural network generates a plurality of position sensitive score maps;combining the regions of interest and the position sensitive score maps to generate a vote array that comprises a position sensitive regions of interest pooling; andaveraging values of the vote array to generate a class score.
  • 13. The robotic process automation system of claim 9 wherein the server processor is further configured to execute instructions that when executed cause the server processor to recognize one or more of the application controls, by: processing each of the first, second and third datasets with a region-based faster R-CNN engine to generate: a first trained region-based faster R-CNN engine that is trained to recognize application controls characterized by dimensions where width is greater than height,a second trained region-based faster R-CNN engine that is trained to recognize application controls characterized by dimensions where width is substantially equal to height, anda third trained region-based faster R-CNN engine that is trained to recognize application controls characterized by dimensions where height is greater than width; andcombining the first, second and third trained region-based faster R-CNN engines to generate a second software robot configured to recognize, from an input image, application controls characterized by dimensions where width is greater than height, where width is substantially equal to height, or where height is greater than width.
  • 14. The robotic process automation system of claim 13 wherein combining the first, second and third trained region-based faster R-CNN engines to generate a software robot comprises: combining the first, second and third trained region-based faster R-CNN engines to cause sequential processing by the first, second and third trained region-based faster R-CNN engines.
  • 15. The robotic process automation system of claim 13 wherein combining the first, second and third trained region-based faster R-CNN engines to generate a software robot comprises: combining the first, second and third trained region-based faster R-CNN engines to cause concurrent processing by the first, second and third trained region-based faster R-CNN engines.
  • 16. The robotic process automation system of claim 13 wherein each of the first, second and third trained region-based faster R-CNN engines comprises: a feature network that generates features from the input image,a region proposal network that generates from the features, regions of interest, wherein the regions of interest comprise areas of the input image that have a high probability of containing an object, anda detection network that generates from the regions of interest, classifications of identified objects and bounding box regressions of the identified objects.
US Referenced Citations (183)
Number Name Date Kind
5949999 Song et al. Sep 1999 A
5983001 Boughner et al. Nov 1999 A
6009192 Klassen et al. Dec 1999 A
6133917 Feigner et al. Oct 2000 A
6226407 Zabih et al. May 2001 B1
6389592 Ayres et al. May 2002 B1
6397215 Kreulen et al. May 2002 B1
6427234 Chambers et al. Jul 2002 B1
6473794 Guheen et al. Oct 2002 B1
6496979 Chen et al. Dec 2002 B1
6640244 Bowman-Amuah Oct 2003 B1
6704873 Underwood Mar 2004 B1
6898764 Kemp May 2005 B2
6954747 Wang et al. Oct 2005 B1
6957186 Guheen et al. Oct 2005 B1
7091898 Arling et al. Aug 2006 B2
7246128 Jordahl Jul 2007 B2
7398469 Kisamore et al. Jul 2008 B2
7441007 Kirkpatrick et al. Oct 2008 B1
7533096 Rice et al. May 2009 B2
7568109 Powell et al. Jul 2009 B2
7571427 Wang et al. Aug 2009 B2
7765525 Davidson et al. Jul 2010 B1
7805317 Khan et al. Sep 2010 B2
7805710 North Sep 2010 B2
7810070 Nasuti et al. Oct 2010 B2
7846023 Evans et al. Dec 2010 B2
8028269 Bhatia et al. Sep 2011 B2
8056092 Allen et al. Nov 2011 B2
8095910 Nathan et al. Jan 2012 B2
8132156 Malcolm Mar 2012 B2
8209738 Nicol et al. Jun 2012 B2
8234622 Meijer et al. Jul 2012 B2
8245215 Extra Aug 2012 B2
8352464 Fotev Jan 2013 B2
8365147 Grechanik Jan 2013 B2
8396890 Lim Mar 2013 B2
8438558 Adams May 2013 B1
8443291 Ku et al. May 2013 B2
8464240 Fritsch et al. Jun 2013 B2
8498473 Chong et al. Jul 2013 B2
8504803 Shukla Aug 2013 B2
8631458 Banerjee Jan 2014 B1
8682083 Kumar et al. Mar 2014 B2
8713003 Fotev Apr 2014 B2
8724907 Sampson et al. May 2014 B1
8769482 Batey et al. Jul 2014 B2
8819241 Washburn Aug 2014 B1
8832048 Lim Sep 2014 B2
8874685 Hollis et al. Oct 2014 B1
8943493 Schneider Jan 2015 B2
8965905 Ashmore et al. Feb 2015 B2
9032314 Mital et al. May 2015 B2
9104294 Forstall et al. Aug 2015 B2
9171359 Lund Oct 2015 B1
9213625 Schrage Dec 2015 B1
9278284 Ruppert et al. Mar 2016 B2
9444844 Edery et al. Sep 2016 B2
9462042 Shukla et al. Oct 2016 B2
9571332 Subramaniam et al. Feb 2017 B2
9600519 Schoning et al. Mar 2017 B2
9621584 Schmidt et al. Apr 2017 B1
9946233 Brun et al. Apr 2018 B2
9965139 Nychis May 2018 B2
9990347 Raskovic et al. Jun 2018 B2
10015503 Ahammad Jul 2018 B1
10025773 Bordawekar et al. Jul 2018 B2
10043255 Pathapati et al. Aug 2018 B1
10078743 Baldi et al. Sep 2018 B1
10282280 Gouskova May 2019 B1
10489682 Kumar et al. Nov 2019 B1
10654166 Hall May 2020 B1
10706218 Milward et al. Jul 2020 B2
11176443 Selva Nov 2021 B1
11182178 Singh et al. Nov 2021 B1
11243803 Anand et al. Feb 2022 B2
11348353 Sundell et al. May 2022 B2
11614731 Anand et al. Mar 2023 B2
20020029232 Bobrow et al. Mar 2002 A1
20030033590 Leherbauer Feb 2003 A1
20030101245 Srinivasan et al. May 2003 A1
20030114959 Sakamoto Jun 2003 A1
20030159089 DiJoseph Aug 2003 A1
20040083472 Rao et al. Apr 2004 A1
20040153649 Rhoads Aug 2004 A1
20040172526 Tann et al. Sep 2004 A1
20040210885 Wang et al. Oct 2004 A1
20040243994 Nasu Dec 2004 A1
20050002566 Federico et al. Jan 2005 A1
20050144462 LaGarde Jun 2005 A1
20050188357 Derks et al. Aug 2005 A1
20050204343 Kisamore et al. Sep 2005 A1
20050257214 Moshir et al. Nov 2005 A1
20060095276 Axelrod et al. May 2006 A1
20060150188 Roman et al. Jul 2006 A1
20060218110 Simske et al. Sep 2006 A1
20070030528 Quaeler et al. Feb 2007 A1
20070101291 Forstall et al. May 2007 A1
20070112574 Greene May 2007 A1
20070156677 Szabo Jul 2007 A1
20080005086 Moore Jan 2008 A1
20080027769 Eder Jan 2008 A1
20080028392 Chen et al. Jan 2008 A1
20080133052 Jones Jun 2008 A1
20080209392 Able et al. Aug 2008 A1
20080222454 Kelso Sep 2008 A1
20080263024 Landschaft et al. Oct 2008 A1
20090037509 Parekh et al. Feb 2009 A1
20090103769 Milov et al. Apr 2009 A1
20090116071 Mantell May 2009 A1
20090172814 Khosravi et al. Jul 2009 A1
20090199160 Vaitheeswaran et al. Aug 2009 A1
20090217309 Grechanik et al. Aug 2009 A1
20090249297 Doshi et al. Oct 2009 A1
20090313229 Fellenstein et al. Dec 2009 A1
20090320002 Peri-Glass et al. Dec 2009 A1
20100013848 Hekstra et al. Jan 2010 A1
20100023602 Martone Jan 2010 A1
20100023933 Bryant et al. Jan 2010 A1
20100100605 Allen et al. Apr 2010 A1
20100106671 Li et al. Apr 2010 A1
20100138015 Colombo et al. Jun 2010 A1
20100235433 Ansari et al. Sep 2010 A1
20100251163 Keable Sep 2010 A1
20110022578 Fotev Jan 2011 A1
20110106284 Catoen May 2011 A1
20110145807 Molinie et al. Jun 2011 A1
20110197121 Kletter Aug 2011 A1
20110276568 Fotev Nov 2011 A1
20110276946 Pletter Nov 2011 A1
20110302570 Kurimilla et al. Dec 2011 A1
20120011458 Xia et al. Jan 2012 A1
20120042281 Green Feb 2012 A1
20120076415 Kahn Mar 2012 A1
20120124062 Macbeth et al. May 2012 A1
20120131456 Lin et al. May 2012 A1
20120143941 Kim Jun 2012 A1
20120154633 Rodriguez Jun 2012 A1
20120324333 Lehavi Dec 2012 A1
20120330940 Caire et al. Dec 2012 A1
20130173648 Tan et al. Jul 2013 A1
20130236111 Pintsov Sep 2013 A1
20130290318 Shapira et al. Oct 2013 A1
20140036290 Miyagawa Feb 2014 A1
20140045484 Kim et al. Feb 2014 A1
20140181705 Hey et al. Jun 2014 A1
20140189576 Cami Jul 2014 A1
20150082280 Betak et al. Mar 2015 A1
20150310268 He Oct 2015 A1
20150347284 Hey et al. Dec 2015 A1
20160019049 Kakhandiki et al. Jan 2016 A1
20160034441 Nguyen et al. Feb 2016 A1
20160078368 Kakhandiki et al. Mar 2016 A1
20160119285 Kakhandiki et al. Apr 2016 A1
20170060108 Kakhandiki et al. Mar 2017 A1
20170154230 Dow et al. Jun 2017 A1
20170213130 Khatri et al. Jul 2017 A1
20170270431 Hosabettu Sep 2017 A1
20170330076 Valpola Nov 2017 A1
20180004823 Kakhandiki et al. Jan 2018 A1
20180113781 Kim Apr 2018 A1
20180157386 Su Jun 2018 A1
20180173698 Dubey et al. Jun 2018 A1
20180218429 Guo et al. Aug 2018 A1
20180275835 Prag Sep 2018 A1
20190005050 Proux Jan 2019 A1
20190028587 Unitt Jan 2019 A1
20190126463 Purushothaman May 2019 A1
20190141596 Gay May 2019 A1
20190188462 Nishida Jun 2019 A1
20190213822 Jain Jul 2019 A1
20190266692 Stach et al. Aug 2019 A1
20190317803 Maheshwari Oct 2019 A1
20190324781 Ramamurthy Oct 2019 A1
20190340240 Duta Nov 2019 A1
20190377987 Price et al. Dec 2019 A1
20200019767 Porter et al. Jan 2020 A1
20200034976 Stone et al. Jan 2020 A1
20200097742 Kumar et al. Mar 2020 A1
20200151591 Li May 2020 A1
20200249964 Fernandes Aug 2020 A1
20220245936 Valk Aug 2022 A1
20220405094 Farquhar Dec 2022 A1
Foreign Referenced Citations (2)
Number Date Country
2019092672 May 2019 WO
2022076488 Apr 2022 WO
Non-Patent Literature Citations (57)
Entry
Dai, Jifeng, et al. “R-fcn: Object detection via region-based fully convolutional networks.” Advances in neural information processing systems 29 (2016). (Year: 2016).
Ren, Shaoqing, et al. “Faster r-cnn: Towards real-time object detection with region proposal networks.” Advances in neural information processing systems 28 (2015). (Year: 2015).
Chollet, Francois et al., Keras: The Python Deep Learning Library, GitHub, https://github/keras-team/keras, 2015.
Al Sallami, Load Balancing in Green Cloud Computation, Proceedings of the World Congress on Engineering 2013 vol. II, WCE 2013, 2013, pp. 1-5 (Year: 2013).
B.P. Kasper “Remote: a Means of Remotely Controlling and Storing Data from a HAL Quadrupole Gass Analyzer Using an IBM-PC Compatible Computer”, Nov. 15, 1995, Space and Environment Technology Center.
Bergen et al., RPC automation: making legacy code releant, May 2013, 6 pages.
Hu et al., Automating GUI testing for Android applications, May 2011, 7 pages.
Konstantinou et al., An architecture for virtual solution composition and deployment in infrastructure clouds, 9 pages (Year: 2009).
Nyulas et al., An Ontology-Driven Framework for Deploying JADE Agent Systems, 5 pages (Year: 2006).
Tom Yeh, Tsung-Hsiang Chang, and Robert C. Miller, Sikuli: Using GUI Screenshots for Search and Automation, Oct. 4-7, 2009, 10 pages.
Yu et al., Deplying and managing Web services: issues, solutions, and directions, 36 pages (Year: 2008).
Zhifang et al., Test automation on moble device, May 2010, 7 pages.
Non-Final Office Action for U.S. Appl. No. 17/230,492, dated Oct. 14, 2022.
Notice of Allowance for U.S. Appl. No. 16/398,532, dated Oct. 23, 2022.
Non-Final Office Action for U.S. Appl. No. 16/876,530, dated Sep. 29, 2020.
Final Office Action for U.S. Appl. No. 16/876,530, dated Apr. 13, 2021.
Notice of Allowance for U.S. Appl. No. 16/876,530, dated Jul. 22, 2021.
Dai, Jifeng et al., “R-fcn: Object detectiom via region-based fully convolutional networks”, Advances in neural information processing systems 29 (2016). (Year: 2016).
Ren, Shaoqing et al., “Faster r-cnn: Towards real0time object detection with region proposal network.” Advances in neutral information processing systems 28 (2015). (Year: 2015).
Zhifang et al., Test automation on mobile device, May 2010, 7 pages.
International Search Report for PCT/US2021/053669, dated May 11, 2022.
Embley et al., “Table-processing paradigms: a research survey”, International Journal on Document Analysis and Recognition, vol. 8, No. 2-3, May 9, 2006, pp. 66-86.
Non-Final Office Action for U.S. Appl. No. 16/925,956, dated Sep. 16, 2021.
Notice of Allowance for U.S. Appl. No. 16/925,956, dated Jan. 7, 2022.
Pre-Interview Office Action for U.S. Appl. No. 16/398,532, dated Jul. 8, 2022.
Notice of Allowance for U.S. Appl. No. 16/398,532, dated Jul. 8, 2022.
Non-Final Office Action for U.S. Appl. No. 17/139,838, dated Feb. 22, 2022.
Final Office Action for U.S. Appl. No. 17/139,838, dated Nov. 15, 2023.
Notice of Allowance for U.S. Appl. No. 17/139,838, dated Apr. 5, 2023.
International Search Report and Written Opinion for PCT/US2021/015691, dated May 11, 2021.
A density-based algorithm for discovering clusters in large spatial databases with noise, Ester, Martin; Kriegel, Hans-Peter; Sander, Jorg; Xu, Xiaowei, Simoudis, Evangelos; Han, Jiawei; Fayyad, Usama M., eds., Proceedings of the Second International Conference on Knowledge Discovery and Data Mining (KDD-96). AMI Press. pp. 226-231 (1996).
Deep Residual Learning for Image Recognition, by K. He, X. Zhang, S. Ren, and J. Sun, arXiv:1512.03385 (2015).
FaceNet: A Unified Embedding for Face Recognition and Clustering, by F. Schroff, D. Kalenichenko, J. Philbin, arXiv:1503.03832 (2015).
Muhammad et al. “Fuzzy multilevel graph embedding”, copyright 2012 Elsevier Ltd.
Sharma et al. Determining similarity in histological images using graph-theoretic description and matching methods for content-based image retrieval in medical diagnostics, Biomed Center, copyright 2012.
First Action Interview Pilot Program Pre-Interview communication for U.S. Appl. No. 16/779,462, dated Dec. 3, 2021.
Reply under 37 CDT 1.111 to Pre-Interview Communication for U.S. Appl. No. 16/779,462, filed Jan. 25, 2022.
Notice of Allowance for U.S. Appl. No. 16/779,462 dated Feb. 9, 2022.
Notice of Allowance for U.S. Appl. No. 17/131,674, dated Jun. 22, 2023.
Non-Final Office Action for U.S. Appl. No. 16/731,044, dated Jan. 25, 2021.
Notice of Allowance for U.S. Appl. No. 16/731,044, dated May 5, 2021.
Alex Krizhevsky, Ilya Sutskever, and Geoffrey E Hinton, Imagenet classification with deep convolutional neural networks, Advances in neural information processing systems (2012).
Andrew G Howard, Menglong Zhu, Bo Chen, Dmitry Kalenichenko, Weijun Wang, Tobias Weyand, Marco Andreetto, and Hartwig Adam, Mobilenets: Efficient convolutional neural networks for mobile vision applications, arXiv:1704.04861v1 (2017).
Christian Szegedy, Wei Liu, Yangqing Jia, Pierre Sermanet, Scott Reed, Dragomir Anguelov, Dumitru Erhan, Vincent Vanhoucke, and Andrew Rabinovich, Going deeper with convolutions, Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 1-9 (2015).
J. Canny, A computational approach to edge detection, IEEE Transactions on Pattern Analysis and Machine Intelligence, PAMI-8(6):679- 698, Nov. 1986.
Jasper RR Uijlings, Koen EA Van De Sande, Theo Gevers, and Arnold WM Smeulders, Selective search for object recognition, International journal of computer vision, 104(2) (2013).
Jifeng Dai, Yi Li, Kaiming He, and Jian Sun, R-fcn: Object detection via region-based fully convolutional networks, Advances in neural information processing systems, 2016.
Joseph Redmon, Santosh Divvala, Ross Girshick, and Ali Farhadi, You only look once: Unified, real-time object detection, Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 779-788, 2016.
Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun, Deep residual leaming for image recognition, Proceedings of the IEEE conference on computer vision and pattern recognition (2016).
M. A. Hearst, S. T. Dumais, E. Osuna, J. Platt, and B. Scholkopf, Support vector machines, IEEE Intelligent Systems and their Applications, 13(4):18-28, Jul. 1998.
Navneet Dalal and Bill Triggs, Histograms of oriented gradients for human detection, International Conference on computer vision & Pattern Recognition (CVPR'05), vol. 1, pp. 886-893. IEEE Computer Society, 2005.
Ross Girshick, Fast r-cnn, Proceedings of the IEEE international conference on computer vision (2015).
Ross Girshick, Jeff Donahue, Trevor Darrell, and Jitendra Malik, Rich feature hierarchies for accurate object detection and semantic segmentation, Proceedings of the IEEE conference on computer vision and pattern recognition (2014).
Shaoqing Ren, Kaiming He, Ross Girshick, and Jian Sun, Faster r-cnn: Towards real-time object detection with region proposal networks, Advances in neural information processing systems, 2015.
Tsung-Yi Lin, Michael Maire, Serge Belongie, James Hays, Pietro Perona, Deva Ramanan, Piotr Dollár, and C Lawrence Zitnick, Microsoft coco: Common objects in context, European conference on computer vision (2014).
Tsung-Yi Lin, Priya Goyal, Ross Girshick, Kairning He, and Piotr Dollár, Focal loss for dense object detection, Proceedings of the IEEE international conference on computer vision (2017).
Wei Liu, Dragomir Anguelov, Durnitru Erhan, Christian Szegedy, Scott Reed, Cheng-Yang Fu, and Alexander C Berg, Ssd: Single shot multibox detector, European conference on computer vision, pp. 21-37. Springer, 2016.