The disclosures made in U.S. Provisional Application No. 63/242,145, filed Sep. 9, 2021, are specifically incorporated by reference herein as if set forth in their entirety.
In one aspect, the present disclosure is directed to surveillance systems and methods, and more specifically, to surveillance systems and methods that facilitate collection and analysis, in relation to various factors, electronic signatures and/or visual identifiers. Other aspects also are described.
Automated License Plate Readers (“ALPR”) typically are used for identifying vehicles in selected locations, e.g., for detecting traffic violations, collecting tolls, etc., However, existing ALPR systems are quite expensive and generally are used for identification of vehicles on roads, in parking lots, other vehicle throughways, etc . . . .
It can be seen that a need exists for surveillance systems and methods that can be used in conjunction with or in place of existing ALPR systems to provide for more precise, reliable, and/or consistent identification, tracking, and/or analysis of vehicles and/or persons.
The present disclosure is directed to the foregoing and other related, and unrelated, problems in the relevant art.
Briefly described, the present disclosure includes surveillance systems and methods for collecting electronic signatures and/or visual identifiers of targets to facilitate association of such target identifiers with collected electronic signature information to help create identification records for known or identified targets, e.g. targets that can be visually identifiable (such as by a license plate read, car make and model, or other visual identifier) can be linked to or otherwise associated with one or more particular electronic signatures. In embodiments, the surveillance systems and methods can utilize an event analysis system or engine and methodologies to classify reads of target identifiers and alarms or “hits” indicating presence of a known target (e.g., a stolen or suspect vehicle, etc . . . ) across subsets of locations and times of occurrence of such hits or alarms for enhancing tracking and movement of targets throughout selected geographic areas or locations. Targets generally refer to persons, vehicles, e.g., an automobile, or both, such a one or more persons within a vehicle. However, targets can include other objects, such as one or more electronic devices, e.g., cell phones or other communication devices, RFID and other sensors or transmitting devices internal to vehicles or as after-market additions, and/or various other, similar devices, without departing from the scope of the present disclosure.
According to aspects of the present disclosure, the surveillance system can include a plurality of collection systems or assemblies each located at selected geographic areas or locations. The collection systems generally are configured to capture or facilitate collection of information related to visual identifiers or electronic signatures associated with targets in or moving about the selected areas/locations. Each capture or collection may be associated with and/or stored with a particular time, date, and location data.
In some embodiments, the collection systems can include at least one sensor configured to collect or otherwise capture information related to visual identifiers and/or electronic signatures of targets. The visual identifiers can include visual vehicle identifiers, such as license plate information or other visual or imaged information associated with vehicles (e.g., stickers, patterns, position(s) of component parts, after-market added parts, damage, and/or various other markings, etc . . . ) that can be used to distinguish or otherwise identify, detect or discern a target vehicle, etc . . . . The electronic signatures can include an electronic signal or combination(s) of electronic signals emanating from transmitting electronic devices and which are associated with and/or can uniquely identify the targets in or moving about the selected areas/locations.
In addition, in some aspects, the surveillance system can include an intelligence system that is in communication with the plurality of collection systems. The intelligence system is configured to receive the information collected or captured by the collection systems, such as license plate reads such as images, video or other captured information from ALPR's that generate hits or alarms; which generally will be time and date stamped information in real time when targets get within proximity of the collection point systems (e.g., ALPR's), and to identify and/or track the targets based on this received information. In addition, in embodiments, the one or more classification and search engines can infer relationships between electronic devices and targets based on consistency of correlation to identify/extract electronic signatures associated with identified targets (based on known identifiers such as vehicle plate, make & model, and other associated information based on historical correlations with a particular target.
In embodiments, the intelligence system can include one or more classification and search engines and an intelligence database in communication therewith. The one or more classification and search engines are configured to identify or extract the electronic signatures associated with the targets using the summary tables/collections of information from the collection systems and applying statistical analysis to break down classes of the collected data in a manner that allows cataloging and visual navigation of statistically significant variations in the date, day-of-week, time, location, violation class, investigative case and/or specific vehicle license plate or plates. This allows an investigator to forecast the presence of a violation type, vehicle, or group of vehicles across time and location through a simple graphic charting interface.
In an embodiment, a method of operation of the surveillance system, includes a three-factor variable set structure for initially fixing day-of-week; time-of-day and source or classification-of-source. The method then allows the user to select 0, 1 or 2 factors and view the variation in the data by the non-selected factor. For example, the user can analyze crime data in data history by day, time and crime-type and then can select a specific day and type and thus be shown the distribution of incidents in history by time across the selected day and type thereby isolating a maximum frequency of incident across the selected variables.
The user can use a frequency and consistency of such crime data electronic signals to deploy investigative or enforcement resources based on the statistical distribution of alarms of the type that is of interest. Thus, a manager who wants to arrest drug suspects, can apply the disclosed statistical analysis to look at when and where drug alarms have occurred in the past, and thereafter can deploy his resources to the locations most likely to generate new alarms. The manager further can use such data to select alarm classes/crime activity based on selected variables such as days of the week, times, etc., in addition to types of crimes, and can determine a relative certainty of association of the selected targets to develop possible predictive patterns. Further, such relationships and/or associations can be visualized in the graphical or web-based user interfaces.
The one or more classification and search engines can be configured to filter or otherwise alter to the received electronic signatures, e.g., to reduce signal noise and facilitate identification or extraction of unique, identifying electronic signatures. In embodiments, the one or more classification and search engines can be configured to locate receipt of a visual identifier and correlated electronic signature to track the target, such as based on a visual vehicle identifier and correlated signatures, to allow independent tracking and location identification of targets based on the associated identifying electronic signatures. That is, once the system has records correlating electronic signatures associated with a specific visual vehicle identifier, e.g., a specific license plate number, the intelligence system will be able to detect the likely presence of a vehicle and its associated license plate without visual information, e.g., without the use of a camera.
The surveillance system thus is configurable to track, map, catalogue, etc., movements of the targets in real time as electronic signals emanating therefrom occur in proximity to the collection systems. The tracking information generated can be used to help confirm and/or authenticate a potential target identification, and further can be configured to generate alerts or notifications when certain targets are in proximity to the collection systems and can be used, as noted, in a number of user selectable visualizations.
In embodiments, the collection systems can be placed in locations or areas not associated with vehicular traffic, such that the intelligence system will be able to identify, and catalogue known electronic signatures away from the vehicles they have typically been associated with, e.g., for tracking, mapping, etc. of persons or electronic devices apart from vehicles. Each collection system can include a plurality of sensor assemblies, such as one or more cameras or camera systems configured to capture or facilitate collection of information related to vehicle identifiers, such as visual information related to a license plate of a vehicle or other visual vehicle identifiers. In addition, the sensor assemblies can include one or more antennas or other signal receivers configured to capture information related to the electronic signatures. The one or more antennas can include a plurality of antennas, such as a Bluetooth® antenna, a WiFi antenna, a RFID antenna, or other RF antennas or combinations thereof, configured to capture information related to electronic signals associated with the targets.
In some embodiments, the collection systems can be used in conjunction with Automated License Plate Readers (“ALPR”) in certain areas, allowing the intelligence system to develop a subset of electronic signals, i.e., an electronic signature, associated with a license plate read at a moment in time and location. Electronic data points from less expensive collectors can then be used to provide more precise tracking than ALPR alone.
In some embodiments, the surveillance system can be configured to capture sample electronic signature information from a target, associate that information with the target's identification, and then search for or alert on receipts of similar electronic signature information at one of the collection point systems.
In additional embodiments, the surveillance system can be configured to allow for search inquiries or scans of suspect's electronic signatures to search known location data points in the database history, placing the suspect at those locations and times.
In still other embodiments, the surveillance system can be configured to allow for labeling of specific electronic signatures with a target and then alert or search for history of those specific electronic signatures in the database, placing the target at various locations.
In further embodiments, the surveillance system further can indicate or determine changes in association or travel of suspects or other individuals of interest based on variations in electronic signatures associated with a target or targets.
Various objects, features, and advantages of the present disclosure will become apparent to those skilled in the art upon a review of the following detail description, when taken in conjunction with the accompanying drawings.
It will be appreciated that for simplicity and clarity of illustration, elements illustrated in the Figures are not necessarily drawn to scale. For example, the dimensions of some elements may be exaggerated relative to other elements. Embodiments incorporating teachings of the present disclosure are shown and described with respect to the drawings herein, in which:
The use of the same reference symbols in different drawings indicates similar or identical items.
The following description in combination with the Figures is provided to assist in understanding the teachings disclosed herein. The description is focused on specific implementations and embodiments of the teachings, and is provided to assist in describing the teachings. This focus should not be interpreted as a limitation on the scope or applicability of the teachings.
In embodiments, the present disclosure is directed to surveillance systems and methods for event analysis of collected event data, such as alarms or hits indicative of illegal activities (e.g., detection of stolen vehicles or vehicles identified with a criminal activity by ALPR systems) or a known target (e.g., a known criminal or wanted suspect/fugitive). The surveillance system and methods will employ a statistical analysis and projection of classifications for read (e.g., license plate reads, and other recognition data) and alarm data (e.g. flagged hits or alarms generated by such reads) across subsets of time and location of the collection of such read and alarms data. From a business/investigative perspective the system and methods are directed to enabling a manager or operator to deploy investigative or enforcement resources based on the statistical distribution of alarms of the type that is of interest. That is, if a manager wants to arrest drug suspects, he might want to look at when and where drug alarms have occurred in the past, and deploy his resources to the locations most likely to generate new alarms.
In embodiments, summary tables and statistical analysis are used to break down classes of data in a manner that allows visual navigation of statistically significant variations in the date, day-of-week, time, location, violation class, investigative case and/or specific vehicle license plate or plates. This allows an investigator to forecast the presence of a violation type, vehicle, or group of vehicles across time and location through a simple graphic charting interface. In addition, a three-factor variable set structure can be used for initially fixing day-of-week; time-of-day and source or classification-of-source. The method then allows the user to select 0, 1, or 2 factors and view the variation in the data by the non-selected factor. For example, the user can analyze crime data in data history by day, time, and crime-type and then can select a specific day and type and thus be shown the distribution of incidents in history by time across the selected day and type thereby isolating a maximum frequency of incident across the selected variables.
As indicated in
In addition, or in the alternative, the at least one sensor or sensor assembly also can include an antenna 114, antenna array, or plurality of antennas configured to capture or otherwise receive electronic signals from transmitting electronic devices 120, 122 associated with the targets for identification/extraction of electronic signatures. The at least one sensor or sensor assembly can include additional sensors, such as IR sensors or other light sensors, without departing from the present disclosure.
The transmitting electronic devices, as shown in
Each sensor or sensor assembly providing collection and input of event data, e.g. hits/alarms from ALPR reads, electronic signature records, etc. will be configured to capture or collect signals transmitted by or otherwise emanating from the transmitting electronic devices 202 when the targets get within proximity of the collection systems. The collection systems can be configured to receive signals at a prescribed or selected proximity in relation thereto. For example, in some embodiments, the collection systems could be configured to look for and receive signals (e.g., for example, via Wifi 204, Bluetooth 206, RFID detection, or via another method of detecting and receiving electronic signals) transmitted within about 200 feet of the collection systems; while in other embodiments, such as to reduce or limit extraneous noise or to help filter such noise, shorter ranges of signals also can be used, i.e. in some locations, the collections systems can be configured to receive signals transmitted within about 100 feet of the collection systems, and in still other embodiments or locations, signals transmitted within about 50 feet of the collection systems. Other, varying ranges also can be used.
In embodiments, the electronic signatures can include information related to the collected electronic signals of the transmitting electronic devices 202 or combinations of collected electronic signals of the transmitting electronic devices 202 that uniquely identify the targets. For example, and without limitation, a combination of one or more signals from a plurality of transmitting electronic devices 202 (e.g., a watch, cell phone/communication device, headphones, etc.) can include an electronic signature that uniquely identifies a person (e.g. the electronic signature may be received as or may include a MAC user ID 208 and/or a GATT profile 210); a combination of one or more signals from a plurality of transmitting vehicle components (e.g., a vehicle sensor, infotainment system, etc.) can include an electronic signature that uniquely identifies a vehicle 116 (e.g., including, for example, received values such as make, model, manufacturing serial number, and/or license plate number as shown at 212), or one or more signals from a transmitting electronic device 120, 122 can include an electronic signature that uniquely identifies that electronic device
In addition, the surveillance system 100 includes an intelligence system 102 that is in communication with the plurality of collection systems. The intelligence system 102 is configured to receive information (e.g., in one embodiment, time and location stamped data points 110) collected or captured by the collection systems and to identify and/or track targets or correlate a target with other targets or electronic devices based on this received information. The intelligence system can be in wireless communication with the collection systems, e.g., through a public or private network using Wi-Fi, cellular, etc . . . .
In addition, or in the alternative, the intelligence system and one or more of the collection systems can be connected through one or more wired connections. In this regard, when targets come within proximity of the collection systems, the collection systems will collect visual information and/or electronic signal information associated with the targets and transmit data points or packets of information, e.g., time and location stamped information, related to collected visual and/or electronic signal information to the intelligence system. The collection systems can be configured to transmit data points or packets substantially simultaneously or generally in real time when targets come within proximity to the collection systems. For example, the collection systems can send a data point including information corresponding to each electronic signal or visual identifier as it is captured or can send a data packet including information corresponding to multiple electronic signals or visual identifiers received. In addition, or in the alternative, the collection systems can transmit the data points or packets at specific time intervals, such as every few seconds, minutes, hours, etc. or at other times or intervals after the electronic signals or visual identifiers are captured, without departing from the scope of the present disclosure.
The correlation and search engine 104 further can be configured to filter or otherwise alter the received electronic signatures (or information related thereto) to reduce or diminish signal noise and facilitate identification or extraction of unique, identifying electronic signatures. For example, the correlation and search engine 104 can apply filtering (e.g., linear or non-linear filters, dynamic noise reduction, etc.) to collected electronic signals to diminish, reduce, or substantially eliminate stationary and variable noise and other values that cannot be usefully correlated with targets, allowing unique electronic signal values to be extracted or identified.
In addition, the correlation and search engine 104 is configured to catalogue the electronic signatures and/or visual identifiers in the intelligence database with specific identifying characteristics allowing these identified electronic signatures and/or visual identifiers to become unique, identifiable, and searchable. The identifying characteristics can include, but are not limited to, geographical coordinates, time stamps, source manufacturer, source type and unique ID, etc., The correlation and search engine 104 also can be configured to build catalogs or groupings of independent data points/data packets in the intelligence database that allow correlation analysis to show what otherwise anonymous or non-unique electronic signals and/or other visual identifiers (e.g., other license plates) consistently appear with the targets. The surveillance system 100 thus can identify, track, map, catalogue, etc., the presence and/or movements of the targets in real time as electronic signals emanating therefrom occur in proximity to the collection systems or based on image captures of visual identifiers. The surveillance system 100 further can generate alerts or notifications when certain targets are in proximity to the collection systems. Still further, the surveillance system 100 further allows for the searches or queries of the intelligence database 106, e.g., for investigating locations or movements of suspects or other persons of interest.
In embodiments, the correlation and search engine 104 can use algorithms, models, statistical models, machine learning algorithms/models, Big Data analysis or statistics, etc., to infer relationships between transmitting electronic devices and/or targets based on consistency or likelihood of correlation of the visual identifiers and/or electronic signals of the transmitting electronic devices 120, 122. For example, the correlation and search engine 104 can be configured to evaluate and combine singular collection events at the collection systems with other catalogued events in the intelligence database 106 to develop correlated information related to the intersection of multiple collected/captured electronic signals and/or visual identifiers that occurred at a specific time and geographical area or location. The correlation and search engine 104 can use the frequency and/or consistency of electronic signals and/or visual identifiers received at collection systems to determine the relative certainty of association of the transmitting electronic devices and/or targets to develop electronic signatures (correlated electronic devices) or correlated targets (e.g., correlated license plates) for the targets.
The correlation and search engine 104 can be programmed to determine a likelihood or probability that a specific electronic signal, a combination or set of electronic signals, and/or other target or targets are associated with a target, and if the determined likelihood or probability meets a prescribed/selected likelihood or probability threshold, the engine will identify or extract an electronic signal or combinations of electronic signals as an electronic signature or electronic signatures to be associated with that target. In one embodiment, the likelihood or probability threshold can be about 70% or more (e.g., above 75%, above 80%, above 85%, above 90%, above 95%, above 98%, etc.) that an electronic signal, combination/set of electronic signals, and/or other targets are associated with a particular target.
For example, the correlation and search engine 104 may correlate two or more license plates and one or more electronic devices based on multiple events that such a combination is received. Based on such a correlation, a prediction of when a particular vehicle may be present at a specific location may be determined by the correlation and search engine. Further, the two or more license plates may be from or may define a convoy (e.g., group of vehicles). In such an example, the electronic devices may be associated with the convoy.
In some embodiments, the correlation and search engine 104 can be configured to determine or identify a location at which a visual identifier and correlated electronic signature and/or other visual identifier are matched to enable tracking and/or verification of targets at such a location. In addition, or in the alternative, the correlation and search engine 104 can be configured to associate identifying electronic signatures and/or other visual identifiers with visual identifiers, such as a visual vehicle identifier, to allow independent tracking and location identification of targets based on the associated identifying electronic signatures and other visual identifiers. For example, once the engine has records correlating electronic signatures and/or other visual identifiers, e.g., a license plate likely to be located at or near a specific visual vehicle identifier, associated with the specific visual vehicle identifier, e.g., a specific license plate number, the correlation and search engine will be able to detect the likely presence of a vehicle and its associated license plate without visual information of that specific vehicle, e.g., a camera may or may not be used. Furthermore, the collection systems can be placed in locations or areas not associated with vehicular traffic, such that the intelligence system 100 will be able to identify, and catalogue known electronic signatures away from the vehicles they have typically been associated with.
In this regard, in embodiments, the collection systems can be used in conjunction with existing ALPRs in certain areas or locations, allowing the intelligence system to develop a subset of electronic signatures and/or other license plate reads associated with a license plate read at a moment in time and location. For example, one or more collection systems can be positioned near or in close proximity to an existing ALPR to allow for correlation or association of received electronic signals with license plate reads. In addition, or in the alternative, an existing ALPR can be modified or retrofitted to include components of the collection point systems to enable collection of electronic signals with license plate reads. Further, in some embodiments, collection systems with or near cameras or ALPRs can be used in connection with collection systems without cameras or ALPRs, as generally indicated in
Additionally, or alternatively, collection systems without cameras (or with cameras 112) can be positioned in areas or locations that cannot be accessed by a vehicle, such as on trains, near railways, around public buildings, etc., to enable collection of electronic signals from persons away from their vehicle, e.g., for cataloguing, tracking, mapping, etc . . . positions or movements thereof.
The intelligence system 102 generally includes one or more processors, controller's, CPUs, etc., and one or more memories, such as RAM, ROM, etc., in communication with the one or more processors. The engine can include computer programming instructions stored in the one or more memories that can be accessed and executed by the one or more processors to facilitate execution of the processes thereof, e.g., correlation of information, identification and tracking of the targets, searching of the intelligence database, etc . . . .
The correlation and search engine 104 can process the information from the received data points or data packages using a statistical analysis and/or machine learning algorithms to correlate the received read/target identification data with received alarms (e.g. hits indicating violations/criminal activities such as a stolen vehicle, warrants for the targets, etc . . . ). Operators then can search, query, and/or analyze the intelligence database, e.g., using a user interface 300 for visual navigation. as shown in
In some embodiments, the surveillance system 100 can be configured to capture an electronic signature and associated information from a target, and can associate such electronic signature, as well as associate other targets, and associated information with the target's identification, e.g., license plate number or other visual identifier, with the correlation and search engine, and then allow searches for or provide alerts or notifications on receipts of similar electronic signature information and/or visual identifier at one or more of the collection systems.
Each capture will include a time stamp and can also include a location/geographical information. The time stamp can include a date, day, and time that the capture was captured. Further, the capture can include location data, as well as related targets. As such, a user may visualize, via a user interface 300, different sets or subsets of captures or reads, as illustrated in
In an embodiment and as illustrated in
In a further embodiment, once a visualization is generated, a user may view subsets of the visualization. In such examples, the user may select a subset of the visualization. For example, the user may input a subset of the time. In another example, the user may click or select a portion of the visualization. For example, if a user would like to review data for Thursday over the selected time period, the user can click on “Thursday” and a new visualization may be generated illustrating the data captured or read over time for that particular day, as illustrated in
In a further embodiment, the user interface may connect to the intelligence database via an API, such as a REST or RESTful API, a JSON based API, a XML based API, or any other API suitable for communication between a database and user interface. As different inputs or selections are made at the user interface, different requests or commands may be sent to the database, such as a GET or SQL query. For example, as a date or other input is entered and the “analyze” button selected, the user interface may send the request or command to the intelligence database for a particular set of data. Once the data is received at the user interface, the user interface may be configured to present the data in a readable format, e.g., a chart as illustrated in
In another embodiment, the user interface may be configured to allow download or sharing (e.g., via electronic transmission of a link) of the data visualization. In another embodiment, the user interface may, as noted, include boxes or sections for inputting data. In another embodiment, the user interface may include a drag and drop style interface, configured to allow a user to select certain data based on boxes dragged to a certain section of the user interface. In yet another embodiment, the user interface may include dropdown selections. The user interface may include some combination of each type of data input.
In an embodiment, a user can select a set of alarm classes to build the analysis on. By adding the ability to perform the analysis on a VRM or set of VRMs, and domains/sets, the analysis becomes an even more powerful tool for locating suspect VRMs or groups/gangs. The resulting output is a complex set of 3 bar charts that show the distribution of alarms by Day-of-week, time-of-day, and Set/location. The tool will have the flexibility for the manager to select/freeze one dimension of the analysis and thereby alter the view of the statistics to match that new filter. For example, if the manager knows he will deploy resources on, for example, a Friday night, he might select the ‘Friday bar’ on the visualization, which would alter the time and location bar charts to show only the statistics for the alarms that occurred on Fridays.
By way of example and not limitation, a user can select a set of data using the following parameters:
A statistical analysis of this data can then be displayed in a linked set of chart/bar graphs presenting:
Clicking on any chart element changes the view of the other two charts to show the distribution of ANPR events in the data set of the clicked element by the other two factors. It is possible to further analyse the data by selecting two of the graph elements to see the effect on the third element. For example, clicking on a Camera chart element and then clicking on a Day of Week chart element will change the Hour of Day chart to show the hourly distribution of ANPR events in the data set for the selected camera on the selected day only.
This operation is enhanced so that historic data analysis preserves the hour of the day across the Daylight Savings Time boundary, i.e., if the user is looking at the 3-4 pm hour the system is automatically collating data from 3-4 pm in the daylight savings period, and 3-4 pm outside the daylight savings period. There also will be a facility to export the Tabular dataset which underpins each histogram chart as a CSV file to facilitate external analysis, the ability to export this data is controlled by the Bulk Export privilege. The CSV data can be used in an externally provided spreadsheet tool to create any required charts. The ability to use the Event Analysis feature will be controlled by the Advanced Data Mining Feature Privilege. In addition, Verification of Identity (VOI) list classes can be used to help further define and/or filter an event analysis operation.
At block 402, an intelligence system and/or surveillance system, via, for example, a user interface (e.g., user interface 300), may receive a selection of one or more of a day, time, or classification of a source of data or information. Further, at block 404, the intelligence system and/or surveillance system, via the user interface, may receive a selection of one or more factors. The one or more factors may include selection of types of events, specific targets, etc. At block 406, the intelligence system and/or surveillance system may generate a distribution of events based on the received day, time, and/or classification and the selection of one or more factors. At block 408, the intelligence system and/or surveillance system may generate a visualization of the events for display to the user interface.
At block 410, the intelligence system and/or surveillance system may determine whether the events exceed a selected frequency and/or exhibit a consistency. If the events do not exceed the selected frequency and/or do not exhibit the consistency, then the intelligence system and/or surveillance system may wait for new selections. If events do exceed the selected frequency and/or exhibit the consistency, then the intelligence system and/or surveillance system, at block 412, may predict a pattern of potential criminal and/or suspect conduct. Further, at block 414, the intelligence system and/or surveillance system may deploy investigative and/or enforcement resources based on the events exceeding a selected frequency and/or exhibiting a consistency.
Example case studies include the following:
The foregoing description generally illustrates and describes various embodiments of the present disclosure. It will, however, be understood by those skilled in the art that various changes and modifications can be made to the above-discussed construction of the present disclosure without departing from the spirit and scope of the disclosure as disclosed herein, and that it is intended that all matter contained in the above description or shown in the accompanying drawings shall be interpreted as being illustrative, and not to be taken in a limiting sense. Furthermore, the scope of the present disclosure shall be construed to cover various modifications, combinations, additions, alterations, etc., above and to the above-described embodiments, which shall be considered to be within the scope of the present disclosure. Accordingly, various features and characteristics of the present disclosure as discussed herein may be selectively interchanged and applied to other illustrated and non-illustrated embodiments of the disclosure, and numerous variations, modifications, and additions further can be made thereto without departing from the spirit and scope of the present invention as set forth in the appended claims.
The present Patent Application claims the benefit of United States Provisional Application No. 63/242,145, filed Sep. 9, 2021.
Number | Name | Date | Kind |
---|---|---|---|
3656111 | Royster, Sr. | Apr 1972 | A |
4843463 | Michetti | Jun 1989 | A |
6140941 | Dwyer et al. | Oct 2000 | A |
6340935 | Hall | Jan 2002 | B1 |
6538580 | Bostrom | Mar 2003 | B2 |
6650765 | Alves | Nov 2003 | B1 |
6754368 | Cohen | Jun 2004 | B1 |
6844841 | Masciulli | Jan 2005 | B1 |
6922138 | Melvin | Jul 2005 | B2 |
6959869 | Tsikos et al. | Nov 2005 | B2 |
6982654 | Rau et al. | Jan 2006 | B2 |
7046169 | Bucholz et al. | May 2006 | B2 |
7173526 | Monroe | Feb 2007 | B1 |
7227974 | Kamijo et al. | Jun 2007 | B2 |
7342511 | Schuette | Mar 2008 | B2 |
7344284 | Lynam et al. | Mar 2008 | B2 |
7382280 | Long | Jun 2008 | B2 |
7460028 | Garibotto et al. | Dec 2008 | B2 |
7465108 | Brown | Dec 2008 | B2 |
7482910 | Melvin | Jan 2009 | B2 |
7504965 | Windover et al. | Mar 2009 | B1 |
7535793 | Frantz | May 2009 | B2 |
7825829 | Madsen | Nov 2010 | B2 |
7868912 | Venetianer | Jan 2011 | B2 |
7889931 | Webb et al. | Feb 2011 | B2 |
7920626 | Fernandez | Apr 2011 | B2 |
8068154 | Jacob et al. | Nov 2011 | B2 |
8117090 | Romero | Feb 2012 | B2 |
8243140 | Howe | Aug 2012 | B1 |
8408772 | Li | Apr 2013 | B2 |
8447112 | Paul et al. | May 2013 | B2 |
8666801 | Cho | Mar 2014 | B2 |
8693733 | Harrison | Apr 2014 | B1 |
8704889 | Hofman | Apr 2014 | B2 |
8774465 | Christopulos et al. | Jul 2014 | B2 |
8781169 | Jackson et al. | Jul 2014 | B2 |
8817098 | Millar | Aug 2014 | B2 |
8953044 | Wu et al. | Feb 2015 | B2 |
8983137 | Hradetzky | Mar 2015 | B2 |
9058619 | Nandakumar et al. | Jun 2015 | B2 |
9141503 | Chen | Sep 2015 | B1 |
9158954 | Thuries et al. | Oct 2015 | B2 |
9171382 | Nerayoff et al. | Oct 2015 | B2 |
9245203 | Hofman et al. | Jan 2016 | B2 |
9280895 | Wang et al. | Mar 2016 | B2 |
9327685 | Wooten | May 2016 | B1 |
9342934 | Chen | May 2016 | B2 |
9426345 | Turner, Jr. et al. | Aug 2016 | B1 |
9436880 | Bos et al. | Sep 2016 | B2 |
9471838 | Miller et al. | Oct 2016 | B2 |
9477917 | Koch | Oct 2016 | B1 |
9491420 | Mimar | Nov 2016 | B2 |
9495869 | Bhogal et al. | Nov 2016 | B2 |
9497393 | Reed et al. | Nov 2016 | B2 |
9538060 | Cilia | Jan 2017 | B1 |
9542609 | Wu et al. | Jan 2017 | B2 |
9594158 | Abl | Mar 2017 | B2 |
9607510 | DeLorean | Mar 2017 | B1 |
9652666 | Nerayoff et al. | May 2017 | B2 |
9685084 | Scholl et al. | Jun 2017 | B2 |
9715703 | Robinson et al. | Jul 2017 | B2 |
9779309 | Fink et al. | Oct 2017 | B1 |
9796331 | Karner et al. | Oct 2017 | B2 |
9818157 | McCall et al. | Nov 2017 | B2 |
9846883 | Angell et al. | Dec 2017 | B2 |
9858493 | Tseng | Jan 2018 | B2 |
9911297 | Lundy | Mar 2018 | B2 |
9972230 | DeLorean | May 2018 | B1 |
9990376 | Voeller et al. | Jun 2018 | B2 |
9996861 | Ryan et al. | Jun 2018 | B2 |
10027361 | Griffin | Jul 2018 | B2 |
10032212 | Winkelman et al. | Jul 2018 | B2 |
10055906 | Fournier et al. | Aug 2018 | B1 |
10057713 | Rackley, III et al. | Aug 2018 | B1 |
10089869 | Makled et al. | Oct 2018 | B1 |
10121172 | Hudson | Nov 2018 | B2 |
10127437 | Bhanu et al. | Nov 2018 | B2 |
10140488 | Bondarenko et al. | Nov 2018 | B1 |
10152858 | Boykin | Dec 2018 | B2 |
10164974 | Spencer et al. | Dec 2018 | B2 |
10198954 | Holtzman et al. | Feb 2019 | B2 |
10217034 | Hoshino | Feb 2019 | B2 |
10223609 | Bhanu | Mar 2019 | B2 |
10373226 | Russell | Aug 2019 | B1 |
10417816 | Satzoda et al. | Sep 2019 | B2 |
10559200 | Langley et al. | Feb 2020 | B1 |
10691968 | Blais-Morin et al. | Jun 2020 | B2 |
10872524 | Kienitz et al. | Dec 2020 | B2 |
11030892 | Langley et al. | Jun 2021 | B1 |
11080533 | Camilus et al. | Aug 2021 | B2 |
11243959 | Gilboa | Feb 2022 | B1 |
11941716 | Child et al. | Mar 2024 | B2 |
20020089157 | Breed et al. | Jul 2002 | A1 |
20020186148 | Trajkovic et al. | Dec 2002 | A1 |
20030131011 | Haunschild et al. | Jul 2003 | A1 |
20040218052 | DiDomenico et al. | Nov 2004 | A1 |
20050012817 | Hampapur | Jan 2005 | A1 |
20050065711 | Dahlgren et al. | Mar 2005 | A1 |
20060007308 | Ide | Jan 2006 | A1 |
20060028556 | Bunn et al. | Feb 2006 | A1 |
20060030985 | Lawida et al. | Feb 2006 | A1 |
20060106659 | Santos et al. | May 2006 | A1 |
20060109083 | Rathus et al. | May 2006 | A1 |
20060180647 | Hansen | Aug 2006 | A1 |
20060269105 | Langlinais | Nov 2006 | A1 |
20060278705 | Hedley et al. | Dec 2006 | A1 |
20070008129 | Soliman | Jan 2007 | A1 |
20070009136 | Pawlenko | Jan 2007 | A1 |
20070069921 | Sefton | Mar 2007 | A1 |
20070112620 | Johnson | May 2007 | A1 |
20070229307 | Pawlenko et al. | Oct 2007 | A1 |
20070291118 | Shu et al. | Dec 2007 | A1 |
20080045245 | Billmaier et al. | Feb 2008 | A1 |
20080117296 | Egnal et al. | May 2008 | A1 |
20080285803 | Madsen | Nov 2008 | A1 |
20090088924 | Coffee et al. | Apr 2009 | A1 |
20090161913 | Son | Jun 2009 | A1 |
20100066537 | Weller et al. | Mar 2010 | A1 |
20100088123 | McCall et al. | Apr 2010 | A1 |
20100195871 | Simon | Aug 2010 | A1 |
20100245582 | Harel | Sep 2010 | A1 |
20110053559 | Klein | Mar 2011 | A1 |
20110134240 | Anderson et al. | Jun 2011 | A1 |
20110228085 | Hofman | Sep 2011 | A1 |
20110228253 | Batten et al. | Sep 2011 | A1 |
20110267222 | Craig | Nov 2011 | A1 |
20110313644 | Grassi | Dec 2011 | A1 |
20120106801 | Jackson | May 2012 | A1 |
20120140079 | Millar | Jun 2012 | A1 |
20120155712 | Paul et al. | Jun 2012 | A1 |
20130082828 | Furey et al. | Apr 2013 | A1 |
20130093886 | Rothschild | Apr 2013 | A1 |
20130129152 | Rodriguez Serrano et al. | May 2013 | A1 |
20130132166 | Wu et al. | May 2013 | A1 |
20130135468 | Kim | May 2013 | A1 |
20130282448 | Rydbeck et al. | Oct 2013 | A1 |
20130303189 | Bennett | Nov 2013 | A1 |
20130306728 | Thuries et al. | Nov 2013 | A1 |
20130308003 | Raghavan | Nov 2013 | A1 |
20130325629 | Harrison | Dec 2013 | A1 |
20140046800 | Chen | Feb 2014 | A1 |
20140066089 | Monks | Mar 2014 | A1 |
20140093125 | Hradetzky | Apr 2014 | A1 |
20140126779 | Duda | May 2014 | A1 |
20140139660 | Zhu | May 2014 | A1 |
20140159925 | Mimeault et al. | Jun 2014 | A1 |
20140160283 | Hofman et al. | Jun 2014 | A1 |
20140225719 | Kesavan et al. | Aug 2014 | A1 |
20140241585 | Zafiroglu et al. | Aug 2014 | A1 |
20140254866 | Jankowski et al. | Sep 2014 | A1 |
20140254877 | Jankowski et al. | Sep 2014 | A1 |
20140254878 | Jankowski et al. | Sep 2014 | A1 |
20140334684 | Strimling | Nov 2014 | A1 |
20150049914 | Alves | Feb 2015 | A1 |
20150054676 | Vinski et al. | Feb 2015 | A1 |
20150113012 | Silver | Apr 2015 | A1 |
20150119069 | Harris et al. | Apr 2015 | A1 |
20150123794 | Hamalainen et al. | May 2015 | A1 |
20150138001 | Davies | May 2015 | A1 |
20150149465 | Unser et al. | May 2015 | A1 |
20150154669 | Wu et al. | Jun 2015 | A1 |
20150205868 | Boncyk et al. | Jul 2015 | A1 |
20150294373 | Schuh | Oct 2015 | A1 |
20150310293 | Dehart | Oct 2015 | A1 |
20150317687 | Ramesh et al. | Nov 2015 | A1 |
20150348342 | Baxley et al. | Dec 2015 | A1 |
20150356498 | Casanova | Dec 2015 | A1 |
20150381948 | Renkis | Dec 2015 | A1 |
20160021344 | Renkis | Jan 2016 | A1 |
20160098864 | Nawrocki | Apr 2016 | A1 |
20160110759 | Polehn et al. | Apr 2016 | A1 |
20160112461 | Othmer | Apr 2016 | A1 |
20160124071 | Baxley et al. | May 2016 | A1 |
20160266579 | Chen | Sep 2016 | A1 |
20160343252 | Davies et al. | Nov 2016 | A1 |
20160344979 | Renkis | Nov 2016 | A1 |
20170018170 | Sayavong et al. | Jan 2017 | A1 |
20170032666 | Pretorius et al. | Feb 2017 | A1 |
20170048376 | Logan | Feb 2017 | A1 |
20170061506 | Dow et al. | Mar 2017 | A1 |
20170124395 | Cohen et al. | May 2017 | A1 |
20170126648 | Bond et al. | May 2017 | A1 |
20170131202 | McClintock et al. | May 2017 | A1 |
20170140237 | Voeller et al. | May 2017 | A1 |
20170186317 | Franklin et al. | Jun 2017 | A1 |
20170193810 | Cao et al. | Jul 2017 | A1 |
20170206559 | Bakshi et al. | Jul 2017 | A1 |
20170308964 | Morris et al. | Oct 2017 | A1 |
20170330455 | Kikuchi et al. | Nov 2017 | A1 |
20170339529 | Buttolo et al. | Nov 2017 | A1 |
20170357662 | Knany et al. | Dec 2017 | A1 |
20170357881 | Hoshino | Dec 2017 | A1 |
20170365105 | Rao et al. | Dec 2017 | A1 |
20170372143 | Barcus et al. | Dec 2017 | A1 |
20170374324 | Loftus | Dec 2017 | A1 |
20180032590 | Recavarren | Feb 2018 | A1 |
20180072177 | Tremblay | Mar 2018 | A1 |
20180096595 | Janzen et al. | Apr 2018 | A1 |
20180107528 | Vizer et al. | Apr 2018 | A1 |
20180107892 | Istenes | Apr 2018 | A1 |
20180137356 | Nerayoff et al. | May 2018 | A1 |
20180144382 | Makke et al. | May 2018 | A1 |
20180165935 | VanBlon et al. | Jun 2018 | A1 |
20180174457 | Taylor | Jun 2018 | A1 |
20180197416 | Hyde et al. | Jul 2018 | A1 |
20180211115 | Klein | Jul 2018 | A1 |
20180211116 | Modi et al. | Jul 2018 | A1 |
20180211117 | Ratti | Jul 2018 | A1 |
20180247519 | Bekanich | Aug 2018 | A1 |
20180253805 | Kelly et al. | Sep 2018 | A1 |
20180300578 | Wilbert et al. | Oct 2018 | A1 |
20180328757 | Chan et al. | Nov 2018 | A1 |
20180349699 | O'Connell et al. | Dec 2018 | A1 |
20180350230 | Kienitz et al. | Dec 2018 | A1 |
20180357247 | Siminoff et al. | Dec 2018 | A1 |
20190031146 | Etonye | Jan 2019 | A1 |
20190039463 | Moghe et al. | Feb 2019 | A1 |
20190043355 | Ferguson et al. | Feb 2019 | A1 |
20190050634 | Nerayoff et al. | Feb 2019 | A1 |
20190050652 | Baker et al. | Feb 2019 | A1 |
20190051142 | Wiles | Feb 2019 | A1 |
20190057601 | Watanabe et al. | Feb 2019 | A1 |
20190087864 | Hacker | Mar 2019 | A1 |
20190147260 | May | May 2019 | A1 |
20190147306 | Hoshino | May 2019 | A1 |
20190163982 | Block | May 2019 | A1 |
20190244498 | Dumas | Aug 2019 | A1 |
20190294890 | Kanga | Sep 2019 | A1 |
20190311327 | Habbaba et al. | Oct 2019 | A1 |
20190329421 | Deyle et al. | Oct 2019 | A1 |
20200097735 | Altuev | Mar 2020 | A1 |
20200104566 | Yamada et al. | Apr 2020 | A1 |
20200111350 | Julian et al. | Apr 2020 | A1 |
20200162701 | Nixon et al. | May 2020 | A1 |
20200211345 | Gordon-Carroll et al. | Jul 2020 | A1 |
20200265258 | Blais-Morin et al. | Aug 2020 | A1 |
20210026019 | Gahagan et al. | Jan 2021 | A1 |
20210044959 | Beaurepaire | Feb 2021 | A1 |
20210225161 | Horn | Jul 2021 | A1 |
20210326563 | Kossor | Oct 2021 | A1 |
20210398298 | Kisara | Dec 2021 | A1 |
20220141426 | Kim et al. | May 2022 | A1 |
20220148320 | Alakarhu | May 2022 | A1 |
20220188953 | Child et al. | Jun 2022 | A1 |
20220294946 | D'Annunzio | Sep 2022 | A1 |
20230010292 | Simon | Jan 2023 | A1 |
20230046309 | May | Feb 2023 | A1 |
20230073717 | Child | Mar 2023 | A1 |
20230073876 | Child | Mar 2023 | A1 |
20230162310 | Child et al. | May 2023 | A1 |
20230177954 | Neff et al. | Jun 2023 | A1 |
20230196780 | Nieuwkerk et al. | Jun 2023 | A1 |
20230237907 | Valcarcel et al. | Jul 2023 | A1 |
20230274647 | Nieuwkerk et al. | Aug 2023 | A1 |
20240021092 | Child | Jan 2024 | A1 |
20240185371 | Child et al. | Jun 2024 | A1 |
Number | Date | Country |
---|---|---|
2013201309 | Aug 2014 | AU |
2620630 | Sep 2014 | CA |
2916418 | Jun 2017 | CA |
3000127 | Jun 2018 | CA |
2924997 | Jun 2021 | CA |
2979406 | Feb 2024 | CA |
102708693 | Oct 2012 | CN |
203050715 | Jul 2013 | CN |
102005035242 | Feb 2007 | DE |
0784838 | Jun 2000 | EP |
1993279 | Nov 2008 | EP |
3726253 | Oct 2020 | EP |
3989194 | Apr 2022 | EP |
2975210 | Nov 2012 | FR |
2328731 | Mar 1999 | GB |
282920 | May 2017 | IN |
140858 | Apr 2008 | SG |
201935419 | Sep 2019 | TW |
WO2004027730 | Apr 2004 | WO |
WO2007032025 | Mar 2007 | WO |
WO2008088880 | Jul 2008 | WO |
WO2011023024 | Mar 2011 | WO |
WO2012038964 | Mar 2012 | WO |
WO2014152923 | Sep 2014 | WO |
WO2015193903 | Dec 2015 | WO |
WO2017180900 | Oct 2017 | WO |
WO2018009567 | Jan 2018 | WO |
WO2019043444 | Mar 2019 | WO |
WO2019043446 | Mar 2019 | WO |
WO2019162970 | Aug 2019 | WO |
WO2020145882 | Jul 2020 | WO |
Entry |
---|
Howington, Larry C.; Automated License Plate Reading; Development and use of Perceptics multi-purpose License Plate Reader Technology; Sep. 1989. |
Waters, Rebecca; Automated License Plate Recognition: An Insider's Look; P&SN Police and Security News 2022; Posted by Publisher—Sep. 15, 2017. |
Ceng, M. Smalley et al.; Videqual—A tool for quantitive meansurement of fixed CCTV Surveillance Systems; Presented and Published at the International Carnahan Conference on Security Technology 1996; 0-7803-3537-6-9/06; 1996 IEEE. |
Kanayama, Kenji et al.; Development of Vehicle-License Number Recognition System Using Real-Time Image Processing and Its Application to Travel-Time Measurement; CH2944-7/91/0000/0798; 1991 IEEE. |
Kwasnicka, Halina and Wawrzyniak, Bartosz; License Plate Localization and Recognition in Camera Pictures; AI-METH 2002—Artificial Intelligence Methods; Gliwice, Poland, Nov. 13-15, 2002. |
Appian Technology PLC; Talon—Description and Technical Specification—Automatic Number Plate Recognition System; Copyright 2003. |
Hoffman, Alwyn Jakobus and Albertus J. Pretorius: “SmartRoad; A new approach to law enforcement in dense traffic environments”; 2015 IEEE 18th International Conference on Intelligent Transportation Systems. IEEE, 2015; Year 2015. |
Yaacoub, et al.; Security analysis of drones systems: Attacks, limitations, and recommendations; Internet of Things 11 (2020) 100218; Elsevier; www.elsevier.com/locate/iot; Retrieved on Apr. 17, 2023 (Apr. 17, 2023) from https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7206421/. |
Makarov, A., et al.; Authenticating Vehicles and Drivers in Motion Based on Computer Vision and RFID Tags; 2012 IEEE 10th Jubilee International Symposium on Intelligent Systems and Informatics; IEEE 2012; Year 2012. |
Notification of Transmittal and the International search Report and the Written Opinion of the International Searching Authority for PCT/US22/42916 dated Feb. 9, 2023. |
Valera, M. and Velastin, S.A., “Intelligent distributed surveillance systems: a review”, Source: vol. 152, Issue 2, Apr. 2005, p. 192-204, DOI: 10. 1049/ip-vis:20041147, Print ISSN 1350-245X, Online ISSN 1359-7108; 13 pages; 2005. |
Hampapur, A.; “smart video surveillance: exploring the concept of multiscale spatiotemporal tracking”, Published in: IEEE Signal Processing Magazine (vol. 22, Issue: 2, Mar. 2005), pp. 38-51; 2005. |
Girgensohn, Andreas, “DOTS: Support for Effective Video Surveillance”, MM '07: Proceedings of the 15th ACM International conference on Multimedia Sep. 2007; pp. 423-432; 2007. |
Hu, Weiming, et al.; “A Survey on Visual Surveillance of Object Motion and Behaviors”; IEEE Transactions on Systems, Man, and Cybernetics, Part C: Applications and Reviews, vol. 34, No. 3, Aug. 2004; pp. 334-352; Year 2004. |
Panganiban, Edward B. et al.; “RFID-Based Vehicle Monitoring System”; 2017 IEEE 9th International Conference on Humanoid, Nanotechnology, Information Technology, Communication and Control, Environmental Management (HNICEM); IEEE, Dec. 1-3, 2017; pp. 1-6; Year 2017. |
Number | Date | Country | |
---|---|---|---|
20230070108 A1 | Mar 2023 | US |
Number | Date | Country | |
---|---|---|---|
63242145 | Sep 2021 | US |