Not Applicable
The present disclosure relates generally to human-computer interfaces and machine learning, and more particularly to configurable monitoring and actioning with distributed programmable pattern recognition edge devices managed by a client software application.
Current home or building security systems are built upon fixed function devices that monitor a specific pattern of inputs from various environmental sensors. Such sensors may be microphones that capture sound from the monitored environment and a controller may evaluate the audio as a specific type of event. For example, glass breakage sensors may monitor for sounds that are characteristic thereof. There may also be electro-mechanical sensors such as those installed on doors, windows, and the like where electrical continuity is broken when opened and triggers an alarm condition. Optical sensors may monitor movement within the environment.
Regardless of the specific inputs or sensor types, when the pattern is detected, an alarm is enabled, or a signal is sent to an electronic device such as a phone or other mobile communication device to signal the detection. Such solutions are rigid and fixed to specific functions. Current systems are capable of providing only a fixed notification of a pre-determined event with no possibility of any follow up action. Furthermore, the user's privacy may be compromised to the extent cloud processing is needed. The ability to take effective action following the detection of an event may also be limited. As time goes on, the functionality of the sensing devices may become less useful or obsolete, and the user is pushed to acquire new ones for newly needed functions. For example, a baby crying detection device may become unnecessary as the child ages.
Accordingly, there is a need in the art for the configurable monitoring and actioning using distributed programmable pattern recognition edge devices managed by a client software application on user devices.
The embodiments of the present disclosure are directed to various methods and systems for a configurable monitoring and actioning system utilizing distributed programmable pattern recognition edge devices managed by a software application. The features of the method and system contemplates the pattern recognition being configurable by the user for time-of-day, environmental conditions, situation of the house, family needs and so on, and provide the user with the ability to take definitive follow up action to resolve the triggering event. Additionally, there are various privacy advantages, in that each edge device performs pattern recognition locally without sharing the audio or the image with an application. Once a pattern is recognized, an interrupt and a notification is sent to the application, which can then take a predetermined action. The application may contain a dashboard of all edge devices and reported detection history, but once an edge device is reprogrammed to a different pattern recognition, the history may be deleted with the start of a new log.
According to one embodiment of the present disclosure, there may be a configurable monitoring and actioning system. It may include one or more programmable edge devices each including a machine learning pattern recognizer, a sensor providing sensor input data to the pattern recognizer, and a memory storing pre-trained machine learning weight values for the pattern recognizer. The machine learning pattern recognizer may generate event detections based upon evaluations of the sensor input data from the sensor against the pre-trained machine learning weight values. The system may also include an application installable on a user device. The application may be in communication with each of the one or more programmable edge devices and execute predetermined actions based upon the event detection evaluations from the machine learning pattern recognizer.
Another embodiment of the present disclosure is directed to a method of monitoring and operating one or more programmable edge devices from a user device. The method may include establishing a communication link to the one or more programmable edge devices. There may also be a step of receiving event detections from one or more programmable edge devices. The event detections may be generated by a machine learning pattern recognizer on an originating one of the one or more programmable edge devices based on sensor input data captured thereby and evaluated against pre-trained machine learning weight values stored thereon. There may also be a step of correlating the event detections to one or more actions, as well as a step of executing the one or more actions on the user device.
An embodiment of the disclosure may also be a non-transitory computer readable medium that includes instructions executable by a data processing device to perform the method of monitoring and operating one or more programmable edge devices from a user device. The present disclosure will be best understood accompanied by reference to the following detailed description when read in conjunction with the drawings.
These and other features and advantages of the various embodiments disclosed herein will be better understood with respect to the following description and drawings, in which like numbers refer to like parts throughout, and in which:
The detailed description set forth below in connection with the appended drawings is intended as a description of the several presently contemplated embodiments of a configurable monitoring and actioning system. It is not intended to represent the only form in which such embodiments may be developed or utilized, and the description sets forth the functions and features in connection with the illustrated embodiments. It is to be understood, however, that the same or equivalent functions may be accomplished by different embodiments that are also intended to be encompassed within the scope of the present disclosure. It is further understood that the use of relational terms such as first and second and the like are used solely to distinguish one from another entity without necessarily requiring or implying any actual such relationship or order between such entities.
With reference to the block diagram of
Referring to
The edge device 10 includes a main processor 18 that executes pre-programmed software instructions that correspond to various functional features of the edge device 10. These software instructions, as well as other data that may be referenced or otherwise utilized during the execution of such software instructions, may be stored in a memory 20. As referenced herein, the memory 20 is understood to encompass random access memory as well as more permanent forms of memory.
To the extent that the edge device 10 is a smart speaker, it is understood to incorporate a loudspeaker/audio output transducer 22 that outputs sound from corresponding electrical signals applied thereto. Furthermore, in order to accept audio input, the edge device 10 includes a microphone/audio input transducer 24. The microphone 24 is understood to capture sound waves and transduces the same to an electrical signal. According to various embodiments of the present disclosure, the edge device 10 may have a single microphone. However, it will be recognized by those having ordinary skill in the art that there may be alternative configurations in which the edge device 10 includes two or more microphones.
Both the loudspeaker 22 and the microphone 24 may be connected to an audio interface 26, which is understood to include at least an analog-to-digital converter (ADC) and a digital-to-analog converter (DAC). The ADC is used to convert the electrical signal transduced from the input audio waves to discrete-time sampling values corresponding to instantaneous voltages of the electrical signal. This digital data stream may be processed by the main processor, or a dedicated digital audio processor. The DAC, on the other hand, converts the digital stream corresponding to the output audio to an analog electrical signal, which in turn is applied to the loudspeaker 22 to be transduced to sound waves. There may be additional amplifiers and other electrical circuits within the audio interface 26, but for the sake of brevity, the details thereof are omitted. Furthermore, although the example edge device 10 shows a unitary audio interface 26, the grouping of the ADC and the DAC and other electrical circuits is by way of example and convenience only, and not of limitation.
In between the audio interface 26 and the main processor 18, there may be a general input/output interface 28 that manages the lower-level functionality audio interface 26 without burdening the main processor 18 with such details. Although there may be some variations in the way the audio data streams to and from the audio interface 26 are handled thereby, the general input/output interface 28 abstracts any such variations. Depending on the implementation of the main processor 18, there may or may not be an intermediary input/output interface 28.
According to some embodiments, the edge device 10 may also incorporate visual input and output peripheral components. Specifically, there may be a display 30 that outputs graphics corresponding to electrical signals the data representative thereof. The display 30 may be a matrix of light emitting elements arranged in rows and columns, with the elements thereof varying in size and technologies, such as liquid crystal displays (LCD), light-emitting diode (LED) displays and so on. It will also be appreciated that the display 30 may include simpler output devices such as segment displays as well as individual LED indicators and the like. The specific type of display that is incorporated into the edge device 10 is driven by the information presentation needs thereof.
The display 30 receives the electrical signals to activate the display elements from a visual interface 32. In some implementations, the visual interface 32 is a graphics card that has a separate graphics processor and memory to offload the graphics processing tasks from the main processor 18. Like the audio interface 26 discussed above, the visual interface 32 may be connected to the general input/output interface 28 to abstract out the functional details of operating the display and the visual interface 32.
The edge device 10 may further include an imager 34 that captures light from the environment and converts the same to electrical signals representative of the scene. A continuous stream or sequence of images may be captured by the imager 34, or a single image may be captured of a time instant in response to the triggering of a shutter. A variety of sensor technologies are known in the art, as are lenses, apertures, shutters, and other optical components that focus the light onto the sensor element for capture. Accordingly, such details of the imager 34 are omitted. The image data output by the imager 34 may be passed to the visual interface 32, and the commands to activate the capture function may be issued through the same. However, this is by way of example only, and some edge devices 10 may utilize a dedicated imager interface separate from that which controls the display 30. The imager 34 and the display 30 are shown connected to a unitary visual interface 32 only for the sake of convenience as representing functional corollaries of the other (e.g., image input vs. image output).
In addition to the foregoing peripheral devices, the edge device 10 may also include more basic input devices 36 such as buttons, keys, and switches with which the user may interact to command the edge device 10. These components may be connected directly to the general input/output interface 28.
The edge device 10 may also include a network interface 38, which serves as a connection point to a data communications network. This data communications network may be a local area network, the Internet, or any other network that enables a communications link between the edge device 10 and a remote note. In this regard, the network interface 38 is understood to encompass the physical, data link, and other network interconnect layers.
In order to communicate with more proximal devices within the same general physical space as the edge device 10, there may be a local communication interface 40. According to various embodiments, the local communication interface 40 may be a wireless modality such as infrared, Bluetooth, Bluetooth Low Energy, RFID, and so on. Alternatively, or additionally, the local communication interface 40 may be a wired modality such as Universal Serial Bus (USB) connections, including different standard generations and physical interconnects thereof (e.g., USB-A, micro-USB, mini-USB, USB-C, etc.). The local communication interface 40 is likewise understood to encompass the physical, data link, and other network interconnect layers, but the details thereof are known in the art and therefore omitted from the present disclosure. In various embodiments, a Bluetooth connection may be established between a smartphone and the edge device 10 to implement certain features of the present disclosure.
As the edge device 10 is electronic, electrical power must be provided thereto in order to enable the entire range of its functionality. In this regard, the edge device 10 includes a power module 42, which is understood to encompass the physical interfaces to line power, an onboard battery, charging circuits for the battery, AC/DC converters, regulator circuits, and the like. Those having ordinary skill in the art will recognize that implementations of the power module 42 may span a wide range of configurations, and the details thereof will be omitted for the sake of brevity.
The main processor 18 is understood to control, receive inputs from, and/or generate outputs to the various peripheral devices as described above. The grouping and segregation of the peripheral interfaces to the main processor 18 are presented by way of example only, as one or more of these components may be integrated into a unitary integrated circuit. Furthermore, there may be other dedicated data processing elements that are optimized for machine learning/artificial intelligence applications. One such integrated circuit is the AONDevices high-performance, ultra-low power edge AI device, AON1100 pattern recognition chip/integrated circuit. However, it will be appreciated by those having ordinary skill in the art that the embodiments of the present disclosure may be implemented with any other data processing device or integrated circuit utilized in the edge device 10. Although a basic enumeration of peripheral devices such as the loudspeaker 22, the microphone 24, the display 30, the imager 34, and the input devices 36 has been presented above, the edge device 10 need not be limited thereto. In some cases, one or more of these exemplary peripheral devices may not be present, while in other cases, there may be other, additional peripheral devices.
Returning to
In a conventional smartphone device, the user primarily interacts with a graphical user interface that is generated on the display and includes various user interface elements that can be activated based on haptic inputs received on the touch screen at positions corresponding to the underlying displayed interface element. Those having ordinary skill in the art will recognize other possible input/output devices that could be integrated into the user device 14, and the purposes such devices would serve. Other smartphone devices may include keyboards (not shown) and other mechanical input devices, and the presently disclosed interaction methods detailed more fully below are understood to be applicable to such alternative input modalities.
With reference to
Based upon the pre-trained weight values 50, the machine learning pattern recognizer 48 evaluates the inputs 46 to generate an event detection 52 if the input 46 corresponds thereto. The event detection 52 is provided to the user device 14, and specifically an application programming interface (API) 54 to the edge device 10 installed thereon. The application 12 may utilize the API 54 to retrieve the event detection and generate an alert on the user device 14 representative of the event detection 52. By way of example, the edge device 10 may be programmed to alert on breaking glass. If the audio data captured by the microphone 24/sensor 44 is evaluated to be broken glass by the machine learning pattern recognizer 48 based upon the pre-trained weight values 50 for breaking glass sound, then the event detection indicating broken glass as detected by the edge device 10 is transmitted to the API 54. The application 12 may, in turn, generate an alert that the sound of breaking glass was detected in the space being monitored by the edge device 10.
In order to improve upon the pattern recognition function, a small excerpt of the captured input 46 may be transmitted to the application 12 for targeted performance enhancement.
The configurable monitoring and actioning system 1 extends this functionality of a single edge device 10 to multiple instances, and each one may be configured to detect different events such as baby crying sounds, television sounds, presence of a human being, coughing sounds, movement of furniture, and so on. Referring back to
The assignment of different patterns to specific rooms 16 may be possible through a user interface of the application 12.
With the assignment of a different pattern to the edge device 10 via the user interface 56, the corresponding pre-trained weight value 66 may be transmitted to the designated edge device 10, where it is stored as the pre-trained weight value 50 in the memory 20. Henceforth, the machine learning pattern recognizer 48 generates the event detection 52 when the input 46 is evaluated to be the selected event. The updates to the pre-trained weight values 50 may be performed wirelessly over the air.
The application 12 is understood to include a scheduler that reprograms the system 1 based upon various secondary conditions such as time of day, environmental conditions, situation of the building/home/facility, as well as personal needs. The block diagram of
In addition to the existing pattern selected from the local library on the user device 14, further patterns and associated weight patterns may retrieved from a remote source. Updated pre-trained weight values 72 may be retrieved by the application 12 via the API 54 from a machine learning training source 74 such as a vendor of the edge device 10. Referring again to
The flowchart of
The flowchart of
The flowchart of
An example of this embodiment of the interactive monitoring and actioning system 1 is where the user has selected dog barking or whimpering monitoring for a certain section of the user's house from the library of pattern recognition. Upon detecting a dog barking or whimpering in the designated section of the house, the system 1 notifies the application 12, and the application 12 will then proceed with a sequence of pre-determined checks of a scenario for dog barking or whimpering. These pre-determined checks may include first checking the temperature and if the temperature is above or below set thresholds, the user may be alerted of this reading. The edge device 10 may then be switched to pre-determined voice command weights so that the temperature can be adjusted by voice command. If the temperature is within set limits, the system 1 may switch to doorbell ringing weights. If a doorbell event is detected and a notification is sent to the application 12, a notification to the user of the event may be generated and switched to image sensing weight values. The application 12 may then send the captured image to the user and switch to pre-determined voice command weights so that the user can instruct the person ringing the doorbell.
The particulars shown herein are by way of example and for purposes of illustrative discussion of the embodiments of a configurable monitoring and actioning system and are presented in the cause of providing what is believed to be the most useful and readily understood description of the principles and conceptual aspects. In this regard, no attempt is made to show details with more particularity than is necessary, the description taken with the drawings making apparent to those skilled in the art how the several forms of the present disclosure may be embodied in practice.
This application relates to and claims the benefit of U.S. Provisional Application No. 63/359,061 filed Jul. 7, 2022, and entitled “METHOD FOR A CONFIGURABLE MONITORING AND ACTION TAKING SYSTEM UTILIZING DISTRIBUTED PROGRAMMABLE PATTERN RECOGNITION DEVICES AT THE EDGE MANAGED BY A SOFTWARE APPLICATION,” the entire disclosure of which is wholly incorporated by reference herein.
Number | Date | Country | |
---|---|---|---|
63359061 | Jul 2022 | US |