Always-on detection systems

Information

  • Patent Grant
  • 10848886
  • Patent Number
    10,848,886
  • Date Filed
    Thursday, September 5, 2019
    4 years ago
  • Date Issued
    Tuesday, November 24, 2020
    3 years ago
Abstract
The present disclosure relates to a system for providing an acknowledgement output at an output transducer of a device in response to detection of a trigger signal. The system includes an input transducer, a signal processor, a main processor separate from the signal processor and an output driver. The signal processor monitors an input signal received at the signal processor from the input transducer and outputs an acknowledgement signal to the output driver if the received input signal corresponds to the trigger signal. The signal processor also outputs an interrupt signal to the main processor to cause the main processor to enter an active state. In this way an acknowledgement output can rapidly be provided to a user of the system, as there is no need to wait until the main processor has entered its active state and processed the trigger signal.
Description
FIELD OF THE INVENTION

The present disclosure relates to the field of always-on detection systems. In particular, the present disclosure relates to a system and method for providing an acknowledgement output in response to detection of a trigger signal in an always-on detection system.


BACKGROUND

Modern devices such as smartphones, tablet computers, smart watches, smart speakers, smart televisions and the like are increasingly being provided with voice recognition capabilities, permitting a user to issue spoken commands to the device to command it to perform particular actions. For example, a user may issue a spoken command to cause the device to commence playback of a chosen audio track, to activate, deactivate or adjust the operation of an appliance such as a lamp or television that is connected to the device, or may issue a spoken command to cause the device to retrieve information such as news, traffic or weather information from the Internet and provide a spoken summary of such information.


Typically such devices are provided with an always-on audio detection system, which is operative continually to monitor an audio input received at an audio input transducer such as a microphone of the device. The always-on audio detection system is configured to distinguish between normal conversation that is not directed at the device and spoken commands that are directed at the device. Audio input recognised as normal conversation is ignored by the device, whereas audio input that is recognised as a spoken command that is directed at the device causes the device to perform an appropriate action in response to the command.


In order to facilitate the distinction between normal conversation and spoken commands that are directed at the device, such devices typically require an audio trigger, such as a specific spoken word or phrase uttered by the user or a sound such as a handclap or whistle, to be provided by the user prior to the user issuing a spoken command. For example, if the user were simply to ask the question “What's the weather forecast for tomorrow?”, the device would interpret this as normal conversation rather than a command directed at the device. However, if the user were to speak a trigger word before asking the question, e.g. uttering the phrase “Device! What's the weather forecast for tomorrow?”, the device would recognise the audio trigger “Device!” and would then respond to the question “What's the weather forecast for tomorrow?” by retrieving relevant weather information from the Internet and providing a spoken summary of that information.


The always-on audio detection system is typically provided as a part of a signal processor that is separate from a main processor or application processor of the device. This arrangement enables the power consumption of the device to be reduced when not in active use, as the main or application processor can be placed in an inactive or sleep state when the device is not in active use, whilst the signal processor incorporating the always-on audio detection system, which has a lower power consumption than the main or application processor, remains on, actively monitoring the input audio signal. When the always-on audio detection system detects the audio trigger it can send an interrupt to the main or application processor to switch the main or application processor into its active or awake state in order to process the trigger signal and respond to a subsequently detected user command.


One problem that can arise in such arrangements is that once the always-on audio detection system has detected the audio trigger, it must send the audio trigger and any subsequent command to the main or application processor, which transmits the audio trigger and command to a remote server via the Internet for verification of the audio trigger, i.e. confirmation that the audio trigger is a valid audio trigger. Only after this confirmation has been received by the device does the device acknowledge to the user that the command has been accepted, typically by outputting a specific audio tone. This causes a delay between the user issuing a command and receiving confirmation that the command has been accepted. The delay is variable but can be as much as 2-6 seconds. As will be appreciated, in the absence of a rapid confirmation that a command has been accepted, a user may re-issue the command, perhaps multiple times, until confirmation is provided that the command has been accepted. This can lead to multiple instances of the command being accepted and responded to, which may cause confusion and frustration to the user.


SUMMARY

According to a first aspect, the invention provides a system for providing an acknowledgement output at an output transducer of a device in response to detection of a trigger signal, the system comprising: an input transducer; a signal processor coupled to the input transducer; a main processor separate from the signal processor; and an output driver, wherein the signal processor is configured to: monitor an input signal received at the signal processor from the input transducer; and if the received input signal corresponds to the trigger signal: output an acknowledgement signal to the output driver to cause the output driver to drive the output transducer to output the acknowledgement output; and output an interrupt signal to the main processor to cause the main processor to enter an active state.


In the system of the first aspect an acknowledgement that a trigger signal issued by the user has been received is output by the system without any need to wait for the main processor to enter its active state to process the trigger signal, thereby rapidly informing the user that the trigger input has been received.


The input transducer may be an audio input transducer, and the signal processor may be configured to monitor an audio input signal received at the signal processor from the audio input transducer.


The signal processor may comprise an always-on audio detection system configured to monitor the audio input signal received at the signal processor from the audio input device.


The system may further comprise a memory, and the memory may be configured to store one or more acknowledgement signals corresponding to the acknowledgement output.


The one or more acknowledgement signals corresponding to the acknowledgement output may comprise one or more of: an audio acknowledgement signal or waveform; a haptic acknowledgement signal or waveform; and a visual acknowledgement signal.


The output driver may comprise a haptic output driver and the acknowledgement output may comprise a haptic output, and the signal processor may be configured to output a haptic acknowledgement signal to the haptic output driver.


The output driver may comprise an audio output driver and the acknowledgement output may comprise an audio output, and the signal processor may be configured to output an audio acknowledgement signal to the audio output driver.


The output driver may comprise a visual output driver and the acknowledgement output may comprise a visual output, and the signal processor may be configured to output a visual acknowledgement signal to the visual output driver.


The memory may be on-board memory of the signal processor.


The main processor may be configured to enter an inactive state when not in use.


The trigger signal may be a voice command.


The system according may further comprise a buffer, and the buffer may be configured to store audio input received by the audio input transducer after the trigger signal has been received, and the system may be configured to transmit the contents of the buffer to the main processor after the main processor has entered its active state.


A further aspect of the invention provides a method for providing an acknowledgement output at an output transducer of a device in response to detection of a trigger signal, the method comprising: monitoring, at a signal processor, an input signal received at the signal processor from an input transducer; and if the received input signal corresponds to the trigger signal: outputting, from the signal processor, an acknowledgement signal to an output driver of the device to cause the output driver to drive the output transducer to output the acknowledgement output; and outputting, from the signal processor, an interrupt signal, to cause a main processor of the device to enter an active state.


The input transducer may be an audio input transducer and the method may further comprise monitoring, at the signal processor, an audio input signal received at the signal processor from the audio input transducer.


The signal processor may comprise an always-on audio detection system configured to monitor the audio input signal received at the signal processor from the audio input device.


The method may further comprise storing, in a memory, one or more acknowledgement signals corresponding to the acknowledgement output.


The one or more acknowledgement signals corresponding to the acknowledgement output may comprise one or more of: an audio acknowledgement signal or waveform; a haptic acknowledgement signal or waveform; and a visual acknowledgement signal.


The output driver may comprise a haptic output driver and the acknowledgement output may comprise a haptic output, and the method may comprise outputting, by the signal processor, a haptic acknowledgement signal to the haptic output driver.


The output driver may comprise an audio output driver and the acknowledgement output may comprise an audio output, and the method may comprise outputting, by the signal processor, an audio acknowledgement signal to the audio output driver.


The output driver may comprise a visual output driver and the acknowledgement output may comprise a visual output, the method comprise outputting, by the signal processor, a visual acknowledgement signal to the visual output driver.


The memory may be on-board memory of the signal processor.


The method may comprise causing the main processor to enter an inactive state when not in use.


The trigger signal may be a voice command.


The method may further comprise storing, in a buffer, audio input received by the audio input transducer after the trigger signal has been received, and transmitting the contents of the buffer to the main processor after the main processor has entered its active state.


A further aspect of the invention provides a signal processor for providing an acknowledgement output at an output transducer of a device in response to detection of a trigger signal, wherein the signal processor is configured to: monitor an input signal received at the signal processor from an input transducer; and if the received input signal corresponds to the trigger signal: output an acknowledgement signal to an output driver of the device to cause the output driver to drive the output transducer to output the acknowledgement output; and output an interrupt signal to cause a main processor of the device to enter an active state.


The input transducer may be an audio input transducer and the signal processor may be configured to monitor an audio input signal received at the signal processor from the audio input transducer.


The signal processor may comprise an always-on audio detection system configured to monitor the audio input signal received at the signal processor from the audio input device.


The output driver may comprise a haptic output driver and the acknowledgement output may comprise a haptic output, and the signal processor may be configured to output a haptic acknowledgement signal to the haptic output driver.


The output driver may comprise an audio output driver and the acknowledgement output may comprise an audio output, and the signal processor may be configured to output an audio acknowledgement signal to the audio output driver.


The output driver may comprise a visual output driver and the acknowledgement output may comprise a visual output, and the signal processor may be configured to output a visual acknowledgement signal to the visual output driver.


The signal processor may comprise a memory.


The memory may be configured to store one or more acknowledgement signals corresponding to the acknowledgement output.


The one or more acknowledgement signals corresponding to the acknowledgement output may comprise one or more of: an audio acknowledgement signal or waveform; a haptic acknowledgement signal or waveform; and a visual acknowledgement signal.


The main processor of the device may be configured to enter an inactive state when not in use.


The trigger signal may be a voice command.


The signal processor may be configured to receive the contents of a buffer after the main processor has entered its active state, the buffer being configured to store audio input received by the audio input transducer after the trigger signal has been received.





BRIEF DESCRIPTION OF THE DRAWINGS

Embodiments of the invention will now be described, strictly by way of example only, with reference to the accompanying drawings, of which:



FIG. 1 is a schematic diagram showing part of a device having an always-on audio detection system;



FIG. 2 is a flow chart illustrating steps performed by a signal processor to provide an acknowledgement output on detection of an audio trigger signal; and



FIG. 3 is a schematic diagram showing part of a device having an always-on detection system for detecting a trigger signal.





DETAILED DESCRIPTION


FIG. 1 is a schematic diagram showing part of a device 100 having a signal processor 102 including an always-on audio detection system 104. The device 100 may be, for example, a mobile telephone, tablet or laptop computer, smart watch, virtual reality (VR) or augmented reality (AR) device, smart speaker or smart television. For the sake of clarity elements of the device 100 which are not relevant to the present disclosure are not shown in FIG. 1, but those skilled in the art will appreciate that the device 100 may include further elements and components in addition to those shown in FIG. 1.


The always-on audio detection system 104 is coupled to an audio input transducer arrangement 106, which in the example shown in FIG. 1 comprises a pair of microphones, such that audio from the environment surrounding the device 100 can be detected in a continuous manner.


The signal processor 102 also interfaces with a memory 108, which in the example shown in FIG. 1 is separate from the signal processor 102, but which may also form part of the signal processor 102. For example, the memory 108 may be on-board memory of the signal processor 102. The memory 108 stores haptic, audio and visual output signals that are output by the signal processor 102 to output drivers of the device, which in turn drive haptic, audio and visual output transducers of the device 100 and external accessory devices, as will be explained in further detail below.


The signal processor 102 has a first output which interfaces with a haptic output amplifier 110. The haptic output amplifier 110 is configured to drive a haptic output transducer 112, which may be, for example, a linear resonant actuator or the like.


The signal processor 102 has a second output which interfaces with an audio output amplifier 114. The audio output amplifier 114 is configured to drive an audio output transducer 116, which may be, for example, one or more loudspeakers.


The signal processor 102 may have a third output which interfaces with a headphone amplifier 118. The headphone amplifier 118 is configured to drive one or more external wired accessory devices such as headphones, earphones, earbuds or the like, which may be connected to the headphone amplifier 118 by means of a jack socket 120 which is configured to receive a corresponding plug of the external wired accessory device.


The signal processor 102 may have a fourth output which interfaces with a wireless communications subsystem 122, which is configured to generate wireless signals complying with one or more wireless communications protocols such as Bluetooth®, WiFi® or the like, for transmission to an external wireless accessory device such as, for example, a wireless headset, via an antenna 124.


The signal processor 102 may have a fifth output which interfaces with a display subsystem 126, which is configured to drive a display 128 of the device 100, which may be, for example, a touch screen display.


The signal processor 102 may have a sixth output which interfaces with a lamp subsystem 130, which is configured to drive a lamp 132 of the device 100.


The device 100 also includes a main processor 134, which may be, for example, an application processor. The main processor 134 is separate from the signal processor 102 and interfaces with the signal processor 102, to provide signals to be processed by the signal processor 102 for subsequent output by one or more of the output drivers (haptic output amplifier 110, audio output amplifier 114, headphone amplifier 118, wireless subsystem 122, display subsystem 126, or lamp subsystem 130).


To increase the power efficiency of the device 100, the main processor 134 is configured to enter an inactive or sleep state when it is not required. The signal processor 102 and the always-on audio detection system 104 remain active at all times, in order to detect audio triggers indicating that the user requires the device 100 to perform an action in response to a user command.


The processor 134 is provided with an interrupt line 128, which is coupled to an output of the always-on audio detection system 104, so that when the always-on audio detection system 104 detects an audio trigger, an interrupt signal can be input to the interrupt line 128 to cause the processor 134 to return to its active or awake state so as to process the received command, allowing the device 100 to respond appropriately.


As discussed above, known devices that respond to audio commands can suffer from a delay between a user issuing a command and the device 100 outputting an acknowledgement signal. The always-on audio detection system 104 is configured to minimize or at least reduce this delay, as will now be explained with reference to FIG. 2.


In the flow chart 200 of FIG. 2, actions performed by the always-on audio detection system 104 are shown in the box labelled 210, whilst actions performed by the signal processor 102 are shown in the box labelled 220 and actions performed by the main processor 134 are shown in the box labelled 230.


At step 212, the always-on audio detection system 104 detects an audio trigger and command. As discussed above, the audio trigger may be, for example, a specific spoken word such as “Device” or a sound such as a handclap or whistle, whilst the command may be a spoken instruction such as “Dim the living room lights” or a question, such as “What is the weather forecast for tomorrow?”. In response to the detection of the audio trigger and command, the always-on audio detection system 104 issues an interrupt signal to the main processor 134 (step 214) and transmits the detected audio trigger and command to the main processor 134 (step 216).


In response to detection of the audio trigger and command by the always-on audio detection system 104, the signal processor 102 retrieves (at step 222) one or more acknowledgement output signals from the memory 108. Alternatively, the one or more acknowledgement output signals may be generated in real-time by the signal processor 102. The acknowledgement output signal(s) retrieved or generated may be dependent upon predefined settings such as user preferences, or upon output devices that are available or connected to the device 100.


For example, if the device 100 includes audio output transducer 116 and haptic output transducer 112, then the signal processor 102 may generate or retrieve from the memory 108 an audio output signal or waveform corresponding to an acknowledge audio tone to be output by the audio output transducer 116 and a haptic output signal or waveform corresponding to an acknowledge vibration effect to be output by the haptic output transducer.


Alternatively, if a wireless accessory device is connected to the device 100, the signal processor 102 may generate or retrieve from the memory 108 an audio output signal or waveform corresponding to an acknowledge audio tone to be output by an audio transducer of the wireless accessory device.


Similarly, if an external wired accessory device such as headphones, earphones or earbuds is connected to the device 100 by means of the jack socket 120, the signal processor 102 may retrieve from the memory 108 an audio output signal or waveform corresponding to an acknowledge audio tone to be output by the external wired accessory device.


If the device 100 includes display 128, then the signal processor 102 may generate or retrieve from memory 108 a video output signal corresponding to an acknowledge video signal to be output by the display 128. The acknowledge video signal may cause the display 128 to display an acknowledge image or message, or may cause a backlight of the display 128 to switch on to illuminate the display, for example.


Similarly, if the device 100 includes lamp 132, then the signal processor 102 may generate or retrieve from memory 108 a signal corresponding to an acknowledge lamp signal to be output by the lamp 132. The acknowledge lamp signal may cause the lamp 132 to switch on for a predetermined period of time, or may cause the lamp to switch on and off in a predetermined sequence, for example.


At step 224 the signal processor 102 outputs the acknowledgement output signals to the appropriate output driver(s) 110, 114, 118, 122, 126, 130, which in turn cause the appropriate acknowledge audio and/or haptic signal to be output by the audio output transducer 116 or haptic output transducer 112, or by a wired or wireless external accessory connected to the device 100 by means of the jack socket 120 or wireless subsystem 122 and antenna 124 respectively, and cause the appropriate acknowledge video or lamp signals to be output by the display 128 or lamp 132 if required.


It will be appreciated that the appropriate acknowledgement output signals may be output by the signal processor 102 to any of, or any combination of, the output drivers 110, 114, 118, 122, 126, 130.


Thus, in the event that a user has configured the device 100 to provide only haptic acknowledgement signals, the signal processor 102 may output only the haptic acknowledgement output signal, thereby causing only the haptic output transducer 112 to output an acknowledge signal.


Similarly, if the user has configured the device to output both audio and haptic acknowledgement signals, the signal processor 102 may output the haptic acknowledge signal to the haptic output amplifier 110 and may output the audio acknowledge signal to the audio output amplifier 114, such that both audio and haptic acknowledgement signals are provided to the user on detection by the always on audio detection system 104 of an audio trigger signal.


If the user has configured the device to output a video or lamp acknowledgement signal (in addition to or instead of other types of acknowledgement output signals), the signal processor 102 may output the video acknowledge signal to the display subsystem 126 and may output the lamp acknowledge signal to the lamp subsystem 130, to cause an appropriate acknowledgement output to be output by the display 128 or the lamp 132.


As a further example, if a wired accessory device is connected to the device 100 by means of the jack socket 120 or a wireless accessory device is connected to the device 100 by means of the wireless sub-system 122 and antenna 124, the signal processor 102 may output the haptic acknowledge signal to the haptic output amplifier 110 and may also output the audio acknowledge signal to the headphone amplifier 118 or wireless sub-system 122 as appropriate, so that a haptic acknowledgement signal is provided at the device 100 (via haptic output transducer 112) and an audio acknowledgement signal is provided at the connected accessory device.


While the signal processor 102 is performing steps 222 and 224 to provide a rapid acknowledgement to the user that the audio trigger and command have been received, the main processor 134, having received the audio trigger and command from the always-on audio detection system 104, enters its active or awake state (at step 232), and, at step 234, transmits the audio trigger via the Internet to a remote server for verification of the audio trigger. If the remote server verifies that the audio trigger is a valid audio trigger, it sends a verification message to the main processor 134, which is received at step 236. On receipt of this verification message, the main processor 134 causes the device 100 to respond to the command in an appropriate manner, at step 238, for example by performing an action or providing a spoken summary of information retrieved by the main processor 134 from the Internet in response to the command.


In the example described above, the received audio trigger and command are transmitted to the main processor 134 together after both the audio trigger and command have been detected. In an alternative approach, the always-on audio detection system 104 may be configured to transmit an interrupt signal to the main processor 134 to cause the main processor 134 to enter its active or awake state as soon as the audio trigger has been detected. In this case the signal processor 102 also outputs appropriate acknowledgement output signals in response to detection of the audio trigger.


In this approach, a buffer 136 (which may be provided as part of the signal processor 102, as shown in FIG. 1, or may be provided elsewhere in the device 100) is used to store audio input received by the audio input transducer arrangement 106 after the audio trigger has been received. Such subsequent audio input may correspond to a voice command issued by the user. The contents of the buffer 136 are then transmitted to the main processor 134 when the main processor 134 has entered its active or awake state in response to receiving the interrupt signal from the always-on audio detection system 104. For example, when the main processor 134 has entered its active or awake state, it may transmit a signal to the signal processor 102 indicating that it has entered its active or awake state, and in response to receiving this signal the signal processor 102 may transmit the contents of the buffer 136 to the main processor 134 for processing to execute the user's command.


It will be appreciated that the principles described above are applicable in devices in which an input transducer other than an audio input transducer is provided to detect a trigger signal that is used to indicate that the user wishes to issue a voice command to a device.



FIG. 3 is a schematic diagram showing part of a device 300 having an always-on detection system for detecting a trigger signal indicating that a user of the device wishes to issue a voice command to the device. The device 300 of FIG. 3 is similar to the device 100 of FIG. 1, so like reference signs have been used to represent like elements, and only those elements of the device 300 of FIG. 3 that differ from the device 100 of FIG. 1 will be described here.


The device 300 of FIG. 3 differs from the device 100 of FIG. 1 in that the always-on audio detection system 104 is replaced by an always-on detection system 304 which is configured to detect a trigger signal generated by an input transducer 306 in response to a user input indicative that the user of the device 300 wishes to issue a voice command to the device. The input transducer 306 need not be an audio input transducer, but instead may be any transducer that is able to generate a trigger signal in response to a user input. The input transducer 306 may be, for example, a push switch or button, a touch sensitive surface such as a touch-screen or touch pad, a movement sensor such an accelerometer or gyroscope, or a force-sensing transducer.


The always-on detection system 304 operates in substantially the same manner as the always-on audio detection system 104 of the device 100 of FIG. 1, as described above with reference to FIG. 2. However, instead of detecting an audio trigger, the always-on detection system is configured to detect a trigger signal of a kind that is appropriate to the type of input transducer 306 that is provided.


For example, where the input transducer 306 is a button or push switch, a trigger signal may be generated by a user depressing the button or push switch. A signal output by the button or push switch may be identified by the always-on detection system 304 as a trigger signal if the duration of the signal (corresponding to the length of time for which the button or switch is depressed by the user) exceeds a predefined threshold, for example.


Where the input transducer 306 is a touch sensitive surface, a signal output by the input transducer 306 corresponding to a predefined touch or gesture on the touch sensitive surface may be identified by the always-on detection system 304 as a trigger signal.


Where the input transducer 306 is a movement sensor such an accelerometer or gyroscope, a signal output by the input transducer 306 corresponding to a predefined action such as shaking or tapping the device 300 may be identified by the always-on detection system 304 as a trigger signal.


Where the input transducer 306 is a force-sensing transducer, a signal output by the input transducer 306 corresponding to a predetermined action such as squeezing or pressing part of the device 300 may be identified by the always-on detection system 304 as a trigger signal.


The always-on detection system 304 responds to detection of an identified trigger signal by issuing an interrupt signal to the main processor 134 to cause the main processor 134 to enter its active state, and transmitting a voice command subsequently received from the user by the audio input transducer arrangement 106 to the main processor 134.


Simultaneously or subsequently, the signal processor 102 retrieves one or more acknowledgement output signals (e.g. an audio acknowledgement signal or a haptic acknowledgement signal) from the memory 108 and outputs the acknowledgement output signal(s) to one or more of the output drivers (haptic output amplifier 110, audio output amplifier 114, headphone amplifier 118 or wireless subsystem 122), thereby providing a rapid acknowledgement to the user that the trigger signal has been received by the device 300.


In response to the interrupt signal the main processor 134 enters its active or awake state and responds to the voice command in an appropriate manner, by performing an action or transmitting the providing a spoken summary of information retrieved by the main processor 134 from the Internet in response to the command.


It will be appreciated that the system described herein provides a rapid acknowledgement that a user's trigger signal and command have been received by the device 100. This rapid acknowledgement ensures that the user is aware that the device has received the trigger signal and command, thereby reducing the likelihood that the user will repeatedly issue the command, thus improving the ease of use of the device 100 and the user's experience of using the device 100.

Claims
  • 1. A system for providing an acknowledgement output at an output transducer of a device in response to detection of a trigger signal, the system comprising: an input transducer;a signal processor coupled to the input transducer;a main processor separate from the signal processor; andan output driver,wherein the signal processor is configured to: monitor an input signal received at the signal processor from the input transducer; andif the received input signal corresponds to the trigger signal: output an acknowledgement signal to the output driver to cause the output driver to drive the output transducer to output the acknowledgement output; andoutput an interrupt signal to the main processor to cause the main processor to enter an active state and to initiate verification that the input signal is a valid trigger signal.
  • 2. A system according to claim 1 wherein the input transducer is an audio input transducer and wherein the signal processor is configured to monitor an audio input signal received at the signal processor from the audio input transducer.
  • 3. A system according to claim 2 wherein the signal processor comprises an always-on audio detection system configured to monitor the audio input signal received at the signal processor from the audio input device.
  • 4. A system according to claim 2 wherein the trigger signal is a voice command.
  • 5. A system according to claim 2 further comprising a buffer, wherein the buffer is configured to store audio input received by the audio input transducer after the trigger signal has been received, and wherein the system is configured to transmit the contents of the buffer to the main processor after the main processor has entered its active state.
  • 6. A system according to claim 1 further comprising a memory, wherein the memory is configured to store one or more acknowledgement signals corresponding to the acknowledgement output.
  • 7. A system according to claim 6 wherein the one or more acknowledgement signals corresponding to the acknowledgement output comprise one or more of: an audio acknowledgement signal or waveform;a haptic acknowledgement signal or waveform; anda visual acknowledgement signal.
  • 8. A system according to claim 6 wherein the memory is on-board memory of the signal processor.
  • 9. A system according to claim 1 wherein the output driver comprises a haptic output driver and the acknowledgement output comprises a haptic output, and wherein the signal processor is configured to output a haptic acknowledgement signal to the haptic output driver.
  • 10. A system according to claim 1 wherein the output driver comprises an audio output driver and the acknowledgement output comprises an audio output, and wherein the signal processor is configured to output an audio acknowledgement signal to the audio output driver.
  • 11. A system according to claim 1 wherein the output driver comprises a visual output driver and the acknowledgement output comprises a visual output, and wherein the signal processor is configured to output a visual acknowledgement signal to the visual output driver.
  • 12. A system according to claim 1 wherein the main processor is configured to enter an inactive state when not in use.
  • 13. A signal processor for providing an acknowledgement output signal at a haptic output transducer of a device in response to detection of a trigger signal when a main processor of the device is in an inactive state, wherein the signal processor is configured to: monitor an input signal received at the signal processor from an input transducer; andif the received input signal corresponds to the trigger signal: retrieve, from a memory of the device, an acknowledgement signal comprising a haptic waveform corresponding to an acknowledge effect to be output by the haptic output transducer;output the acknowledgement signal to an output driver of the device to cause the output driver to drive the haptic output transducer to output the acknowledgement output; andoutput an interrupt signal to cause the main processor of the device to enter an active state and to initiate verification that the input signal is a valid trigger signal.
  • 14. A signal processor according to claim 13 wherein the input transducer is an audio input transducer and wherein the signal processor is configured to monitor an audio input signal received at the signal processor from the audio input transducer.
  • 15. A signal processor according to claim 14 wherein the signal processor comprises an always-on audio detection system configured to monitor the audio input signal received at the signal processor from the audio input device.
  • 16. A signal processor according to claim 14 wherein the trigger signal is a voice command.
  • 17. A system for providing an acknowledgement output at a haptic output transducer of a device in response to detection of a trigger signal, the system comprising: an input transducer;a signal processor coupled to the input transducer;a main processor separate from the signal processor;an output driver; anda memory, wherein the memory is configured to store one or more acknowledgement signals corresponding to the acknowledgement output,wherein the signal processor is configured to: monitor an input signal received at the signal processor from the input transducer;andif the received input signal corresponds to the trigger signal: retrieve, from the memory, an acknowledgement signal comprising a haptic waveform corresponding to an acknowledge effect to be output by the haptic output transducer;output the acknowledgement signal to the output driver to cause the output driver to drive the haptic output transducer to output the acknowledgement output; andoutput an interrupt signal to the main processor to cause the main processor to enter an active state and to initiate verification that the input signal is a valid trigger signal.
REFERENCE TO PREVIOUSLY-FILED APPLICATION

The present disclosure is a continuation of U.S. Non-provisional patent application Ser. No. 15/875,175, filed Jan. 19, 2018, which is incorporated by reference herein in its entirety.

US Referenced Citations (150)
Number Name Date Kind
4902136 Mueller et al. Feb 1990 A
5684722 Thorner et al. Nov 1997 A
5748578 Schell May 1998 A
5857986 Moriyasu Jan 1999 A
6050393 Murai et al. Apr 2000 A
6278790 Davis et al. Aug 2001 B1
6332029 Azima et al. Dec 2001 B1
6388520 Wada et al. May 2002 B2
6580796 Kuroki Jun 2003 B1
6683437 Tierling Jan 2004 B2
6703550 Chu Mar 2004 B2
6762745 Braun et al. Jul 2004 B1
6906697 Rosenberg Jun 2005 B2
7154470 Tierling Dec 2006 B2
7623114 Rank Nov 2009 B2
7639232 Grant et al. Dec 2009 B2
7791588 Tierling et al. Sep 2010 B2
7979146 Ullrich et al. Jul 2011 B2
8068025 Devenyi et al. Nov 2011 B2
8098234 Lacroix et al. Jan 2012 B2
8102364 Tierling Jan 2012 B2
8325144 Tierling et al. Dec 2012 B1
8427286 Grant et al. Apr 2013 B2
8466778 Hwang et al. Jun 2013 B2
8480240 Kashiyama Jul 2013 B2
8572293 Cruz-Hernandez et al. Oct 2013 B2
8572296 Cruz-Hernandez et al. Oct 2013 B2
8593269 Grant et al. Nov 2013 B2
8648829 Shahoian et al. Feb 2014 B2
8659208 Rose et al. Feb 2014 B1
8754757 Ullrich et al. Jun 2014 B1
8947216 Da Costa et al. Feb 2015 B2
8981915 Birnbaum et al. Mar 2015 B2
8994518 Gregorio et al. Mar 2015 B2
9030428 Fleming May 2015 B2
9063570 Weddle et al. Jun 2015 B2
9083821 Hughes Jul 2015 B2
9092059 Bhatia Jul 2015 B2
9117347 Matthews Aug 2015 B2
9128523 Buuck et al. Sep 2015 B2
9164587 Da Costa et al. Oct 2015 B2
9196135 Shah et al. Nov 2015 B2
9248840 Truong Feb 2016 B2
9329721 Buuck et al. May 2016 B1
9354704 Lacroix et al. May 2016 B2
9368005 Cruz-Hernandez et al. Jun 2016 B2
9489047 Jiang et al. Nov 2016 B2
9507423 Gandhi et al. Nov 2016 B2
9513709 Gregorio et al. Dec 2016 B2
9520036 Buuck et al. Dec 2016 B1
9588586 Rihn Mar 2017 B2
9640047 Choi et al. May 2017 B2
9652041 Jiang et al. May 2017 B2
9697450 Lee Jul 2017 B1
9842476 Rihn et al. Dec 2017 B2
9864567 Seo Jan 2018 B2
9881467 Levesque Jan 2018 B2
9946348 Ullrich et al. Apr 2018 B2
9959744 Koskan et al. May 2018 B2
10032550 Zhang et al. Jul 2018 B1
10055950 Saboune et al. Aug 2018 B2
10074246 Da Costa et al. Sep 2018 B2
10110152 Hajati Oct 2018 B1
10175763 Shah Jan 2019 B2
10447217 Zhao et al. Oct 2019 B2
10564727 Billington et al. Feb 2020 B2
10620704 Rand et al. Apr 2020 B2
10732714 Rao et al. Aug 2020 B2
20020018578 Burton Feb 2002 A1
20030068053 Chu Apr 2003 A1
20030214485 Roberts Nov 2003 A1
20060028095 Maruyama et al. Feb 2006 A1
20060284856 Soss Dec 2006 A1
20080293453 Atlas et al. Nov 2008 A1
20080316181 Nurmi Dec 2008 A1
20090079690 Watson et al. Mar 2009 A1
20090088220 Persson Apr 2009 A1
20090102805 Meijer et al. Apr 2009 A1
20090153499 Kim et al. Jun 2009 A1
20100141408 Doy et al. Jun 2010 A1
20100141606 Bae et al. Jun 2010 A1
20110141052 Bernstein et al. Jun 2011 A1
20110167391 Momeyer et al. Jul 2011 A1
20120105358 Momeyer May 2012 A1
20120206246 Cruz-Hernandez et al. Aug 2012 A1
20120206247 Bhatia et al. Aug 2012 A1
20120229264 Company Bosch et al. Sep 2012 A1
20120306631 Hughes Dec 2012 A1
20130027359 Schevin et al. Jan 2013 A1
20130127755 Lynn et al. May 2013 A1
20130141382 Simmons et al. Jun 2013 A1
20130275058 Awad Oct 2013 A1
20130289994 Newman et al. Oct 2013 A1
20140056461 Afshar Feb 2014 A1
20140064516 Cruz-Hernandez et al. Mar 2014 A1
20140079248 Short et al. Mar 2014 A1
20140085064 Crawley et al. Mar 2014 A1
20140139327 Bau et al. May 2014 A1
20140292501 Lim et al. Oct 2014 A1
20140340209 Lacroix et al. Nov 2014 A1
20150070260 Saboune et al. Mar 2015 A1
20150084752 Heubel et al. Mar 2015 A1
20150216762 Oohashi et al. Aug 2015 A1
20150324116 Marsden et al. Nov 2015 A1
20150341714 Ahn et al. Nov 2015 A1
20160074278 Muench et al. Mar 2016 A1
20160132118 Park et al. May 2016 A1
20160162031 Westerman et al. Jun 2016 A1
20160179203 Modarres et al. Jun 2016 A1
20160239089 Taninaka et al. Aug 2016 A1
20160246378 Sampanes et al. Aug 2016 A1
20160358605 Ganong, III et al. Dec 2016 A1
20170052593 Jiang et al. Feb 2017 A1
20170078804 Guo et al. Mar 2017 A1
20170083096 Rihn et al. Mar 2017 A1
20170153760 Chawda et al. Jun 2017 A1
20170256145 Macours et al. Sep 2017 A1
20170277350 Wang et al. Sep 2017 A1
20170357440 Tse Dec 2017 A1
20180059733 Gault et al. Mar 2018 A1
20180059793 Hajati Mar 2018 A1
20180074637 Rosenberg et al. Mar 2018 A1
20180082673 Tzanetos Mar 2018 A1
20180084362 Zhang et al. Mar 2018 A1
20180151036 Cha et al. May 2018 A1
20180158289 Vasilev et al. Jun 2018 A1
20180160227 Lawrence et al. Jun 2018 A1
20180178114 Mizuta et al. Jun 2018 A1
20180182212 Li et al. Jun 2018 A1
20180183372 Li et al. Jun 2018 A1
20180237033 Hakeem et al. Aug 2018 A1
20180253123 Levesque et al. Sep 2018 A1
20180294757 Feng et al. Oct 2018 A1
20180301060 Israr et al. Oct 2018 A1
20180321748 Rao et al. Nov 2018 A1
20180329172 Tabuchi Nov 2018 A1
20180367897 Bjork et al. Dec 2018 A1
20190227628 Rand et al. Jan 2019 A1
20190064925 Kim et al. Feb 2019 A1
20190073078 Sheng et al. Mar 2019 A1
20190103829 Vasudevan et al. Apr 2019 A1
20190138098 Shah May 2019 A1
20190206396 Chen Jul 2019 A1
20190215349 Adams et al. Jul 2019 A1
20190114496 Lesso Aug 2019 A1
20190294247 Hu et al. Sep 2019 A1
20190296674 Janko et al. Sep 2019 A1
20190297418 Stahl Sep 2019 A1
20190311590 Doy et al. Oct 2019 A1
20190341903 Kim Nov 2019 A1
Foreign Referenced Citations (22)
Number Date Country
2002347829 Apr 2003 AU
103165328 Jun 2013 CN
0784844 Jun 2005 EP
2363785 Sep 2011 EP
2600225 Jun 2013 EP
2846218 Mar 2015 EP
2846329 Mar 2015 EP
3125508 Feb 2017 EP
3379382 Sep 2018 EP
201747044027 Aug 2018 IN
H02130433 May 1990 JP
6026751 Nov 2016 JP
6250985 Dec 2017 JP
6321351 May 2018 JP
20120126446 Nov 2012 KR
2013104919 Jul 2013 WO
2013186845 Dec 2013 WO
2014018086 Jan 2014 WO
2016105496 Jun 2016 WO
2016164193 Oct 2016 WO
2017113651 Jul 2017 WO
2018053159 Mar 2018 WO
Non-Patent Literature Citations (12)
Entry
International Search Report and Written Opinion of the International Searching Authority, International Application No. PCT/GB2019/050964, dated Sep. 3, 2019.
International Search Report and Written Opinion of the International Searching Authority, International Application No. PCT/GB2019/050770, dated Jul. 5, 2019.
Communication Relating to the Results of the Partial International Search, and Provisional Opinion Accompanying the Partial Search Result, of the International Searching Authority, International Application No. PCT/US2018/031329, dated Jul. 20, 2018.
Combined Search and Examination Report, UKIPO, Application No. GB1720424.9, dated Jun. 5, 2018.
International Search Report and Written Opinion of the International Searching Authority, International Application No. PCT/US2020/023342, dated Jun. 9, 2020.
International Search Report and Written Opinion of the International Searching Authority, International Application No. PCT/GB2020/050823, dated Jun. 30, 2020.
International Search Report and Written Opinion of the International Searching Authority, International Application No. PCT/US2020/024864, dated Jul. 6, 2020.
International Search Report and Written Opinion of the International Searching Authority, International Application No. PCT/GB2020/051037, dated Jul. 9, 2020.
Communication Relating to the Results of the Partial International Search, and Provisional Opinion Accompanying the Partial Search Result, of the International Searching Authority, International Application No. PCT/GB2020/050822, dated Jul. 9, 2020.
International Search Report and Written Opinion of the International Searching Authority, International Application No. PCT/GB2020/051035, dated Jul. 10, 2020.
International Search Report and Written Opinion of the International Searching Authority, International Application No. PCT/GB2020/050822, dated Aug. 31, 2020.
International Search Report and Written Opinion of the International Searching Authority, International Application No. PCT/GB2020/051438, dated Sep. 28, 2020.
Related Publications (1)
Number Date Country
20190394592 A1 Dec 2019 US
Continuations (1)
Number Date Country
Parent 15875175 Jan 2018 US
Child 16562037 US