Computing devices have gained in sophistication to users by processing audio instructions and providing responses. Users can recite audio instructions which may be used to control these computing devices. For example, users can speak to the computing devices to provide information, such as instructions to provide directions to a particular location.
In the accompanying drawings, like numerals refer to like components or blocks. The following detailed description references the drawings, wherein:
In audio information processing, a user typically activates the application to process audio by pressing a button and/or reciting instructions. Once launching the audio processing application, the user additionally needs to recite explicit instructions they would desire a computing device to perform. Thus, processing speech instructions from a user can be time consuming and repetitive. In addition, continuously monitoring for instructions from the user consumes much power, draining the battery.
To address these issues, example embodiments disclosed herein use a low power integrated circuit to continuously monitor for occurrence of a keyword in an audio stream (e.g., the user's speech), while relying on a processor for more thorough analysis of the user's speech. For example, various examples disclosed herein provide for receiving an audio stream in a low power integrated circuit, digitizing the audio stream, and analyzing the digitized audio stream to recognize a keyword. Once recognizing the keyword within the digitized audio stream, the integrated circuit sends a signal to the processor to increase power. Once increasing power to the processor, the digitized audio stream is retrieved to determine a response. This decreases the amount of time consumed for the user to launch the specific audio processing application and prevents repetition of the user's speech. Determining the response from the retrieved audio stream prevents the user from providing additional explicit instructions for the computing device to perform the speech analysis.
Additionally, in the various examples disclosed herein, once increasing power to the processor, the processor retrieves the digitized audio stream from a memory and converts the digitized audio stream to a text stream. After converting to the text stream, the processor determines a response based on text within the text stream. Determining the response from the text stream reduces the time for the user of the computing device to instruct the computing device. Additionally still, the processor may determine the appropriate response, based on the context of the audio stream. Further, the computing device determines which application needs to execute in order to fulfill the response to the user. Further still, by increasing power to the processor once recognizing the keyword within the digitized audio stream, the computing device consumes less power while listening for the user's speech.
In one embodiment, the computing device may also determine the response by receiving the response from a server or from the processor. In a further embodiment, the memory maintains the stored digitized audio stream for a predetermined period of time. In this embodiment, the processor can retrieve the digitized audio stream in time increments. For example, the processor may retrieve the complete digitized audio stream or may retrieve a shorter time interval of the digitized audio stream. The retrieval of the digitized audio stream allows the processor to analyze the context of the audio stream to determine the appropriate response.
In this manner, example embodiments disclosed herein save a user time by preventing repetitive audio instructions to a computing device since the computing device determines an appropriate response based on the context of the audio stream. Further, the computing device consumes less power while receiving and processing audio streams.
Referring now to the drawings,
The audio stream 102 is received by the computing device 100, specifically, the low power integrated circuit 104. The audio stream 102 is an input analog signal that is digitized 106 to provide the digitized audio stream 114. Embodiments of the audio stream 102 include speech from a user or audio from another computing device. For example, there may be several computing devices 300 receiving audio streams 102, which may be confusing. Thus, the computing devices may designate one device as a central point to receive the audio stream 102. In this embodiment, the low power integrated circuit 104 operates as part of an ad-hoc network that may be a central unit of one or more computing devices.
For example, the user may discuss with another person the shortest route from New York to Los Angeles, Calif. In this example, the audio stream would be the discussion of the shortest route from New York to Los Angeles. In a further embodiment, the audio stream 102 may include audio for a predetermined period of time. For example, the audio stream 102 may include a few seconds or minutes when received by the low power integrated circuit 104. In this example, the low power integrated circuit 104 may distinguish the audio stream 102 from other audio streams 102.
The low power integrated circuit 104 includes the module 106 to digitize the audio stream 102 and module 108 to compare the digitized audio stream 114 to the keyword. The low power integrated circuit 104 is an electronic circuit with patterned trace elements on the surface of a material that form interconnections between other electronic components. For example, the low power integrated circuit 104 forms connections between the processor 118 and the memory 112. Embodiments of the low power integrated circuit 104 include a microchip, chipset, electronic circuit, chip, microprocessor, semiconductor, microcontroller, or other electronic circuit capable of receiving audio stream 102 and transmitting signal 116. The low power integrated circuit 104 may continuously monitor the audio stream 102, utilize digitize module 106 to digitize the audio stream, and store the digitized audio stream at the memory 112. As such, further embodiments of the low power integrated circuit 104 include a transmitter, receiver, microphone, or other suitable component to receive the audio stream 102.
The audio stream is digitized at module 106 to provide the digitized audio stream 114. The digitize module 106 converts the audio stream to a discrete time signal representation. Embodiments of the digitize module 106 include an analog to digital converter (ADC), digital conversion device, instruction, firmware and/or software operating in conjunction with low power integrated circuit 104. For example, the digitize module 106 may include an electronic device to convert an input analog voltage to a digital number proportional to the magnitude of the analog signal.
Once the audio stream 102 is digitized at module 106, it is compared to the keyword at module 108. The audio stream 102 and at module 108, it is compared against the keyword which operates as an indication to signal 116 the processor 118 to increase power 122 and obtain the digitized audio stream 114 to analyze at module 120. Embodiments of 108 include an instruction, process, operation, logic, algorithm, technique, logical function, firmware and/or software. Once the keyword is recognized, the low power integrated circuit 104 transmits the signal 116 to increase power 122 to the processor 118.
Embodiments of the keyword include a digital signal, analog signal, pattern, database, commands, directions, instructions, or other representation to compare at module 108. For example, the user of a computing device may discuss the difference between a shrimp and prawn with a friend and subsequently desire to perform a web search to identify the answer. As such, the user may state the predetermined keyword to trigger recognition of the keyword by compare to keyword module 108 and subsequent analysis of the previous discussion by analyze module 120.
The keyword may include, for example, a phrase, a single keyword, or a single keyword that is private to the user of the computing device. In keeping with the previous example, the keyword may be the phrase, “Computer, what do you think?”. In this example, the phrase causes the low power integrated circuit 104 to send the signal 116 to the processor 118 to obtain the digitized audio stream 114 which may include the audio before or after the phrase. Thus, the user does not need to repeat the instructions since the processor 118 analyzes the digitized audio stream 114 to determine the context of the audio stream 102 for an appropriate response. Yet, in a further example, the single keyword, may include “Shazam.” Thus, as a specific example, when the user speaks the word “Shazam,” circuit 104 may detect the keyword and transmit the signal 116 to instruct the processor 118 to obtain the digitized audio stream 114 and convert the stream to a text stream. Supposing the text stream is an instruction to compose a text message to the user's mother, the appropriate response would be to compose the text message. Thus, as described above, using the predetermined keyword(s), the low power integrated circuit 104 recognizes when the user of the computing device needs to complete a further response, such as directions or perform a web search.
In a further embodiment of module 108, when no keyword is recognized within the digitized audio stream 114, the low power integrated circuit 104 continues monitoring for another audio stream 102 which is digitized at module 106 and stored in the memory 112. In yet a further embodiment, the low power integrated circuit 104 compresses the digitized audio stream 114 and this compressed digitized audio stream is used to recognize the keyword by comparing it to the keyword at module 108.
The memory 112 stores and/or maintains the digitized audio stream 114. Embodiments of the memory 112 may include a memory buffer, cache, non-volatile memory, volatile memory, random access memory (RAM), an Electrically Erasable Programmable Read-Only Memory (EEPROM), storage drive, a Compact Disc Read-Only Memory (CDDROM), or other memory capable of storing and/or maintaining the digitized audio stream 114.
The digitized audio stream 114 is stored in the memory 112. Embodiments may include the low power integrated circuit 104 compressing the audio stream 102 after the digitization module 106 to obtain a compressed digitized audio stream prior to placement in the memory 112. Although
The signal 116 is transmitted from the low power integrated circuit 104 to the processor 118 upon recognition of the keyword within the digitized audio stream 114. The signal 116 instructs the processor 118 to increase power 122 and analyze the digitized audio stream 114 from the memory 12. Embodiments of the signal 116 include a communication, transmission, electrical signal, instruction, digital signal, analog signal, or other type of communication to increase power 122 to the processor 118. A further embodiment of the signal 116 includes an interrupt transmitted to the processor 118 upon recognition of the keyword within the digitized audio stream 114.
The processor 118 receives the signal 116 to increase power 122 and obtains the digitized audio stream 114 to analyze at module 120. Embodiments of the processor 118 may include a central processing unit (CPU), visual processing unit (VPU), microprocessor, graphics processing unit (GPU), or other programmable device suitable to analyze 120 the digitized audio stream 14.
Once the processor 118 obtains the digitized audio stream 114 from the memory 112, the processor analyzes the digitized audio stream 114 at module 120. Embodiments of the analyze module 120 include an instruction, process, operation, logic, algorithm, technique, logical function, firmware and/or software the processor 18 may fetch, decode, and/or execute. Additional embodiments of module 120 include converting the digitized audio stream 114 to a text stream to determine an appropriate response based on the context of the audio stream 102. Further embodiments of module 120 include determining a response to render to the user of the computing device 100 as will be seen in later figures.
The power 122 supplies electrical energy in the form of electrical potential to the processor 118. Specifically, the power 122 increases electrical energy to the processor 118 once receiving the signal 116 from the low power integrated circuit 104. Increasing the power 122 to the processor 118 wakes or triggers the processor 118 to obtain the digitized audio stream 114. Embodiments of the power 122 include a power supply, power management device, battery, energy storage, electromechanical system, solar power, power plug, or other device capable of delivering power 122 to the processor 118. In a further embodiment, power 122 supplies the electrical energy to the computing device 100.
Referring now to
The audio stream 202 is received by the low power integrated circuit 204. The audio stream 202 may be similar in structure to audio stream 102 of
The low power integrated circuit 204 includes the circuitry 210 to digitize the audio stream 202 and compare the digitized audio stream 214 to a keyword. The low power integrated circuit 204 may be similar in functionality and structure of the low power integrated circuit 104 as above in
The circuitry 210 includes digitize circuitry 206 and compare circuitry 208. Embodiments of the circuitry 210 include logic, analog circuitry, electronic circuitry, digital circuitry, or other circuitry capable of digitizing the audio stream 102 and comparing the digitized audio stream 214 to the keyword. In further embodiments, the circuitry includes an application and/or firmware which may be utilized independently and/or in conjunction with the low power integrated circuit 204 to fetch, decode, and or execute circuitry 206 and 208.
The audio stream 202 is received and digitized by circuitry 206 to produce the digitized audio stream 214. The digitize circuitry 206 is a type of conversion for the audio stream 202. Further, the digitize circuitry 206 may be similar in functionality of the digitize module 106 as described in connection with
The low power integrated circuit 204 receives the audio stream 202 to digitize at circuitry 206 and produces the digitized audio stream 214. The digitized audio stream 214 may be similar in structure to the digitized audio stream 114 as described in connection with
The circuitry 208 included in the circuitry 210 of the low power integrated circuit 204, compares the digitized audio stream 214 to the keyword. Further, 208 is used to recognize the keyword within the digitized audio stream 214 to transmit the signal 216 to increase power to the processor. The compare circuitry 208 may be similar in functionality to the module 108 as described in connection with
The signal 216 instructs a device to increase power upon recognition of the keyword within the digitized audio stream 214 by compare circuitry 208. The signal 216 may be similar in structure and functionality to signal 116 of
The audio stream 302 is received by the computing device 300, specifically, the low power integrated circuit 304. The audio stream 302 may be similar in structure to the audio stream 102 and 202 in
The low power integrated circuit 304 includes a digitize module 306 and an analyze module 308. In one embodiment, the low power integrated circuit 304 includes circuitry to comprise modules 306 and 308. The low power integrated circuit 304 may be similar in structure and functionality of the low power integrated circuit 104 and 204 described in connection with
The audio stream 302 once received by the computing device 300 is digitized 306 to produce a digitized audio stream 314. The digitize module 306 may be similar in structure and functionality to the digitize module 106 and digitize circuitry 206 in
Once the audio stream 314 is digitized, the low power integrated circuit analyzes the digitized audio stream 314 at module 308. In one embodiment, module 308 compares a keyword to the digitized audio stream 114. In this embodiment, 308 includes the functionality of the compare module 108 as above in
The memory 312 stores the digitized audio stream 314 from the low power integrated circuit 304. In one embodiment, the memory 312 maintains the digitized audio stream 314 received during a predetermined period of time. For example, the audio stream 302 may be monitored for the predetermined time of a few seconds and as such, this few seconds of the audio stream 302 is digitized at module 306 and sent to the memory 312. In this example, the memory 312 stores the digitized audio stream 314 of the few seconds to be retrieved and/or obtained by the processor 318 to analyze once receiving the signal 316. Also, in this example, when another audio stream 302 of a few seconds is received and digitized, this other digitized audio stream 314 replaces the prior digitized audio stream 314. This allows the memory 312 to maintain the most recent audio stream 302 for the processor 318 to obtain and/or retrieve. The memory 312 may be similar in structure and functionality of the memory 112 as described in connection with
The audio stream 302 is digitized 306 to produce the digitized audio stream 314. The digitized audio stream 314 is stored and/or maintained in the memory 312. In an embodiment, the processor 318 obtains the digitized audio stream 314 to analyze at module 320 once receiving the signal 316. The digitized audio stream 314 may be similar in structure and functionality of digitized audio stream 114 and 214 as described, in connection with
The signal 316 is a transmission from the low power integrated circuit 304 to the processor 316 to increase power 322. In an embodiment of the signal 316, additionally instructs the processor 316 to obtain the digitized audio stream 314 to analyze at module 320. The signal 316 may be similar in structure and functionality of the signal 116 and 216 as described in connection with
The power 322 supplies electrical energy to the processor 318 and/or computing device 300. The power 322 may be similar in structure and functionality of the power 122 as described in connection with
The processor 318 includes the analyze module 320 and text stream 324. Specifically, the processor 318 receives the signal 316 to increase power 322. Once receiving this signal 316, the processor 318 obtains the digitized audio stream 314 to analyze at module 320. In a further embodiment, the processor 318 converts the digitized audio stream 314 to the text stream 324. In this embodiment, the text within the text stream 324 dictates a response for the computing device 300. The text stream is a string of finite sequence of symbols or representations from an alphabet, numbered set, or alphanumeric set. For example, the digitized audio stream 314 may be in a binary language, thus the processor translates bytes of the binary representation to a word. In a further example, the digitized audio stream 314 may be in a language representative of words and/or numbers, thus the processor 318 translates this language into text the processor 318 comprehends. Embodiments of the response include performing a web search, dialing a phone number, opening an application, recording text, streaming media, composing a text message, listing direction, or speaking directions. In a further embodiment, the processor 318 determines the response to render to a user of the computing device 300. The processor 318 may be similar in structure and functionality of the processor 118 as described in connection with
The processor 318 analyzes the stored digitized audio stream 314 at module 320. Embodiments of the analyze module 320 include transmitting the digitized audio stream 314 obtained from the memory 314 to the server 326. Other embodiments of module 320 include converting the digitized audio stream 314 obtained from the memory 312 to the text stream 324 and transmitting the text stream 324 to the server 326. Other embodiments of module 320 include converting the digitized audio stream 314 to the text stream 324 to determine the appropriate response by analyzing the context of the audio stream 302. For example, the digitized audio stream 314 may be converted to the text stream 324 at module 320 and the processor 318 may utilize a natural language processing to analyze the text within the text stream 324 to determine the appropriate response based on the context of the audio stream 302.
The text stream 324 includes text to determine the appropriate response for the computing device 300. In one embodiment, the text stream 324 is processed by the processor to determine the appropriate response to render to the user of the computing device 300 on the output device 328. In another embodiment, the text stream 324 is processed by the server 326 to determine the appropriate response which is transmitted to the computing device 300. In this embodiment, the response is sent from the server 326 to the computing device 300. In a further embodiment, the computing device 300 renders the response to the user of the computing device 300. For example, the text stream 324 may include text that discusses sending a text message to mom. Thus, the text within the text stream 324 dictates for the computing device 300 to respond by composing a text message to mom.
The server 326 provides services across a network and may include, for example, a web server, a network server, a Local Area Network (LAN) server, a file server, or any other computing device suitable to process the text stream 324 to transmit the response to the computing device 300.
The output device 328 renders the response as determined from the text within the text stream 324 to the user of the computing device 300. Embodiments of the output device 328 include a display device, a screen, or a speaker to render the response to a user of the computing device 300. In keeping with the text message to mom example, the user of the computing device 300 may have a display that shows the text message being composed to mom and/or speaker to communicate to the user the text message.
Turning now to
At operation 402, the computing device operating in conjunction with a low power integrated circuit receives an audio stream. In one embodiment, the audio stream is of a predetermined amount of time. For example, the audio stream may be a few seconds or milliseconds. In this embodiment, the computing device may continuously monitor audio. In further embodiments, the audio stream includes at least one of a speech from a user or audio from the other computing device.
At operation 404, the low power integrated circuit operating in conjunction with the computing device digitizes the audio stream received at operation 402 to produce a digitized audio stream. Embodiments of operation 404 include the use of an analog to digital converter (ADC), digital conversion device, instruction, firmware, and/or software operating in conjunction with the low power integrated circuit. Embodiments of operation 404 include transmitting the digitized audio stream to a memory. Further embodiments of 404 include compressing the audio stream received at operation 402, while another embodiment of 404 includes compressing the digitized audio stream.
At operation 406, the digitized audio stream produced at operation 404 is stored in the memory. Embodiments of operation 406 include the memory storing and/or maintaining the digitized audio stream. In another embodiment of operation 406, the audio stream received during the predetermined amount of time at operation 402 is digitized at operation 404, thus when another audio stream is received at operation 402 and digitized at operation 404, this current digitized audio stream replaces the prior digitized audio stream. In this embodiment, the memory maintains the stored digitized audio stream received during the predetermined period of time prior to the current time.
At operation 408, the low power integrated circuit analyzes the digitized audio stream produced at operation 404. Embodiments of operation 408 include processing the digitized audio stream while other embodiments include comparing the digitized audio stream to a keyword. In these embodiments of operation 408, the low power integrated circuit processes the digitized audio stream for the keyword. Upon recognition of the keyword within the digitized audio stream, the method moves to operation 410 to transmit a signal. In a further embodiment, if the low power integrated circuit does not recognize the keyword within the digitized audio stream, the method returns to operation 402. Yet, in a further embodiment includes comparing the digitized audio stream to an analog or digital representation that indicates the user of the computing device desires a response by the computing device. In yet a further embodiment, operations 402, 404, 406, and 408 occur in parallel. For example, if the computing device analyzes the digitized audio stream at 408, the integrated circuit continues receiving audio streams at operation 402, digitizing, and storing the audio stream at operations 404 and 406.
At operation 410, the low power integrated circuit transmits the signal to the processor to increase power. Specifically, upon recognition of the keyword within the digitized audio stream, the low power integrated circuit transmits a signal to the processor to increase power. In an embodiment of operation 410, the processor increases power or electrical energy delivered to the processor and/or the computing device.
At operation 412, the processor obtains the stored digitized audio stream from the memory at operation 406. In one embodiment of operation 412, the memory transmits the digitized audio stream to the processor, while in another embodiment of operation 412, the processor retrieves the digitized audio stream from the memory.
At operation 414, the processor converts the digitized audio stream obtained at operation 412 to a text stream. After converting the digitized audio stream to the text stream, the processor analyzes the text within the text stream to determine the appropriate response. Embodiments of operation 414 include using speech to text (STT), voice to text, digital to text, or other type of text conversion. A further embodiment of operation 414 includes using a natural language processing after conversion to the text stream. In this embodiment, the computing device processes the text within the text stream to determine an appropriate response based on the context of the audio stream received at operation 402. For example, once detecting the keyword within the digitized audio stream at 408, the processor obtains at operation 412, and the digitized audio stream is converted to the text stream at operation 414. In a further example, the audio stream may include a conversation regarding directions between two locations, thus once this digitized audio stream is converted at operation 412 to the text stream, —the processor can determine the appropriate response by analyzing the text within the text stream.
At operation 416, the processor determines the response based on the text stream produced at operation 414. Embodiments of the response include performing a web search, dialing a phone number, opening an application, recording text, streaming media, composing a text message, listing directions, or speaking directions. In one embodiment, the text within the text stream dictates the appropriate response for the processor. In a further embodiment, the response is rendered to a user of the computing device. For example, the text stream may include speech inquiring how to reach China and as such directions to China would be the appropriate response. Additionally, in this example, a map display listing and/or speaking directions to China may be included.
Referring now to
At operation 502, the computing device compresses a digitized audio stream. In one embodiment, operation 502 is performed in conjunction with operation 404 prior to operation 406 in
At operation 504, the computing device renders a response to the user of the computing device. Embodiments of operation 504 include occurring during or after operation 416 in
The embodiments described in detail herein relate to digitizing an audio stream to detect a keyword and based upon recognition of the keyword within the digitized audio stream, transmitting a signal to a processor to increase power and further analyze the digitized audio stream to determine a response. In this manner, example embodiments save a user time by preventing repetitive audio instructions to a computing device, while reducing power consumption of the computing device.
The present application is a continuation of U.S. patent application Ser. No. 16/453,779, filed on Jun. 26, 2019, entitled “LOW POWER INTEGRATED CIRCUIT TO ANALYZE A DIGITIZED AUDIO STREAM”, which is a continuation of U.S. patent application Ser. No. 15/400,817, filed on Jan. 6, 2017, U.S. Pat. No. 10,381,007, issued Aug. 13, 2019 and entitled “LOW POWER INTEGRATED CIRCUIT TO ANALYZE A DIGITIZED AUDIO STREAM”, which is a continuation of U.S. patent application Ser. No. 14/363,783, filed on Oct. 21, 2014, U.S. Pat. No. 9,564,131, issued Feb. 7, 2017 and entitled “LOW POWER INTEGRATED CIRCUIT TO ANALYZE A DIGITIZED AUDIO STREAM”, which is a U.S. National Phase Application of PCT/US2011/063804, filed Dec. 7, 2011, entitled “LOW POWER INTEGRATED CIRCUIT TO ANALYZE A DIGITIZED AUDIO STREAM”, all of which are incorporated herein by reference in their entirety.
Number | Name | Date | Kind |
---|---|---|---|
5278944 | Sasaki et al. | Jan 1994 | A |
5721938 | Stuckey | Feb 1998 | A |
5946050 | Wolff | Aug 1999 | A |
5983186 | Miyazawa et al. | Nov 1999 | A |
6070140 | Tran | May 2000 | A |
6101473 | Scott et al. | Aug 2000 | A |
6253176 | Janek et al. | Jun 2001 | B1 |
6408272 | White et al. | Jun 2002 | B1 |
6594630 | Zlokarnik et al. | Jul 2003 | B1 |
6718307 | Buil et al. | Apr 2004 | B1 |
7567827 | Kim | Jul 2009 | B2 |
7774626 | Fleming | Aug 2010 | B2 |
7818176 | Freeman et al. | Oct 2010 | B2 |
8078469 | White et al. | Dec 2011 | B2 |
8108218 | Huboi | Jan 2012 | B1 |
8111839 | Goldstein et al. | Feb 2012 | B2 |
8150044 | Goldstein et al. | Apr 2012 | B2 |
8150700 | Shin et al. | Apr 2012 | B2 |
8160884 | Roth et al. | Apr 2012 | B2 |
8266451 | Eydier et al. | Sep 2012 | B2 |
8359020 | Ebeau et al. | Jan 2013 | B2 |
8370145 | Endo | Feb 2013 | B2 |
8488799 | Goldstein et al. | Jul 2013 | B2 |
8553905 | Goldstein et al. | Oct 2013 | B2 |
8560307 | El-Maleh et al. | Oct 2013 | B2 |
8606293 | Kim et al. | Dec 2013 | B2 |
8666751 | Murthi et al. | Mar 2014 | B2 |
8682667 | Haughay | Mar 2014 | B2 |
8798995 | Edara | Aug 2014 | B1 |
8880405 | Cerra et al. | Nov 2014 | B2 |
8918141 | Goldstein et al. | Dec 2014 | B2 |
8924219 | Bringert | Dec 2014 | B1 |
9031847 | Sarin et al. | May 2015 | B2 |
9112989 | Lee | Aug 2015 | B2 |
9124982 | Goldstein et al. | Sep 2015 | B2 |
9444939 | Ryan et al. | Sep 2016 | B2 |
9564131 | Liu et al. | Feb 2017 | B2 |
9992745 | Hwang et al. | Jun 2018 | B2 |
10381007 | Liu et al. | Aug 2019 | B2 |
10405082 | Usher | Sep 2019 | B2 |
20020046033 | Ono et al. | Apr 2002 | A1 |
20020077830 | Suomela et al. | Jun 2002 | A1 |
20030157968 | Boman et al. | Aug 2003 | A1 |
20030193964 | Bae et al. | Oct 2003 | A1 |
20040002862 | Kim et al. | Jan 2004 | A1 |
20040054532 | Staiger | Mar 2004 | A1 |
20050091052 | Chang et al. | Apr 2005 | A1 |
20050114132 | Hsu | May 2005 | A1 |
20060074658 | Chadha | Apr 2006 | A1 |
20070192109 | Likens et al. | Aug 2007 | A1 |
20070239837 | Jablokov et al. | Oct 2007 | A1 |
20070254680 | Nassimi | Nov 2007 | A1 |
20080168189 | Aldaz et al. | Jul 2008 | A1 |
20080221891 | Konig et al. | Sep 2008 | A1 |
20090055005 | Oxman et al. | Feb 2009 | A1 |
20090070119 | Yoo et al. | Mar 2009 | A1 |
20090209300 | Furbeck | Aug 2009 | A1 |
20090234655 | Kwon | Sep 2009 | A1 |
20090259865 | Sheynblat et al. | Oct 2009 | A1 |
20090271190 | Niemisto et al. | Oct 2009 | A1 |
20100142619 | Suzuki et al. | Jun 2010 | A1 |
20110066431 | Ju et al. | Mar 2011 | A1 |
20110213613 | Cohen et al. | Sep 2011 | A1 |
20110245946 | Kim et al. | Oct 2011 | A1 |
20110246206 | Kim et al. | Oct 2011 | A1 |
20120010890 | Koverzin | Jan 2012 | A1 |
20120224706 | Hwang et al. | Sep 2012 | A1 |
20130054243 | Ichikawa | Feb 2013 | A1 |
20140214429 | Pantel | Jul 2014 | A1 |
20180197533 | Yon et al. | Jul 2018 | A1 |
20190385612 | Liu et al. | Dec 2019 | A1 |
Number | Date | Country |
---|---|---|
1302397 | Jul 2001 | CN |
1564256 | Jan 2005 | CN |
1886918 | Dec 2006 | CN |
101076099 | Nov 2007 | CN |
101083809 | Dec 2007 | CN |
101164102 | Apr 2008 | CN |
101483683 | Jul 2009 | CN |
101938391 | Jan 2011 | CN |
101968791 | Feb 2011 | CN |
101971250 | Feb 2011 | CN |
201752079 | Feb 2011 | CN |
102023705 | Apr 2011 | CN |
102118886 | Jul 2011 | CN |
1562180 | Aug 2005 | EP |
1879000 | Jan 2008 | EP |
1978765 | Oct 2008 | EP |
2342828 | Apr 2000 | GB |
H07121195 | May 1995 | JP |
H07244494 | Sep 1995 | JP |
H10312194 | Nov 1998 | JP |
2000315097 | Nov 2000 | JP |
2003505712 | Feb 2003 | JP |
2004226698 | Aug 2004 | JP |
2004265217 | Sep 2004 | JP |
2004294946 | Oct 2004 | JP |
2005070367 | Mar 2005 | JP |
2007219207 | Aug 2007 | JP |
2007255897 | Oct 2007 | JP |
2008309864 | Dec 2008 | JP |
2009265219 | Nov 2009 | JP |
2010268324 | Nov 2010 | JP |
20010063844 | Jul 2001 | KR |
20030081835 | Oct 2003 | KR |
100447667 | Sep 2004 | KR |
100744301 | Jul 2007 | KR |
0063879 | Oct 2000 | WO |
2004084443 | Sep 2004 | WO |
2008004037 | Jan 2008 | WO |
2009097020 | Aug 2009 | WO |
2010078386 | Jul 2010 | WO |
2011037264 | Mar 2011 | WO |
2011059781 | May 2011 | WO |
Entry |
---|
Alon G., “Key-Word Spotting The Base Technology for Speech Analytics,” White Paper, Natural Speech Communication Ltd., Jul. 2005, 8 pages. |
Boucheron L.E., et al., “Hybrid Scalar/Vector Quantization of Mel-Frequency Cepstral Coefficients for Low Bit-Rate Coding of Speech,” IEEE Data Compression Conference (DCC), Mar. 2011, pp. 103-112. |
Bourke P.J., “A Low-Power Hardware Architecture for Speech Recognition Search,” Carnegie Mellon University, Pittsburgh, PA, May 2011, 166 pages. |
Ceron I.F.C., et al., “A Keyword Based Interactive Speech Recognition System for Embedded Applications,” School of Innovation, Design and Engineering, Malardalen University, Vasteras, Sweden, Jun. 2011, 96 pages. |
Delaney B., et al., “Energy Aware Distributed Speech Recognition for Wireless Mobile Devices”, Copyright Hewlett-Packard Company 2004, HP Laboratories Palo Alto HPL-2004-106, Jun. 17, 2004. |
English translation of Appeal against the Final Rejection on Korean Patent Application No. 10-2014-7018606, dated Apr. 18, 2019, 14 Pages. |
International Search Report and Written Opinion—PCT/US2011/063804—ISA/EPO—dated Aug. 29, 2012 (141997WO). |
Korean Office Action dated Apr. 18, 2019 for Korean Patent Application Serial No. 10-2014-7018606, 29 Pages. |
State Intellectual Property Office of the People's Republic of China Notification of Examination Decision for Invalidation Request, Application/Patent No. 201180076332.2; Case No. 4W107002, Oct. 9, 2018, 44 Pages. |
State Intellectual Property Office of the People's Republic of China Notification of Examination Decision for Invalidation Request, Application/Patent No. 201180076332.2; Case No. 4W107724, Feb. 2, 2019, 29 pages. |
Supplementary European Search Report—EP11877053—Search Authority—Munich—dated Sep. 30, 2015 (141997EP). |
First Instance Judgment, Docket No. 141997CN, Patent No. ZL201180076332.2, Oct. 27, 2021, pp. 1-13. |
Number | Date | Country | |
---|---|---|---|
20210304770 A1 | Sep 2021 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 16453779 | Jun 2019 | US |
Child | 17347434 | US | |
Parent | 15400817 | Jan 2017 | US |
Child | 16453779 | US | |
Parent | 14363783 | US | |
Child | 15400817 | US |