Systems and methods for automatic speech recognition

Information

  • Patent Grant
  • 11770649
  • Patent Number
    11,770,649
  • Date Filed
    Tuesday, February 23, 2021
    3 years ago
  • Date Issued
    Tuesday, September 26, 2023
    7 months ago
Abstract
Systems and methods for speech recognition are provided. Some methods can include detecting a relative location of a user within one of a plurality of voice pickup areas of a multi-directional microphone array and the multi-directional microphone array activating a setting of the multi-directional microphone array based on the relative location of the user within the one of the plurality of voice pickup areas.
Description
FIELD

The present invention relates generally to speech recognition systems and methods. More particularly, the present invention relates to systems and methods for activating a direction setting of a multi-directional microphone array in response to a sensor determining a relative location of a user within one of a plurality of voice pickup areas of the multi-directional microphone array.


BACKGROUND

Known speech recognition systems and methods use a multi-directional microphone array operating initially in a directionless setting. The multi-directional microphone array detects a wakeup command from within one of a plurality of voice pickup areas and, in response, activates a direction setting associated with the one of the plurality of voice pickup areas that receives the wakeup command to enhance speech recognition of forthcoming speech coming from the one of the plurality of voice pickup areas. However, such known systems and methods do not activate the direction setting prior to receiving the wakeup command, which results in unintended or missed activation of the multi-directional microphone array caused by interference during operation in the directionless setting.


In view of the above, there is a continuing, ongoing need for improved systems and methods.





BRIEF DESCRIPTION OF THE DRAWINGS


FIG. 1 is a block diagram of a sensor and a multi-directional microphone array in accordance with disclosed embodiments;



FIG. 2 is a block diagram of a sensor and a multi-directional microphone array in accordance with disclosed embodiments;



FIG. 3 is a block diagram of a sensor, a multi-directional microphone array, and a control panel in accordance with disclosed embodiments;



FIG. 4 is a block diagram of a sensor, a multi-directional microphone array, and a control panel in accordance with disclosed embodiments; and



FIG. 5 is a block diagram of a plurality of sensors, a plurality of multi-directional microphone arrays, and a control panel deployed in a region in accordance with disclosed embodiments.





DETAILED DESCRIPTION

While this invention is susceptible of an embodiment in many different forms, there are shown in the drawings and will be described herein in detail specific embodiments thereof with the understanding that the present disclosure is to be considered as an exemplification of the principles of the invention. It is not intended to limit the invention to the specific illustrated embodiments.


Embodiments disclosed herein can include systems and methods for automatic speech recognition. For example, systems and methods disclosed herein can include a sensor detecting a relative location of a user within one of a plurality of voice pickup areas of a multi-directional microphone array and the multi-directional microphone array activating a direction setting of the multi-directional microphone array based on the relative location of the user within the one of the plurality of voice pickup areas, wherein the direction setting enhances speech recognition by the multi-directional microphone array from the relative location of the user within the one of the plurality of voice pickup areas.


Systems and methods as disclosed herein may be deployed as standalone devices interacting together or as a part of home automation systems or security systems. It is to be understood that such systems can include, but are not limited to systems that include a control panel in communication with a sensor or a plurality of sensors and a multi-directional microphone array or a plurality of multi-directional microphone arrays deployed in a region


In accordance with disclosed embodiments, each of a plurality of sensors may be positioned to detect a respective location of the user within a respective one of the plurality of voice pickup areas of the multi-directional microphone array. Accordingly, the sensor may be positioned to detect the relative location of the user within the one of the plurality of voice pickup areas of the multi-directional microphone array. Each of the plurality of voice pickup areas may detect sound or recognize voice commands from a respective direction surrounding the multi-directional microphone array. Accordingly, as the user enters a field of view or detection area of the sensor, the sensor may detect the relative location of the user that corresponds to the one of the plurality of voice pickup areas. In response to the sensor detecting the relative location of the user, the multi-directional microphone array may activate a direction setting associated with the one of the plurality of voice pickup areas. The direction setting may be used to enhance speech recognition by the multi-directional microphone array in the direction of the one of the plurality of voice pickup areas, thereby enhancing the ability of the multi-directional microphone array to receive wakeup commands from the relative location of the user, which can facilitate avoiding or reducing unintended or missed wakeup commands. In some embodiments, the sensor can include a microwave-based location sensor. However, additional types of sensors known in the art for detecting the relative location of the user may come within the spirit and scope of disclosed embodiments.


In some embodiments, when the direction setting is activated, the multi-directional microphone array may detect the wakeup command from the relative location of the user and, in response, may enter an active listening mode to identify additional commands or speech commands coming from the direction of the relative location of the user. In some embodiments, the direction setting may include directing a beamforming function of the multi-directional microphone array toward the relative location of the user within the one of the plurality of voice pickup areas.


In some embodiments, a control panel or similar control unit may be electrically coupled to the sensor and the multi-directional microphone array. The control panel may transfer data indicative of the relative location of the user between the sensor and the multi-directional microphone array using either wired or wireless communication methods. In some embodiments, the sensor can be paired with the multi-directional microphone array to directly communicate the relative location of the user to the multi-directional microphone array. In some embodiments, communication of the relative location of the user from the sensor to the multi-directional microphone array, either directly or through the control panel, can be accomplished using an Internet-based communication protocol.


In some embodiments, a plurality of sensors and a plurality of multi-directional microphone arrays may be placed throughout the region and coupled to the control panel. Each of the plurality of multi-directional microphone arrays may include a respective plurality of voice pickup areas that operate as described herein. The respective plurality of voice pickup areas may detect sound or recognize voice commands from multiple different directions surrounding a respective one of the plurality of multi-directional microphone arrays.


In some embodiments, as the user enters the field of view or the detection area of one of the plurality of sensors, the one of the plurality of sensors may detect the relative location of the user within the region and forward the relative location to the control panel. In response to the one of the plurality of sensors detecting the relative location of the user, the control panel may identify one of the plurality of multi-directional microphone arrays for which one of the respective plurality of voice pickup areas is closest to or within range of the relative location of the user and may send the relative location to the one of the plurality of multi-directional microphone arrays.


In some embodiments, the one of the plurality of multi-directional microphone arrays may activate a respective direction setting associated with the one of the respective plurality of voice pickup areas. The respective direction setting may be used to enhance speech recognition by the one of the plurality of multi-directional microphone arrays in the direction of the one of the respective plurality of voice pickup areas, thereby enhancing the ability of the plurality of multi-directional microphone arrays to receive wakeup commands from the relative location of the user within the region, which can facilitate avoiding or reducing unintended or missed wakeup commands. In some embodiments, the plurality of sensors can include microwave-based location sensors. However, additional types of sensors known in the art for detecting the relative location of the user come within the spirit and scope of disclosed embodiments.


In some embodiments, when the respective direction setting is activated, the one of the plurality of multi-directional microphone arrays may detect wakeup commands from the relative location of the user within the region and, in response, may enter the active listening mode to identify the additional commands or the speech commands from within the one of the respective plurality of voice pickup areas. In some embodiments, the respective direction setting may include directing the beamforming function associated with the one of the respective plurality of voice pickup areas of the one of the plurality of multi-directional microphone arrays toward the relative location of the user within the region.



FIG. 1 and FIG. 2 are block diagrams of a system 20 for automatic speech recognition in accordance with disclosed embodiments. As shown in FIG. 1, the system 20 can include a sensor 22 (e.g. a microwave sensor, PIR sensor, RF sensor, location sensor, etc.) and a multi-directional microphone array 26. The sensor 22 can detect when a relative location of a user U is within a detection area 24 of the sensor 22 corresponding to a specific voice pickup area 28 of the multi-directional microphone array 26 and can communicate the relative location of the user U to the multi-directional microphone array 26. In some embodiments, the specific voice pickup area 28 can be one of a plurality of voice pickup areas of the multi-directional microphone array 26.


As shown in FIG. 2, the multi-directional microphone array 26 can activate a direction setting 30 in response to receiving the relative location of the user U from the sensor 22. The direction setting may correspond to the specific voice pickup area 28 that is associated with the relative location of the user U. In some embodiments, the direction setting can activate a beamforming function of the multi-directional microphone array directed toward the relative location of the user U.


In some embodiments, as shown in FIG. 3 and FIG. 4, the system 20 for automatic speech recognition can include a control panel 32 that can communicate with the sensor 22 and the multi-directional microphone array 26. The control panel 32 can receive first data indicative of the relative location of the user U from the sensor 22 and send second data indicative of the relative location of the user U to the multi-directional microphone array 26.


In some embodiments, as shown in FIG. 5, the system 20 for automatic speech recognition can include a plurality of sensors 22 and a plurality of multi-directional microphone arrays 26 deployed in a region R each of which can communicate with the control panel 32. As the user U enters a field of view detection area of one of the plurality of sensors 22, the one of the plurality of sensors 22 may detect the relative location of the user U within the region R and forward the relative location to the control panel 32. In response to the one of the plurality of sensors 22 detecting the relative location of the user U, the control panel 32 may identify one of the plurality of multi-directional microphone arrays 26 for which one of a respective plurality of voice pickup areas is proximate to the relative location of the user U and may send the relative location to the one of the plurality of multi-directional microphone arrays. Upon receiving the relative location of the user U from the control panel 32, the one of the plurality of multi-directional microphone arrays may operate to activate a direction setting as described herein.


Although a few embodiments have been described in detail above, other modifications are possible. For example, the steps described above do not require the particular order described or sequential order to achieve desirable results. Other steps may be provided, steps may be eliminated from the described flows, and other components may be added to or removed from the described systems. Other embodiments may be within the scope of the invention.


From the foregoing, it will be observed that numerous variations and modifications may be effected without departing from the spirit and scope of the invention. It is to be understood that no limitation with respect to the specific system or method described herein is intended or should be inferred. It is, of course, intended to cover all such modifications as fall within the spirit and scope of the invention.

Claims
  • 1. A security or home automation system comprising: a multi-directional microphone array that includes a plurality of microphones each of which corresponds to a respective one of a plurality of voice pickup areas; andcontrol circuitry in communication with the multi-directional microphone array,wherein the control circuitry is configured to receive information identifying a location of a user relative to the multi-directional microphone array,wherein, after receiving information identifying the location of the user relative to the multi-directional microphone array, the control circuitry is configured to transmit instructions to the multi-directional microphone array to enter a first microphone setting, wherein the first microphone setting comprises activation of a first microphone of the plurality of microphones,wherein, after entering the first microphone setting, the first microphone is configured to detect a wakeup command at the multi-directional microphone array, andwherein, responsive to detecting the wakeup command at the first microphone and while in the first microphone setting, the multi-directional microphone array is configured to enter an active listening mode at the multi-directional microphone array.
  • 2. The security or home automation system of claim 1, wherein the control circuitry is configured to receive a first signal from a first sensor in or proximate to a first area of the plurality of voice pickup areas.
  • 3. The security or home automation system of claim 2, wherein the first signal comprises the information identifying the location of the user relative to the multi-directional microphone array.
  • 4. The security or home automation system of claim 3, wherein the first signal is indicative of the first sensor detecting the user in the first area of the plurality of voice pickup areas.
  • 5. The security or home automation system of claim 1, wherein each of the plurality of microphones is capable of detecting sound from a respective one of a plurality of directions surrounding the multi-directional microphone array.
  • 6. The security or home automation system of claim 1, wherein the wakeup command includes a voice command of the user.
  • 7. The security or home automation system of claim 1, wherein, responsive to detecting the wakeup command at the multi-directional microphone array and while in the first microphone setting, the multi-directional microphone array directs a beamforming function of a first microphone of the plurality of microphones towards a first area of the plurality of voice pickup areas.
  • 8. The security or home automation system of claim 1, wherein the control circuitry uses the information identifying the location of the user relative to the multi-directional microphone array to configure the first microphone setting.
  • 9. The security or home automation system of claim 8, wherein the first microphone setting includes activation of a first microphone of the plurality of microphones corresponding to a first voice pickup area of the plurality of voice pickup areas.
  • 10. A method comprising: associating each of a plurality of microphones of a multi-directional microphone array with a respective one of a plurality of voice pickup areas;communicating with the multi-directional microphone array via control circuitry that receives information identifying a location of a user relative to the multi-directional microphone array;after receiving information identifying the location of the user relative to the multi-directional microphone array, transmitting, via the control circuitry, instructions to the multi-directional microphone array to enter a first microphone setting, wherein entering the first microphone setting comprises activating a first microphone of the plurality of microphones;after entering the first microphone setting, detecting, via the first microphone, a wakeup command at the multi-directional microphone array; andresponsive to detecting the wakeup command at the first microphone and while in the first microphone setting, entering an active listening mode at the multi-directional microphone array.
  • 11. The method of claim 10, further comprising: receiving a first signal from a first sensor in or proximate to a first of the plurality of voice pickup areas.
  • 12. The method of claim 11, wherein the first signal comprises the information identifying the location of the user relative to the multi-directional microphone array.
  • 13. The method of claim 12, wherein the first signal is indicative of the first sensor detecting the user in the first of the plurality of voice pickup areas.
  • 14. The method of claim 10, wherein each of the plurality of microphones is capable of detecting sound from a respective one of a plurality of directions surrounding the multi-directional microphone array.
  • 15. The method of claim 10, wherein the wakeup command includes a voice command of the user.
  • 16. The method of claim 10, wherein, responsive to detecting the wakeup command at the multi-directional microphone array and while in the first microphone setting, directing, via the multi-directional microphone array, a beamforming function of a first microphone of the plurality of microphones towards a first area of the plurality of voice pickup areas.
  • 17. The method of claim 10, wherein the information identifying the location of the user relative to the multi-directional microphone array is used to configure the first microphone setting.
  • 18. The method of claim 17, wherein the first microphone setting includes activation of a first microphone of the plurality of microphones corresponding to a first voice pickup area of the plurality of voice pickup areas.
US Referenced Citations (183)
Number Name Date Kind
4223831 Szarka Sep 1980 A
4253153 Bitterli et al. Feb 1981 A
4340797 Takano et al. Jul 1982 A
4353502 Myers Oct 1982 A
4381549 Stamp, Jr. et al. Apr 1983 A
4598397 Nelson et al. Jul 1986 A
4790143 Hanson Dec 1988 A
5065813 Berkeley et al. Nov 1991 A
5086385 Launey et al. Feb 1992 A
5123252 Hanson Jun 1992 A
5161606 Berkeley et al. Nov 1992 A
5284024 Hanson et al. Feb 1994 A
5345226 Rice et al. Sep 1994 A
5395042 Riley et al. Mar 1995 A
5422822 Toyota et al. Jun 1995 A
5422824 Biehler et al. Jun 1995 A
5481714 Pipkin et al. Jan 1996 A
5553064 Paff et al. Sep 1996 A
5596507 Jones et al. Jan 1997 A
5682329 Seem et al. Oct 1997 A
5690277 Flood Nov 1997 A
5761606 Wolzien Jun 1998 A
5794205 Walters et al. Aug 1998 A
5818438 Howe et al. Oct 1998 A
5892508 Howe et al. Apr 1999 A
5903816 Broadwin et al. May 1999 A
5929850 Broadwin et al. Jul 1999 A
5940118 Van Schyndel Aug 1999 A
5961603 Kunkel et al. Oct 1999 A
5982445 Eyer et al. Nov 1999 A
5995092 Yuen et al. Nov 1999 A
6002394 Schein et al. Dec 1999 A
6005561 Hawkins et al. Dec 1999 A
6006256 Zdepski et al. Dec 1999 A
6009392 Kanevsky et al. Dec 1999 A
6083270 Scott Jul 2000 A
6172677 Stautner et al. Jan 2001 B1
6385510 Hoog et al. May 2002 B1
6413079 Lyons et al. Jul 2002 B1
6477493 Brooks et al. Nov 2002 B1
6477508 Lazar et al. Nov 2002 B1
6526335 Treyz et al. Feb 2003 B1
6584439 Geilhufe et al. Jun 2003 B1
6584613 Dunn et al. Jun 2003 B1
6608560 Abrams Aug 2003 B2
6624750 Marman et al. Sep 2003 B1
6813221 Barr Nov 2004 B1
6988070 Kawasaki et al. Jan 2006 B2
7042349 Bergman et al. May 2006 B2
7108194 Hankins, II Sep 2006 B1
7113090 Saylor et al. Sep 2006 B1
7119658 Stilp Oct 2006 B2
7139716 Gaziz Nov 2006 B1
7146253 Hoog et al. Dec 2006 B2
7162253 Vare et al. Jan 2007 B2
7183907 Simon et al. Feb 2007 B2
7184960 Deisher et al. Feb 2007 B2
7280643 Howard et al. Oct 2007 B2
7349758 Miro et al. Mar 2008 B2
7397363 Joao Jul 2008 B2
7436292 Rourke et al. Oct 2008 B2
7436296 Rourke et al. Oct 2008 B2
7451937 Flood et al. Nov 2008 B2
7454269 Dushane et al. Nov 2008 B1
7464035 Funk et al. Dec 2008 B2
7522063 Ehlers Apr 2009 B2
7634504 Amundson Dec 2009 B2
7675402 Lee et al. Mar 2010 B2
7720683 Vermeulen et al. May 2010 B1
7747446 Blass et al. Jun 2010 B2
7752047 Morris Jul 2010 B2
7890334 Park et al. Feb 2011 B2
7899912 Bisdikian et al. Mar 2011 B2
7957974 Cho et al. Jun 2011 B2
8013730 Oh et al. Sep 2011 B2
8019567 Steinberg et al. Sep 2011 B2
8042048 Wilson et al. Oct 2011 B2
8068881 Schrager Nov 2011 B2
8078472 Resch et al. Dec 2011 B2
8099289 Mozer et al. Jan 2012 B2
8107989 Budampati et al. Jan 2012 B2
8155767 ElMankabady et al. Apr 2012 B2
8175884 Morris May 2012 B1
8234119 Dhawan et al. Jul 2012 B2
8239922 Sullivan et al. Aug 2012 B2
8340975 Rosenberger Dec 2012 B1
8350666 Kore Jan 2013 B2
8411590 Wang Apr 2013 B2
8509954 Imes et al. Aug 2013 B2
8630742 Stefanski et al. Jan 2014 B1
8699944 Malamud et al. Apr 2014 B2
8918219 Sloo et al. Dec 2014 B2
8976937 Shapiro et al. Mar 2015 B2
9043210 Adcock et al. May 2015 B1
9240182 Lee et al. Jan 2016 B2
9354310 Visser et al. May 2016 B2
10030878 Nemcek et al. Jul 2018 B2
10145579 Stoner et al. Dec 2018 B2
20020034956 Mekuria Mar 2002 A1
20020123896 Diez et al. Sep 2002 A1
20020193989 Geilhufe et al. Dec 2002 A1
20030034898 Shamoon et al. Feb 2003 A1
20030036909 Kato Feb 2003 A1
20030088642 Price et al. May 2003 A1
20030117486 Ferren Jun 2003 A1
20030118200 Beaucoup et al. Jun 2003 A1
20030177012 Drennan Sep 2003 A1
20030233432 Davis et al. Dec 2003 A1
20040019489 Funk et al. Jan 2004 A1
20040215750 Stilp Oct 2004 A1
20050002535 Liu et al. Jan 2005 A1
20060161270 Luskin et al. Jul 2006 A1
20060180676 Park et al. Aug 2006 A1
20060282649 Malamud et al. Dec 2006 A1
20070012793 Flood et al. Jan 2007 A1
20070135969 Curl Jun 2007 A1
20070192486 Wilson et al. Aug 2007 A1
20070204228 Minear Aug 2007 A1
20070247301 Browne Oct 2007 A1
20070263600 Sutardja et al. Nov 2007 A1
20070286181 Bushmitch et al. Dec 2007 A1
20080037727 Sivertsen et al. Feb 2008 A1
20080072314 Frenette Mar 2008 A1
20080091432 Dalton et al. Apr 2008 A1
20080150679 Bloomfield Jun 2008 A1
20080175261 Wang Jul 2008 A1
20080221714 Schoettle Sep 2008 A1
20080233983 Park et al. Sep 2008 A1
20090059786 Budampati Mar 2009 A1
20090064295 Budampati Mar 2009 A1
20090086940 Diethorn et al. Apr 2009 A1
20090134993 Ashworth May 2009 A1
20090143918 Amundson et al. Jun 2009 A1
20090204262 Nishimura Aug 2009 A1
20090204410 Mozer et al. Aug 2009 A1
20090323904 Shapiro et al. Dec 2009 A1
20100223055 Mclean Sep 2010 A1
20100332235 David Dec 2010 A1
20110046792 Imes et al. Feb 2011 A1
20110140914 Pelech et al. Jun 2011 A1
20110210816 Wang Sep 2011 A1
20110257973 Chutorash et al. Oct 2011 A1
20110260832 Ross et al. Oct 2011 A1
20110290893 Steinberg Dec 2011 A1
20120021684 Schultz et al. Jan 2012 A1
20120062729 Hart et al. Mar 2012 A1
20120065973 Cho et al. Mar 2012 A1
20120066286 Heredia et al. Mar 2012 A1
20120116748 Kore et al. May 2012 A1
20120123561 Park et al. May 2012 A1
20120136666 Corpier et al. May 2012 A1
20120158161 Cohn et al. Jun 2012 A1
20120327115 Chhetri et al. Dec 2012 A1
20130006400 Caceres et al. Jan 2013 A1
20130035774 Warren et al. Feb 2013 A1
20130077797 Hoy et al. Mar 2013 A1
20130117395 Bushmitch et al. May 2013 A1
20130138250 Mowery et al. May 2013 A1
20130183944 Mozer et al. Jul 2013 A1
20130297078 Kolavennu Nov 2013 A1
20140098233 Martin et al. Apr 2014 A1
20140153281 Zhao et al. Jun 2014 A1
20140153747 Contolini et al. Jun 2014 A1
20140163751 Davis et al. Jun 2014 A1
20140195233 Bapat Jul 2014 A1
20140266687 Britton Sep 2014 A1
20140278394 Bastyr et al. Sep 2014 A1
20140330435 Stoner et al. Nov 2014 A1
20150053779 Adamek et al. Feb 2015 A1
20150053780 Nelson et al. Feb 2015 A1
20150053781 Nelson et al. Feb 2015 A1
20150077553 Dawes Mar 2015 A1
20150187354 Kim Jul 2015 A1
20150276254 Nemcek et al. Oct 2015 A1
20150279134 Warren Oct 2015 A1
20150324179 Nye Nov 2015 A1
20160134632 Cregg May 2016 A1
20160134966 Fitzgerald May 2016 A1
20160249132 Oliaei Aug 2016 A1
20160360526 Lehmann et al. Dec 2016 A1
20170019362 Kim et al. Jan 2017 A1
20180176680 Knight et al. Jun 2018 A1
20180367944 Heo et al. Dec 2018 A1
Foreign Referenced Citations (33)
Number Date Country
201225714 Apr 2009 CN
101947788 Jan 2011 CN
202092667 Dec 2011 CN
102332204 Jan 2012 CN
102436273 May 2012 CN
1054387 Nov 2000 EP
1119191 Jul 2001 EP
1260886 Nov 2002 EP
1345360 Jul 2006 EP
1135757 Nov 2006 EP
1415218 Feb 2011 EP
2801972 Nov 2014 EP
2094508 Sep 1982 GB
2006208460 Aug 2006 JP
2006317573 Nov 2006 JP
2010181064 Aug 2010 JP
2010236759 Oct 2010 JP
201000263563 Mar 2010 KR
20110012048 Feb 2011 KR
20110045314 May 2011 KR
20120017492 Feb 2012 KR
101151571 May 2012 KR
1993013507 Jul 1993 WO
1999034339 Jul 1999 WO
0021053 Apr 2000 WO
2006033760 Mar 2006 WO
2006126192 Nov 2006 WO
2007101164 Sep 2007 WO
2008018102 Feb 2008 WO
2009107211 Sep 2009 WO
2009147927 Dec 2009 WO
2014004911 Jan 2014 WO
2015026933 Feb 2015 WO
Non-Patent Literature Citations (34)
Entry
European Search Report for European Patent Application 21168434.5, dated Sep. 2, 2021, 12 pgs.
Action Talking Products LLC, “Kelvin Installation Manual,” V1.5, pp. 1-15, downloaded Jul. 14, 2015.
Storify, “A Thermostat That Speaks to a Consumer's Needs,” https://storify.com/plumbingtips969/a-thermostat-that-speaks-to-a-consumer-s-needs, printed Jul. 14, 2015, 3 pgs.
SmartWay Solutions, Inc., “Talking Thermostat Model VT3000, User's Guide,” downloaded Jul. 14, 2015, 20 pgs.
Systems Controls & Instruments, LLC, “CEM-24 Series Owner's Manual—Installation and Operating Instructions,” Rev 6.4, downloaded Jul. 14, 2015, 32 pgs.
“T2 RERC Rehabilitation Engineering Research Center on Technology Transfer, Accessible Thermostat,” downloaded Jun. 6, 2013, 40 pgs.
Brown, et al., “Joint DoD/Industry Study on Opportunities in Integrated Diagnostics,” Institute for Defense Analysis, Jan. 1990, 110 pgs.
Carrier Comfort Network, “Carrier Comfort System VVTIII,” Aug. 1996, 6 pgs.
Carvalho et al., “Voice Thermostat Remote Control,” Proceedings of the IEEE 25th Annual Northeast Bioengineering Conference pp. 59-60, Apr. 8-9, 1999.
Coleman, “Hints for Homeowner's, A Thermostat That Speaks to Consumer's Needs,” on or before Jul. 20, 2006, 1 pg.
Honeywell, “PM7006A Network Compatible Computer Room Air Conditioning (CRAC) Program Module for W7600 Control Module,” Aug. 1993, 72 pgs.
“Kelvin Voice Operated Thermostat,” http://www.accendaproducts.com/kelvin/index.html, printed Dec. 9, 2013, 2 pgs.
“Control Your Home by Voice or the Web, Anytime Anywhere,” http://www.automatedliving.com, printed Dec. 9, 2013, 1 pg.
“VoiceIR Environmental Voice Controller Configurator—Broadened Horizons,” http://www.broadenedhorizons.com/voiceir/, Dec. 9, 2013, 17 pgs.
Voice Recognition & Voice Interactive Products, “http://www.innotechsystems.com/voice.htm,” printed Dec. 9, 2013, 1 pg.
Telephone Controlled Thermostat, “http://www.rciautomation.com/thermostat_phone.htm,” printed Dec. 9, 2013, 4 pgs.
http://www.talkingthermostats.com/blind.shtml, printed Dec. 12, 2013, 3 pgs.
“Siri Proxy-NestLearning Thermostat,” https://github.com/chilitechno/SiriProxy-NestLearningthermostat, printed Jan. 6, 2014, 3 pgs.
Jacobson, “Design: A Voice Activated Thermostat,” Biomedical Sciences Instrumentation, Technical Papers Composing the Proceedings of the 29th Annual Rocky Mountain Bioengineering Symposium & 29th International ISA Biomedical Sciences Instrumentation Symposium, vol. 28, 1992, pp. 15-19.
Lacquet et al., “Short Communications, An Affordable Digital-Display-to-Natural-Voice Converter for Visually Impaired Radio Amateurs,” IEEE Transactions on Rehabilitation Engineering, vol. 4, No. No. 4, Dec. 1996, 6 pgs.
Lee et al., “Fault Detection in an Air-Handling Unit Using Residual and Recursive Parameter Identification Methods,” ASHRAE Transactions vol. 102, Pt. 1, pp. 1-22, 1996.
Lopez et al., “Temperature and Humidity Laboratory Remote Controller,” Journal of the Mexican Society of Instrumentation, pp. 14-20, 1996.
Miles et al., “An Audible Thermostat for the Blind or Visually Challenged,” Proceedings of the IEEE 23rd Northeast Bioengineering Conference, pp. 68-69, May 21-22, 1997.
Piette et al., “Model-Based Chiller Energy Tracking for Performance Assurance at a University Building,” Cool Sense National Forum on Integrated Chiller Retrofits, San Francisco, Calif, LBNL Report-40781, Sep. 1997 17 pgs.
“Comfort Solutions for Persons Who are Blind or have Low Vision, VIP 3000 Residential Talking Thermostat that Promotes Independent Living,” Talkingthermostats.com, 2003, 2 pgs.
Venstar, “Comfort Call System ACCO433 Owner's Manual,” Aug. 2007, 24 pgs.
Walters, “Siri Hack Allows Voice Control of Home Thermostat,” found at http://www.extremetech.com/computing/106073-sir-hack-allows-voice-control-of-home..., Nov. 21, 2011, 2 pgs.
Watt, “Development of Empirical Temperature and Humidity-Based Degraded-Condition Indicators for Low-Tonnage Air-Conditioners,” Thesis, Dec. 1997, 205 pgs.
Watt, “Predictive Maintenance Programs (PMP's) in Small HVAC Applications: Analysis of Available Products and Technology,” ESL Report from Paper Prepared for MEEN 662, Apr. 1994, 4 pgs.
“Stargate Interactive Automation System: JDS Technologies,” Web.archive.org/web20030215020919/http:www.idstechnologies.com/stargate.html, printed May 9, 2013, 9 pgs.
Nusca, Andrew: “How Apple's Siri really works,” https://zdnet.com/article/how-apples-siri-really-works/, Nov. 3, 2011, 4 pgs.
Lennox “icomfort Thermostat Manual,” Sep. 2012 edition Retrieved from the internet Jun. 14, 2017 URL:http://airmastersheating.com/Lennox_icomfortWiFi_Manual.pdf, 21 pgs.
2 gig Technologies, “Go Control Wireless Security System, Installation & Programming Instructions,” 2010, 52 pgs.
Extended European Search Report for European Patent Application 18189290.2, dated Nov. 13, 2018, 6 pgs.
Related Publications (1)
Number Date Country
20210185434 A1 Jun 2021 US
Continuations (2)
Number Date Country
Parent 16694078 Nov 2019 US
Child 17182381 US
Parent 15833045 Dec 2017 US
Child 16694078 US