Electronic devices continue to make significant contributions to modern society, such as in the realms of safety, transportation, communication, and more, propelling their integration into the daily lives of users. Unfortunately, many users have difficulty tangibly operating electronic devices. For instance, some users are physically handicapped, while other users experience impediments while navigating operating systems of electronic devices due to a lack of familiarity. As a result, such users may experience many legitimate psychological feelings including frustration, inadequacy, and fear of missing out (“FOMO”). Lacking the ability to manage and optimize one's personal property, including electronic devices, in a familiar way can often be upsetting. Moreover, being uninformed of preferred practices when operating these electronic devices can lead to serious consequences, including battery fires and waste.
The present document describes systems and techniques for battery management and optimization using voice integration systems. These techniques include one or more electronic devices defining a voice integration system and having connectivity to a server system. These devices are configured to provide audio data relating to battery operations, including battery charging, to a user of an electronic device. Further, these techniques enable the user of the electronic device to provide additional audio data to the voice integration system and, thereby, instruct the server system to perform one or more actions relating to battery operations associated with the electronic device. In this way, the user can manage and optimize battery operations using the voice integration system.
In some aspects, a method is disclosed that: provides, via a speaker of a voice integration system, an audio inquiry requesting audio input from the user, the audio inquiry relating to one or more charging options for a rechargeable battery of an electronic device; receives, from the user and at a microphone of the voice integration system, the requested audio input, the requested audio input indicating one or more charging including at least a desired charge level for the rechargeable battery of the electronic device; and directs, via a processor or a network communication module associated with the voice integration system and configured to communicating with at least one electronic component of the electronic device or charging unit configured to charging the rechargeable battery of the electronic device, the electronic device or the charging unit configured to charging the rechargeable battery of the electronic device to charge the rechargeable battery of the electronic device to the desired charge level indicated by the requested audio input.
This document also describes computer-readable media having instructions for performing the above-summarized methods and other methods set forth herein, as well as systems and means for performing these methods.
This summary is provided to introduce simplified concepts of battery management and optimization using voice integration systems, which are further described below in the Detailed Description. This summary is not intended to identify essential features of the claimed subject matter, nor is it intended for use in determining the scope of the claimed subject matter.
The details of one or more aspects of battery management and optimization using voice integration systems are described in this document with reference to the following drawings. The same numbers are used throughout the drawings to reference like features and components:
The present document describes systems and techniques for battery management and optimization using voice integration systems. The techniques described herein enable users of electronic devices to procure greater degrees of knowledge, control, and assurance of rechargeable battery operability (e.g., battery state of health, charging and discharging operations).
Many users are unaware of preferred practices to sustain battery health and longevity (e.g., battery state of health) for their electronic devices. As a result, users often unknowingly establish charging habits that accelerate battery deterioration, leading to various types of battery degradation including reduced maximum charge storage capacity, diminished battery efficiency, increased charge times, greater power expenditure, and many others. In some cases, as an example, simply leaving an electronic device to charge for prolonged periods or longer than what is required to achieve a full charge (e.g., overcharging) can negatively affect battery life. In another example, charging an electronic device above a certain threshold charge level (e.g., 70%) or allowing an electronic device to discharge below a certain threshold charge level (e.g., 20%) can introduce battery impairments such as over-potential, gas formation, or accelerated aging.
Due to many users being unaware of these preferred practices, as well as many users lacking some measure of control over a charge duration and a charge rate while charging their electronic devices, electronic devices are often discarded or replaced largely due to their degraded battery. As a consequence, otherwise usable electronic devices are relegated as waste, increasing worldwide waste landfills, and harming the environment. If, however, a user continues to use their electronic device with a degraded battery, the electronic device may more frequently demand charging due to a higher self-discharge rate. Further, more energy may be expended trying to charge the degraded battery of the electronic device. Consequently, an electronic device with a degraded battery may consume more power than an electronic device with an efficient battery. On a large scale, millions of electronic devices having batteries with various degrees of degradation can increase worldwide energy expenditure.
To address these problems, the techniques described herein for battery management and optimization via voice integration systems may, on a large scale, reduce worldwide power expenditure and waste. In more detail, by informing users of battery estimates (e.g., state of health, state of charge) and best-use practices, as well as granting users the ability to manage battery operations via voice integration systems, battery usage and longevity can be optimized. Moreover, providing users convenient methods by which to access, control, or manage a battery of their electronic device, such as voice control, not only facilitates user interaction (e.g., user-input speeds, device-output speeds), but also enables more users to engage with such features. For instance, some users who are visually impaired or are physically handicapped in some fashion may normally have difficulty accessing, controlling, and navigating their electronic devices. As a result, voice integration systems configured to receive and identify voice commands can greatly assist such users.
The following discussion describes operating environments, techniques that may be employed in the operating environments, and example methods.
To provide user access to functions implemented using the wireless network devices 102 in the HAN, a cloud service 112 connects to the HAN via a border router 106, via a secure tunnel 114 through the external network 108 and the access point 110. The cloud service 112 facilitates communication between the HAN and internet clients 116, such as apps on mobile devices, using a web-based application programming interface (API) 118. The cloud service 112 also manages a home graph that describes connections and relationships between the wireless network devices 102, elements of the structure 104, and users. The cloud service 112 hosts controllers which orchestrate and arbitrate home automation experiences, as described in greater detail below.
The HAN may include one or more wireless network devices 102 that function as a hub 120. The hub 120 may be a general-purpose home automation hub, or an application-specific hub, such as a security hub, an energy management hub, a heating, ventilation, and air conditioning (HVAC) hub, and so forth. The functionality of a hub 120 may also be integrated into any wireless network device 102, such as a smart thermostat device or the border router 106. In addition to hosting controllers on the cloud service 112, controllers can be hosted on any hub 120 in the structure 104, such as the border router 106. A controller hosted on the cloud service 112 can be moved dynamically to the hub 120 in the structure 104, such as moving an HVAC zone controller to a newly installed smart thermostat.
Hosting functionality on the hub 120 in the structure 104 can improve reliability when the user's internet connection is unreliable, can reduce latency of operations that would normally have to connect to the cloud service 112, and can satisfy system and regulatory constraints around local access between wireless network devices 102.
The wireless network devices 102 in the HAN may be from a single manufacturer that provides the cloud service 112 as well, or the HAN may include wireless network devices 102 from partners. These partners may also provide partner cloud services 122 that provide services related to their wireless network devices 102 through a partner Web API 124. The partner cloud service 122 may optionally or additionally provide services to internet clients 116 via the web-based API 118, the cloud service 112, and the secure tunnel 114.
The network environment 100 can be implemented on a variety of hosts, such as battery-powered microcontroller-based devices, line-powered devices, and servers that host cloud services. Protocols operating in the wireless network devices 102 and the cloud service 112 provide a number of services that support operations of home automation experiences in a distributed computing environment (e.g., the network environment 100). These services include, but are not limited to, real-time distributed data management and subscriptions, command-and-response control, real-time event notification, historical data logging and preservation, cryptographically controlled security groups, time synchronization, network and service pairing, and software updates.
In the environment 130, any number of the wireless network devices can be implemented for wireless interconnection to wirelessly communicate and interact with each other. The wireless network devices may be modular, intelligent, multi-sensing, network-connected devices that can integrate seamlessly with each other and/or with a central server or a cloud-computing system to provide any of a variety of useful automation objectives and implementations. In an example, a first wireless network device may wirelessly communicate with a second wireless network device to exchange information therebetween. Using this wireless interconnection, the wireless network devices may exchange stored information, including information relating to one or more users, such as radar characteristics, user settings, audio data, and so forth. In addition, these wireless network devices can be communicatively coupled to other electronic devices (e.g., wired speakers, wired microphones). Furthermore, wireless network devices may exchange information regarding operations in progress (e.g., timers, music being played) to preserve a continuity of operations and/or information regarding operations across various rooms in the structure 104. These operations may be performed by one or more wireless network devices simultaneously or independently based on, for instance, the detection of a user's presence in a room. An example of a wireless network device that can be implemented as any of the devices described herein is shown and described with reference to
In implementations, the thermostat 132 may include a Nest® Learning Thermostat that detects ambient climate characteristics (e.g., temperature and/or humidity) and controls an HVAC system 144 in the home environment. The learning thermostat 132 and other network-connected devices “learn” by capturing occupant settings to the devices. For example, the thermostat learns preferred temperature set-points for mornings and evenings, and when the occupants of the structure are asleep or awake, as well as when the occupants are typically away or at home.
A hazard detector 134 can be implemented to detect the presence of a hazardous substance or a substance indicative of a hazardous substance (e.g., smoke, fire, or carbon monoxide). In examples of wireless interconnection, a hazard detector 134 may detect the presence of smoke, indicating a fire in the structure, in which case the hazard detector that first detects the smoke can broadcast a low-power wake-up signal to all of the connected wireless network devices. The other hazard detectors 134 can then receive the broadcast wake-up signal and initiate a high-power state for hazard detection and to receive wireless communications of alert messages. Further, the lighting units 138 can receive the broadcast wake-up signal and activate in the region of the detected hazard to illuminate and identify the problem area. In another example, the lighting units 138 may activate in one illumination color to indicate a problem area or region in the structure, such as for a detected fire or break-in, and activate in a different illumination color to indicate safe regions and/or escape routes out of the structure.
In various configurations, the wireless network devices 140 can include an entryway interface device 146 that functions in coordination with a network-connected door lock system 148, and that detects and responds to a person's approach to or departure from a location, such as an outer door of the structure 104. The entryway interface device 146 can interact with the other wireless network devices based on whether someone has approached or entered the smart home environment. An entryway interface device 146 can control doorbell functionality, announce the approach or departure of a person via audio or visual means, and control settings on a security system, such as to activate or deactivate the security system when occupants come and go. The wireless network devices 140 can also include other sensors and detectors, such as to detect ambient lighting conditions, detect room-occupancy states (e.g., with an occupancy sensor 150), and control a power and/or dim state of one or more lights. In some instances, the sensors and/or detectors may also control a power state or speed of a fan, such as a ceiling fan 152. Further, the sensors and/or detectors may detect occupancy in a room or enclosure and control the supply of power to electrical outlets 154 or devices 140, such as if a room or the structure is unoccupied.
The wireless network devices 140 may also include connected appliances and/or controlled systems 156, such as refrigerators, stoves and ovens, washers, dryers, air conditioners, pool heaters 158, irrigation systems 160, security systems 162, and so forth, as well as other electronic and computing devices, such as televisions, entertainment systems, computers, intercom systems, garage-door openers 164, ceiling fans 152, control panels 166, and the like. When plugged in, an appliance, device, or system can announce itself to the home area network as described above and can be automatically integrated with the controls and devices of the home area network, such as in the home. It should be noted that the wireless network devices 140 may include devices physically located outside of the structure, but within wireless communication range, such as a device controlling a swimming pool heater 158 or an irrigation system 160.
As described above, the HAN includes a border router 106 that interfaces for communication with an external network, outside the HAN. The border router 106 connects to an access point 110, which connects to the external network 108, such as the Internet. A cloud service 112, which is connected via the external network 108, provides services related to and/or using the devices within the HAN. By way of example, the cloud service 112 can include applications for connecting end-user devices 168, such as smartphones, tablets, and the like, to devices in the home area network, processing and presenting data acquired in the HAN to end-users, linking devices in one or more HANs to user accounts of the cloud service 112, provisioning and updating devices in the HAN, and so forth. For example, a user can control the thermostat 132 and other wireless network devices in the home environment using a network-connected computer or portable device, such as a mobile phone or tablet device. Further, the wireless network devices can communicate information to any central server or cloud-computing system via the border router 106 and the access point 110. The data communications can be carried out using any of a variety of custom or standard wireless protocols (e.g., Wi-Fi, ZigBee for low power, 6LoWPAN, Thread, etc.) and/or by using any of a variety of custom or standard wired protocols (CAT6 Ethernet, HomePlug, and so on).
Any of the wireless network devices in the HAN can serve as low-power and communication nodes to create the HAN in the home environment. Individual low-power nodes of the network can regularly send out messages regarding what they are sensing, and the other low-powered nodes in the environment—in addition to sending out their own messages—can repeat the messages, thereby communicating the messages from node to node (e.g., from device to device) throughout the home area network. The wireless network devices can be implemented to conserve power, particularly when battery-powered, utilizing low-powered communication protocols to receive the messages, translate the messages to other communication protocols, and send the translated messages to other nodes and/or to a central server or cloud-computing system. For example, the occupancy sensor 150 and/or an ambient light sensor 170 can detect an occupant in a room as well as measure the ambient light, and activate the light source when the ambient light sensor 170 detects that the room is dark and when the occupancy sensor 150 detects that someone is in the room. Further, the sensor can include a low-power wireless communication chip (e.g., an Institute of Electrical and Electronics Engineers (IEEE) 802.15.4 chip, a Thread chip, a ZigBee chip) that regularly sends out messages regarding the occupancy of the room and the amount of light in the room, including instantaneous messages coincident with the occupancy sensor detecting the presence of a person in the room. As mentioned above, these messages may be sent wirelessly, using the home area network, from node to node (e.g., network-connected device to network-connected device) within the home environment as well as over the Internet to a central server or cloud-computing system.
In other configurations, various ones of the wireless network devices can function as “tripwires” for an alarm system in the home environment. For example, in the event a perpetrator circumvents detection by alarm sensors located at windows, doors, and other entry points of the structure or environment, the alarm could still be triggered by receiving an occupancy, motion, heat, sound, etc. message from one or more of the low-powered mesh nodes in the HAN. In other implementations, the home area network can be used to automatically turn on and off the lighting units 138 as a person transitions from room to room in the structure. For example, the wireless network devices can detect the person's movement through the structure and communicate corresponding messages via the nodes of the HAN. Using the messages that indicate which rooms are occupied, other wireless network devices that receive the messages can activate and/or deactivate accordingly. As referred to above, the home area network can also be utilized to provide exit lighting in the event of an emergency, such as by turning on the appropriate lighting units 138 that lead to a safe exit. The lighting units 138 may also be turned on to indicate the direction along an exit route that a person should travel to safely exit the structure.
The various wireless network devices may also be implemented to integrate and communicate with wearable computing devices 172, such as may be used to identify and locate an occupant of the structure and adjust the temperature, lighting, sound system, and the like accordingly. In other implementations, RFID sensing (e.g., a person having an RFID bracelet, necklace, or key fob), synthetic vision techniques (e.g., video cameras and face recognition processors), audio techniques (e.g., voice, sound pattern, vibration pattern recognition), ultrasound sensing/imaging techniques, and infrared or near-field communication (NFC) techniques (e.g., a person wearing an infrared or NFC-capable smartphone), along with rules-based inference engines or artificial intelligence techniques can draw useful conclusions from the sensed information as to the location of an occupant in the structure or environment.
In other implementations, personal comfort-area networks, personal health-area networks, personal safety-area networks, and/or other such human-facing functionalities of service robots can be enhanced by logical integration with other wireless network devices and sensors in the environment according to rules-based inferencing techniques or artificial intelligence techniques for achieving better performance of these functionalities. In an example relating to a personal health area, the system can detect whether a household pet is moving toward the current location of an occupant (e.g., using any of the wireless network devices and sensors), along with rules-based inferencing and artificial intelligence techniques. Similarly, a hazard detector service robot can be notified that the temperature and humidity levels are rising in a kitchen, and temporarily raise a hazard detection threshold, such as a smoke detection threshold, under an inference that any small increases in ambient smoke levels will most likely be due to cooking activity and not due to a genuinely hazardous condition. Any service robot that is configured for any type of monitoring, detecting, and/or servicing can be implemented as a mesh node device on the home area network, conforming to the wireless interconnection protocols for communicating on the home area network.
The wireless network devices may also include a network-connected alarm clock 174 for each of the individual occupants of the structure in the home environment. For example, an occupant can customize and set an alarm device for a wake time, such as for the next day or week. Artificial intelligence can be used to consider occupant responses to the alarms when they go off and make inferences about preferred sleep patterns over time. An individual occupant can then be tracked in the home area network based on a unique signature of the person, which is determined based on data obtained from sensors located in the wireless network devices, such as sensors that include ultrasonic sensors, passive IR sensors, and the like. The unique signature of an occupant can be based on a combination of patterns of movement, voice, height, size, etc., as well as using facial recognition techniques.
In an example of wireless interconnection, the wake time for an individual can be associated with the thermostat 132 to control the HVAC system in an efficient manner so as to pre- heat or cool the structure to desired sleeping and awake temperature settings. The preferred settings can be learned over time, such as by capturing the temperatures set in the thermostat before the person goes to sleep and upon waking up. Collected data may also include biometric indications of a person, such as breathing patterns, heart rate, movement, etc., from which inferences are made based on this data in combination with data that indicates when the person actually wakes up. Other wireless network devices can use the data to provide other automation objectives, such as adjusting the thermostat 132 so as to pre-heat or cool the environment to a desired setting and turning on or turning off the lighting units 138.
In implementations, the wireless network devices can also be utilized for sound, vibration, and/or motion sensing such as to detect running water and determine inferences about water usage in a home environment based on algorithms and mapping of the water usage and consumption. This can be used to determine a signature or fingerprint of each water source in the home and is also referred to as “audio fingerprinting water usage.” Similarly, the wireless network devices can be utilized to detect the subtle sound, vibration, and/or motion of unwanted pests, such as mice and other rodents, as well as by termites, cockroaches, and other insects. The system can then notify an occupant of the suspected pests in the environment, such as with warning messages to help facilitate early detection and prevention.
The environment 130 may include one or more wireless network devices that function as a hub 176. The hub 176 (e.g., hub 120) may be a general-purpose home automation hub, or an application-specific hub, such as a security hub, an energy management hub, an HVAC hub, and so forth. The functionality of a hub 176 may also be integrated into any wireless network device, such as a network-connected thermostat device or the border router 106. Hosting functionality on the hub 176 in the structure 104 can improve reliability when the user's internet connection is unreliable, can reduce latency of operations that would normally have to connect to the cloud service 112, and can satisfy system and regulatory constraints around local access between wireless network devices.
Additionally, the example environment 130 includes a network-connected speaker 178. The network-connected speaker 178 provides voice assistant services that include providing voice control of network-connected devices. The functions of the hub 176 may be hosted in the network-connected speaker 178. The network-connected speaker 178 can be configured to communicate via the HAN, which may include a wireless mesh network, a Wi-Fi network, or both. In additional examples, other wireless network devices 102, including end-user devices 168, can provide voice assistant services that include providing voice control of network-connected devices.
Some low-power nodes may be unconfigured to bidirectional communication. These low-power nodes send messages but are unable to “listen”. Thus, other devices in the network environment 100, such as the spokesman nodes, cannot send information to these low-power nodes. Some low-power nodes may be configured to only have limited bidirectional communication. As a result of such limited bidirectional communication, other devices may be able to communicate with these low-power nodes only during a certain time period.
As described, in some implementations, the smart devices serve as low-power and spokesman nodes to create a mesh network in the network environment 100. In some implementations, individual low-power nodes in the network environment regularly send out messages regarding what they are sensing, and the other low-powered nodes in the network environment—in addition to sending out their own messages—forward the messages, thereby causing the messages to travel from node to node (e.g., device to device) throughout the HAN 202. In some implementations, the spokesman nodes in the HAN 202, which are able to communicate using a relatively high-power communication protocol (e.g., IEEE 802.11), are able to switch to a relatively low-power communication protocol (e.g., IEEE 802.15.4) to receive these messages, translate the messages to other communication protocols, and send the translated messages to other spokesman nodes and/or the server system 206 (using, e.g., the relatively high-power communication protocol). Thus, the low-powered nodes using low-power communication protocols are able to send and/or receive messages across the entire HAN 202, as well as over the Internet (e.g., network 108) to the server system 206. In some implementations, the mesh network enables the server system 206 to regularly receive data from most or all of the smart devices in the home, make inferences based on the data, facilitate state synchronization across devices within and outside of the HAN 202, and send commands to one or more of the smart devices to perform tasks in the network environment.
As described, the spokesman nodes and some of the low-powered nodes are configured to “listening.” Accordingly, users, other devices, and/or the server system 206 may communicate control commands to the low-powered nodes. The spokesman nodes may use a low-power protocol to communicate the commands to the low-power nodes throughout the HAN 202, as well as to other spokesman nodes that did not receive the commands directly from the server system 206. In another example, a user may use the end-user device 168 (e.g., a smartphone) to send commands over the Internet to the server system 206, which then relays the commands directly back to the end-user device 168. In still further examples, a user may speak voice commands to voice integration system using the end-user device 168 (e.g., a laptop) to send commands over the HAN 202, which then relays the commands to one or more spokesman nodes in the HAN 202.
In some implementations, a lighting unit 138 (
Other examples of low-power nodes include battery-operated versions of the hazard detectors 134. These hazard detectors 134 are often located in an area without access to constant and reliable power and may include any number and type of sensors, such as smoke/fire/heat sensors (e.g., thermal radiation sensors), carbon monoxide/dioxide sensors, occupancy/motion sensors, ambient light sensors, ambient temperature sensors, humidity sensors, and the like. Furthermore, hazard detectors 134 may send messages that correspond to each of the respective sensors to the other devices and/or the server system 206, such as by using the mesh network as described above.
Examples of spokesman nodes include entryway interface devices 146 (e.g., smart doorbells), thermostats 132, control panels 166, electrical outlets 154, charging unit (e.g., docking stations), and other wireless network devices 140. These devices are often located near and connected to a reliable power source, and therefore may include more power-consuming components, such as one or more communication chips configured to bidirectional communication in a variety of protocols.
In some implementations, the network environment 100 includes controlled systems 156, such as service robots, that are configured to carry out, in an autonomous manner, any of a variety of household tasks.
As explained with reference to
As described herein, a voice integration system may include one or more electronic devices, including smart devices 204 or other wireless network devices, and be configured to receiving audio data (e.g., voice commands via one or more microphones), transmitting streams of audio data, or commands therein, (e.g., via a network communication module), and/or providing audio output (e.g., audio data via one or more speakers). For example, a user can speak voice commands to the voice integration system using the end-user device 168 (e.g., a smartphone) to send commands over the Internet to the server system 206, which can then relay the commands to one or more spokesman nodes in the HAN 202. In some implementations, the voice integration system may include the server system 206 and/or one or more electronic devices in the HAN 202 configured to implement the techniques of the server system 206.
In some implementations, the server system 206 receives additional data from one or more smart devices 204 (e.g., metadata, numerical data, etc.). These data may be analyzed to provide context (e.g., actions, a power state, time, location) for audio/video data detected by video cameras 224, network-connected speaker 178, proximity sensors, motion sensors, electrical outlets 154, charging stations, or others. In some implementations, the data indicates where and at what time an audio event (e.g., detected by an audio device such as an audio sensor integrated with the network-connected speaker 178), a security event (e.g., detected by a perimeter monitoring device such as the camera 136 and/or a motion sensor), a hazard event (e.g., detected by the hazard detector 134), medical event (e.g., detected by a health-monitoring device), or the like has occurred within a network environment 100.
In some implementations, each of the audio/video sources 210 captures video or audio and sends the captured audio/video data to the server system 206 substantially in real-time. In some implementations, each of the audio/video sources 210 has its own on-board processing capabilities to perform some preliminary processing on the captured audio/video data before sending the audio/video data (e.g., along with metadata obtained through the preliminary processing) to a controller device and/or the server system 206. In some implementations, one or more of the audio/video sources 210 is configured to, optionally, locally store the video data (e.g., for later transmission if requested by a user). In some implementations, an audio/video source 210 is configured to perform some processing of the captured audio/video data and based on the processing, either send the audio/video data in substantially real-time, store the video data locally, or disregard the audio/video data.
In some implementations, the smart devices 204 communicate with a server-side module 212 executed on the server system 206 through the one or more networks 108. The server-side module 212 provides server-side functionality for data processing for any number of electronic devices, including smart devices 204 (e.g., any one of smart devices 204-1 to 204-p, any one of audio/video sources 210-1 to 210-n). In implementations, the smart devices 204 may also be implemented as audio/video sources 210. Further, one or more of the audio/video sources 210 may be a smart device 204.
In some implementations, the server system 206 includes one or more processors 214, a server database 216, an input/output (I/O) interface 218 to one or more smart devices 204, and an I/O interface 220 to one or more audio/video sources 210. The I/O interface 220 to one or more smart devices 204 facilitates the client-facing input and output processing. The I/O interface 220 to one or more audio/video sources 210 facilitates communications with one or more audio/video sources 210 (e.g., groups of one or more network-connected speakers 178, cameras 136, and associated controller devices). The server database 216 stores raw audio/video data received from the audio/video sources 210, as well as various types of metadata, such as activities, events, and categorization models, for use in data processing.
In some implementations, the server system 206 is implemented on one or more standalone data processing apparatuses or a distributed network of computers. The server system 206 may also employ various virtual devices and/or services of third-party service providers (e.g., third-party cloud service providers) to provide the underlying computing resources and/or infrastructure resources of the server system 206. In some implementations, the server system 206 includes, but is not limited to, a server computer, a handheld computer, a tablet computer, a laptop computer, a desktop computer, or a combination of any two or more of these data processing devices or other data processing devices.
The server-client environment shown in
In some aspects, an audio/video source 210 (e.g., a video camera 136, a network- connected speaker 178) transmits one or more streams 222 of audio/video data to the server system 206. In some implementations, the one or more streams 222 include multiple streams (e.g., 222-1 through 222-q), having respective resolutions and/or rates (e.g., sample rate, frame rate), of the raw audio/video captured by an image sensor and/or microphone. In some implementations, the multiple streams include a “primary” stream (e.g., 222-1) with a certain resolution and rate, corresponding to the raw audio/video captured by the image sensor and/or microphone, and one or more additional streams (e.g., 222-2 through 222-q). An additional stream is optionally the same audio/video stream as the “primary” stream but at a different resolution and/or rate, or a stream that captures a portion of the “primary” stream (e.g., cropped) at the same or different resolution and/or rate as the “primary” stream. In some implementations, the primary stream and/or the additional streams are dynamically encoded (e.g., based on network conditions, server operating conditions, audio/video source operating conditions, characterization of data in the stream (e.g., whether motion is present), user preferences, and the like).
In some implementations, one or more of the streams 222 is sent from the video source 224 directly to a smart device 204 (e.g., without being routed to, or processed by, the server system 206). In some implementations, one or more of the streams 222 is stored at a local memory of the audio/video source 210 and/or at a local storage device (e.g., a dedicated recording device), such as a digital video recorder (DVR). For example, in accordance with some implementations, the network-connected speaker 178 stores the most-recent 24 hours of audio footage recorded by the microphone. In some implementations, portions of the one or more streams 222 are stored at the network-connected speaker 178 and/or the local storage device (e.g., portions corresponding to particular events or times of interest).
In some implementations, the server system 206 transmits one or more streams 224 of audio/video data to a smart device 204 to facilitate voice control. In some implementations, the one or more streams 224 may include multiple streams (e.g., 224-1 through 224-t), of respective resolutions and/or rates, of the same audio/video feed. In some implementations, the multiple streams include a “primary” stream (e.g., 224-1) with a certain resolution and rate, corresponding to the audio/video feed, and one or more additional streams (e.g., 224-2 through 224-t). An additional stream may be the same audio/video stream as the “primary” stream but at a different resolution and/or rate, or a stream that shows a portion of the “primary” stream (e.g., cropped to include a portion of the field of view or pixels of the primary stream) at the same or different resolution and/or rate as the “primary” stream.
Each of the above-identified elements may be stored in one or more of the previously mentioned memory devices and may correspond to a set of instructions for performing a function described above. The above-identified modules or programs (e.g., sets of instructions) need not be implemented as separate software programs, procedures, or modules, and thus various subsets of these modules may be combined or otherwise rearranged in various implementations. In some implementations, the memory 306, optionally, stores a subset of the modules and data structures identified above. Furthermore, the memory 306, optionally, stores additional modules and data structures not described above.
In an example implementation, the data receiving module 316 may receive data from a network-connected speaker (e.g., network-connected speaker 178) and prepare the received data for further processing and storage in the server database. In implementations, the network-connected speaker may include a speaker, a microphone, and a network communication module, and may be configured to receive audio input via a microphone in response to, for example, an audio inquiry provided by the network-connected speaker. The audio input may be converted to a digital signal (“data”) and transmitted via the network communication module to the data receiving module 316. Upon receiving the data, the audio/video processor sub-module 322 may perform voice recognition. For example, the audio/video processor sub-module 322 may analyze and extract speech in the data. Using the extracted speech, optionally in combination with other information determined via the user interface sub-module 324, the entity recognition module 326, the context-manager module 328, and/or data stored in the server database 330, the device control module 318 generate and send server-initiated control commands, based on the extracted speech in the data, to, for example, a smartphone. The server-initiated control commands may direct an operating system, or an application, to perform any of a variety of actions, including enter a low-power mode, initiate a charging operation, cease a charging operation, etc.
The sensor(s) 422 include, for example, one or more thermal radiation sensors, ambient temperature sensors, humidity sensors, infrared (IR) sensors such as passive infrared (PIR) sensors, proximity sensors, range sensors, occupancy sensors (e.g., using radio frequency identification (RFID) sensors), ambient light sensors (ALS), motion sensors 422, location sensors (e.g., GPS sensors), accelerometers, and/or gyroscopes.
In some implementations, the smart device 204 includes an energy storage component 424, including one or more batteries (e.g., a Lithium Ion rechargeable battery). In some implementations, the energy storage component 424 includes a power management integrated circuit (IC). In some implementations, the energy storage component 424 includes circuitry to harvest energy from signals received via an antenna (e.g., the radios 406) of the smart device. In some implementations, the energy storage component 424 includes circuitry to harvest thermal, vibrational, electromagnetic, and/or solar energy received by the smart device 204. In some implementations, the energy storage component 424 includes circuitry to monitor a stored energy level and adjust operation and/or generate notifications based on changes to the stored energy level.
The communication interfaces 404 include, for example, hardware configured to data communications using any of a variety of custom or standard wireless protocols (e.g., IEEE 802.15.4, Wi-Fi, ZigBee, 6LoWPAN, Thread, Z-Wave, Bluetooth Smart, ISA100.5A, WirelessHART, MiWi, etc.) and/or any of a variety of custom or standard wired protocols (e.g., Ethernet, HomePlug, etc.), or any other suitable communication protocol, including communication protocols not yet developed as of the filing date of this document. The radios 406 enable one or more radio communication networks in the network environments 100 and enable a smart device 204 to communicate with other devices. In some implementations, the radios 406 are configured to data communications using any of a variety of custom or standard wireless protocols (e.g., IEEE 802.15.4, Wi-Fi, ZigBee, 6LoWPAN, Thread, Z-Wave, Bluetooth Smart, ISA100.5A, WirelessHART, MiWi, etc.).
The memory 426 includes high-speed random access memory (e.g., DRAM, SRAM, DDR RAM, or other random access solid-state memory devices) and, optionally, includes non-volatile memory (e.g., one or more magnetic disk storage devices, one or more optical disk storage devices, one or more flash memory devices, or one or more other non-volatile solid-state storage devices). The memory 426, or alternatively the non-volatile memory within the memory 426, includes a non-transitory computer-readable storage medium. In some implementations, the memory 426, or the non-transitory computer-readable storage medium of the memory 426, stores the following programs, modules, and data structures, or a subset or superset thereof:
Each of the above-identified elements may be stored in one or more of the previously mentioned memory devices and corresponds to a set of instructions for performing a function described above. The above-identified modules or programs (e.g., sets of instructions) need not be implemented as separate software programs, procedures, or modules, and thus various subsets of these modules may be combined or otherwise rearranged in various implementations. In some implementations, the memory 426, optionally, stores a subset of the modules and data structures identified above. Furthermore, the memory 426, optionally, stores additional modules and data structures not described above, such as a sensor management module for managing operation of the sensor(s) 412. In at least some implementations, at least some techniques of the server-side module 232 may be implemented on or by the battery manager 438.
In some implementations, the voice interface module 440 is, or includes, a voice assistant (e.g., a system-specific voice assistant associated with a particular brand or type of home-automation system or a generic voice assistant that can work with a variety of home-automation systems and devices). The voice assistant may be activated by an activation word or phrase and may be configured to perform tasks as instructed by a user using voice commands.
Examples of a representative smart device 204 include a handheld computer, a wearable computing device, a personal digital assistant (PDA), a tablet computer, a laptop computer, a desktop computer, a cellular telephone, a smartphone, an enhanced general packet radio service (EGPRS) mobile phone, a media player, a navigation device, a game console, an automobile, a television, a remote control, a point-of-sale (POS) terminal, a vehicle-mounted computer, an eBook reader, or a combination of any two or more of these data processing devices or other data processing devices.
Examples of the one or more networks 108 include local area networks (LAN) and wide-area networks (WAN) such as the Internet. The one or more networks 108 are implemented using any known network protocol, including various wired or wireless protocols, such as Ethernet, Universal Serial Bus (USB), FIREWIRE, Long Term Evolution (LTE), Global System for Mobile Communications (GSM), Enhanced Data GSM Environment (EDGE), code division multiple access (CDMA), time division multiple access (TDMA), Bluetooth, Wi-Fi, voice over Internet Protocol (VoIP), Wi-MAX, or any other suitable communication protocol.
The device data 442 is described in more detail in
The calendar 502 of the user may be accessible via the network 108 and may include the user's schedule of events (e.g., appointments, meetings, notifications, announcements, reminders). In aspects, the user's schedule may include information usable to predict battery-related contexts including one or more of: (i) a potential, future battery usage; (ii) an estimated duration of a scheduled event; (iii) which electronic devices a user might utilize or bring during the scheduled event; (iv) a future activity of the user; and (v) a future location of the user. For example, if the calendar 502 indicates that the user is going bird watching, then the server-side module 232 may determine that the user may utilize a digital camera.
Messages, notifications, or other communications sent or received via the user's email messages 504, SMS messages 506, social media account 508, and/or applications 510 associated with the device data 442 may be analyzed to detect whether the user is planning an unscheduled event. For example, if the user purchases airline tickets and receives an email message indicating confirmation of air travel, the context-manager module 328 can use such information to determine an event and predict battery-related contexts. In another example, the user may receive an SMS message 506 from a friend indicating that they will arrive in one hour and want to play basketball. The context- manager module 328 can use such information to determine an event and predict, for example, that a user may use a wearable computing device.
The device settings 512 include information regarding the current settings of the device such as positioning information, mode of operation information, and the like. In some implementations and instances, the device settings 512 are user-specific and are set by respective users of the device.
The associated user information 514 includes information regarding users associated with the device such as users receiving notifications from the device, users registered with the device, users associated with the network environment of the device, and the like.
A voice integration system and the server system 206 can cooperatively operate to perform the method 600. For example, a voice integration system including a smart device 204 (e.g., a smartphone, a hub 176) having a microphone, a speaker, and a network communication module can transmit audio data to and receive commands from the server system 206, and thereby perform the method 600. In another example, a voice integration system including a wireless network device (e.g., a smartphone) having a microphone, as well as a smart device 204 (e.g., a network-connected speaker 178) having a speaker and a network communication module, may be configured to transmit audio data to and receive commands from the server system 206, and thereby perform the method 600.
In at least some implementations, a voice integration system having one or more entities (e.g., modules) of the server system 206 can perform the method 600. For example, a voice integration system including a smart device 204 (e.g., hub 176) can include one or more entities of the server-side module 314, and thereby perform the method 600.
At 602, an audio inquiry requesting audio input from the user is provided via a speaker of the voice integration system. The audio inquiry may relate to one or more charging options for a rechargeable battery of an electronic device. In some implementations, the voice interface module 440 (e.g., a voice assistant) may provide the audio inquiry. The audio inquiry may be provided in response to or based on any of a variety of events (e.g., triggers), including an activity of a user, a direction of travel of the user, a proximity of the user to a charging unit configured to charge the electronic device, a charge level of the rechargeable battery in the electronic device, an intent of the user to charge the rechargeable battery, a schedule of the user indicative of one or more events, event locations, or event durations, etc.
For example, a radar system in one or more smart devices, such as a lightbulb, within a room can sense a proximity of a user for an extended length of time and generate data usable to determine a velocity (e.g., a direction of travel, a speed) of the user. The determined velocity may indicate that the user is traveling towards a charging unit configured to charging a rechargeable battery of a smartphone. The battery manager 438 of the smartphone may, concurrently, determine, or have recently determined, a charge level of the rechargeable battery. Both smart devices (e.g., the lightbulb and the smartphone) may transmit, via their respective network communication module 430, data relating to the user and/or device data 442 to the server system 206. Upon receiving the data relating to the user and/or the device data 442, the server-side module 314 may determine that the user may desire to charge the rechargeable battery of the smartphone after being prompted. As a result, the server system 206 may direct, via the network communication module 312, the smartphone to provide to the user, via an on-device speaker, an audio alert, warning the user that the rechargeable battery of the smartphone is below or near a first threshold charge level (e.g., 15%, 20%, 25%), which may be based on user settings, manufacturer preferred battery operating charges, etc. In further implementations, the threshold charge level may be associated with a battery state of charge at which the rechargeable battery of the smartphone experiences a greater degree (e.g., an increased rate) of degradation due to a high depth (e.g., more than 50%, more than 70%) of discharge. For example, the threshold charge level may be based on a charge level at which the rechargeable battery of the smartphone is expected (e.g., through statistical quality control) to degrade quicker than at other charge levels. In additional implementations, the server system 206 may direct, via the network communication module 312, the smartphone to provide to the user, via an on-device speaker, an audio alert, warning the user that the rechargeable battery of the smartphone is above or near a second threshold charge level (e.g., 75%, 80%, 85%), which may be based on user settings, manufacturer preferred battery operating charges, etc. In further implementations, the threshold charge level may be associated with a battery state of charge at which the rechargeable battery of the smartphone experiences a greater degree (e.g., an increased rate) of degradation due to a low depth (e.g., less than 50%, less than 70%) of discharge. Soon after providing the audio alert, or in response to the user initiating a charging sequence, the server-side module 314 may determine an intent of the user to charge the rechargeable battery of the smartphone and provide the audio inquiry requesting audio input from the user.
At 604, the requested audio input from the user may be received via a microphone of the voice integration system. The requested audio input may indicate (e.g., select) at least one of the one or more charging options. For example, the requested audio input may select a desired charge level for the rechargeable battery of the electronic device. In some implementations, the requested audio input may further select any of a desired charge start time, a desired charge finish time, a desired charge rate, or a desired charge duration.
Continuing with the previous example, the smartphone may receive the requested audio input from the user via a microphone of the smartphone. An electronic device in the HAN 202 (e.g., the smartphone, a hub 176) or the server-side module 314 (e.g., in response to the smartphone transmitting the audio data via the network communication module 430 to the server system 206) may then determine that the requested audio input indicates that the user desires to charge the electronic device to, e.g., 80%).
At 606, the electronic device, or a charging unit configured to charging the rechargeable battery of the electronic device, is directed to charge the rechargeable battery of the electronic device to the desired charge level indicated by the requested audio input. The electronic device or charging unit may be directed via at least one processor and/or a network communication module associated with the voice integration system to charge the rechargeable battery of the electronic device.
In further continuation of the previous example, the smartphone may be directed via one or more processors 402 or a network communication module (e.g., 312, 430) to cease charging after the rechargeable battery charges to, e.g., 80%. The smartphone may cease charging by using one or more on-device electronic circuits configured to gate incoming current. In some implementations, a state of charge of the rechargeable battery corresponding to the desired charge level may be determined and then the smartphone may be charged based on the state of charge corresponding to the desired charge level.
In implementations, the server system 206, or a smart device configured to implement one or more techniques of the server system 206, can receive sensor data from sensors (e.g., a passive infrared (PIR) sensor, an image capture device, a radar unit) associated with one or more devices (e.g., smart devices, electronic devices, wireless network devices) of the HAN 202, as well as device data and/or data relating to a user, and fuse the data (e.g., sensor fusion, data fusion). Using the fused data, the server system 206, or the smart device configured to implement one or more techniques of the server system 206, can, as non-limiting examples, cause one or more devices to be charged and/or suggest a charging operation for one or more devices (e.g., based on a current or future activity of a user, a current battery state of charge, a location of the user, a duration of charging) in substantially real-time. Further, the server system 206, or the smart device configured to implement one or more techniques of the server system 206, can direct one or more devices to facilitate battery discharging (e.g., draining) at any of a variety rates based on the fused data.
In some aspects, one or more electronic devices, including smart devices (e.g., smart devices 204) and/or wireless network devices (e.g., wireless network devices 102), associated with or connected to a HAN (e.g., HAN 202) may be configured to sense and generate data relating to a user, as well as device data (e.g., device data 442). Further, the data relating to the user and/or the device data, may be transmitted between devices of the HAN or a server system (e.g., server system 206) via a network communication module (e.g., network communication module 430, network communication module 312). It should be noted that one or more techniques described as being performed on or by the server system can be implementable on another electronic device associated with the HAN. Based on the data relating to the user and/or the device data, at least one entity (e.g., modules, managers, processors) may determine actions and direct the one or more electronic devices associated with or defining a voice integration system to perform the action, including audibly communicating with the user.
In at least some implementations, the battery manager may direct a voice interface module (e.g., voice interface module 440) to provide an audio alert and an audio inquiry (e.g., step 602). In some cases, the user 702 may have previously defined (e.g., a user preference) a threshold charge level. In additional cases, the user 702 may have previously defined a battery setting configuring the battery manager to direct the processor to transmit information when the charge level of the rechargeable battery is decreasing at a specified rate. In additional implementations, concurrently, or at substantially similar times (e.g., within 2 minutes), electronic devices associated with the HAN, such as a wearable computing device, may transmit data relating to an identity of the user 702 or a location of the user 702 within a structure (e.g., structure 104) to the server system.
Based on at least the information originating from the battery manager, the server system, or other electronic device associated with the HAN, may analyze the information using at least one module and direct at least one electronic device associated with the voice integration system to provide the audio alert and the audio inquiry via an on-device and/or operatively coupled speaker. Additionally, based on an activity of the user 702 and/or a user preference, the server system or other electronic device associated with the HAN may direct the electronic device to perform additional actions, including dimming the display of the smartphone 704, pausing a video for a duration of the audio alert and the audio inquiry, turning off the electronic device, decreasing a volume, entering an operating mode (e.g., a low-power mode), or other actions.
As illustrated, a hub 706 (e.g., hub 176) associated with the voice integration system and the HAN, based on a received instruction from a server system, provides the audio alert and the audio inquiry to the user 702 via an on-device speaker, stating, “Your smartphone is approaching 20% charge capacity. To maintain battery life, it is recommended that the charge be maintained above 20%.” The hub 706, as opposed to other electronic devices, may have been directed to provide the audio alert and the audio inquiry based on one or more conditions, including a user preference, a proximity to a user in relation to other electronic devices, a speaker quality, an authorization, and so on. In this way, the user 702 can be audibly notified of battery best-use operations. In implementations, the audio alert may include information relating to a duration of a charging sequence (e.g., how long an electronic device has been charging).
In some implementations, the server system may also obtain device data from the smartphone 804. The device data may indicate, or be usable to determine, a schedule of a user, detailing future events, activities, locations, and durations. In additional implementations, the server system may further obtain battery usage statistics from the battery manager. The server system, using one or more modules (e.g., context-manager module 328), can analyze the device data to determine a potential, future battery usage. Based on a potential, future battery usage and a charge level of the smartphone 804, the server system can direct the smartphone 804, or any other electronic device associated with the voice integration system, to provide an audio alert. As illustrated in
In some implementations, the hub 908 may be directed to provide an audio message that includes information pertaining to benefits or consequences associated with any of the one or more charging options for the rechargeable battery of the smartphone. For example, the audio message can include information regarding an increased longevity of the rechargeable battery if charged to certain charge levels. In another example, the audio message can include information regarding an increased longevity of the rechargeable battery if charged at certain charging rates.
In response to the audio inquiry, the user may provide a voice command instructing that the smartphone 904 do one of the following: be charged to 100%, cease charging, charge to another desired charge level, or perform another instruction relating to the charging of the rechargeable battery of the smartphone.
In additional implementations, the user may have previously defined, at the smartphone 904, the threshold charge level at which the voice integration system can notify him. In still further implementations, the wireless charging unit 906, an electrical outlet to the wireless charging unit 906 may be coupled to, and/or the smartphone 904 may be smart devices. Any of the smart devices may be configured to determine a charge level of the rechargeable battery of the smartphone, cease the charging, or provide the audio inquiry and audio message.
A microphone of the smartphone 1004, or of any other electronic device associated with the voice integration system, can receive the requested audio input. The signal produced by the microphone, or instructions associated with the requested audio input, may then be passed to the server system. In this way, when the smartphone 1004 is charged by the wireless charging unit 1006 to the requested charge level (e.g., 80%), the server system can direct the smartphone 1004 to cease charging to preserve battery longevity.
In implementations, the server system may direct the wireless charging unit 1006 or the electrical outlet to cease charging. In still further implementations, the server system can implement steps to maintain the charge level of the rechargeable battery at 80%. For example, the server system can, based on a calculated self-discharge rate and/or battery usage statistics for a given battery operating mode, charge the rechargeable battery at a first rate before reaching the desired charge level and then at a second rate configured to maintain the charge level of the rechargeable battery at the desired charge level. In addition, the wireless charging unit and/or smartphone can be configured to indicate a charge completion (e.g., a light color change) when the rechargeable battery of the smartphone has been charged to the desired charge level.
Further illustrated, the user 1102 may provide a voice command requesting, “Please charge my smartphone in a manner that best preserves battery life.” A microphone of one or more electronic devices associated with a voice integration system may receive the voice command and transmit it, or instructions associated with the voice command, to the server system. The server system may instruct the smartphone 1104, the wireless charging unit 1106, and/or an electrical outlet coupled to the wireless charging unit 1106 to charge the smartphone 1104 in a fashion that optimally preserves battery longevity. For example, the server system can instruct the wireless charging unit 1106 to charge the smartphone 1104 at a slow rate, minimizing heat generation, over-potential, or gas formation. In another implementation, the wireless charging unit 1106 may charge the smartphone 1104 at variable rates (e.g., step profile, step charge). For example, between threshold values, the server system may instruct the smartphone 1104, the wireless charging unit 1106, and/or an electrical outlet coupled to the wireless charging unit 1106 to charge the smartphone at a first rate. Outside the threshold values (e.g., above 80%, below 20%), the server system may instruct the smartphone 1104, the wireless charging unit 1106, and/or an electrical outlet coupled to the wireless charging unit 1106 to charge at a second, slower rate.
Further illustrated, the server system may direct the headphones 1204 to provide an audio inquiry asking, “Your laptop is charging at home. Would you like to complete charging by the time you get home?” In implementations, an audio inquiry, such as the one illustrated in
Using the techniques described herein, users of electronic devices can procure greater levels of knowledge concerning rechargeable battery operability using voice control. In addition, control over rechargeable batteries is easily accessible to all users. As a result, rechargeable batteries can be managed and optimized more effectively, assisting in the reduction of worldwide waste and power management. In addition to the above descriptions, electronic device safety as it relates to battery operability can be improved. For example, rechargeable battery degradation-related impairments in response to subpar battery management can be reduced. Although techniques have been described herein relating to rechargeable battery management, the techniques are also extendable to other power systems that may not use rechargeable batteries. For example, systems that connect directly to an electrical outlet or grid may benefit from the techniques described herein by minimizing power expenditure using voice control.
Throughout this document, examples are described where an electronic device or electronic system (e.g., the server system, a smart device, a wireless network device) may analyze information (e.g., calendar data, location data) associated with a user, for example, the device data mentioned with respect to
Although systems and techniques for battery management and optimization via voice integration systems are described, it is to be understood that the subject of the appended Claims is not necessarily limited to the specific features or methods described. Rather, the specific features and methods are disclosed as example implementations and reference is made to the operating environment by way of example only.
In the following section, additional examples are provided.
Example 1: A method comprising: providing, via a speaker of a voice integration system, an audio inquiry requesting audio input from a user, the audio inquiry relating to one or more charging options for a rechargeable battery of an electronic device; receiving, from the user and at a microphone of the voice integration system, the requested audio input, the requested audio input selecting at least one of the one or more charging options; and causing the electronic device to be charged according to the selected one of the one or more charging options.
Example 2: The method as described in example 1, further comprising: determining, prior to providing the audio inquiry, a charge level of the rechargeable battery of the electronic device; and providing, based on the determined charge level of the rechargeable battery, an audio alert via the speaker at the voice integration system, the audio alert pertaining to the charge level for the rechargeable battery.
Example 3: The method as described in any of the previous examples, wherein the determined charge level for the rechargeable battery is near or less than a first threshold charge level, the first threshold charge level being associated with a battery state of charge at which the rechargeable battery experiences a greater degree of degradation due to a high depth of discharge.
Example 4: The method as described in any of the previous examples, wherein the determined charge level of the rechargeable battery is near or greater than a second threshold charge level, the second threshold charge level being associated with a battery state of charge at which the rechargeable battery experiences a greater degree of degradation due to a low depth of discharge.
Example 5: The method as described in any of the previous examples, further comprising determining, prior to providing the audio inquiry requesting audio input from the user, a proximity of the user to a charging unit configured to charging the rechargeable battery of the electronic device, and wherein providing the audio inquiry provides the audio inquiry to the user within the proximity.
Example 6: The method as described in any of the previous examples, wherein the requested audio input is determined to select a desired charge level for the rechargeable battery of the electronic device based on voice recognition.
Example 7: The method as described in any of the previous examples, wherein the requested audio input further selects one or more of a desired charge start time, a desired charge finish time, a selected charge rate, or a desired charge duration.
Example 8: The method as described in any of the previous examples, further comprising: determining a schedule of the user, the schedule indicative of one or more events, event locations, or event durations; and determining a suggested charge level for the rechargeable battery of the electronic device based on the determined schedule of the user, and wherein providing the audio inquiry requesting audio input from the user is based on the suggested charge level for the rechargeable battery of the electronic device.
Example 9: The method as described in example 8, further comprising: determining, based at least in part on the determined schedule of the user, another electronic device associated with the determined scheduled of the user; and determining a suggested battery charge level for the other electronic device based on the determined schedule of the user; and providing another audio inquiry requesting audio input from the user, the other audio inquiry relating to another one or more charging options for another rechargeable battery of the other electronic device.
Example 10: The method as described in any of the previous examples, wherein directing the electronic device or the charging unit directs via a network communication module associated with the voice integration system and configured to communicating with the electronic device or the charging unit configured to charging the rechargeable battery of the electronic device.
Example 11: The method as described in any of the previous examples, wherein directing causes the electronic device or the charging unit configured to charging the rechargeable battery of the electronic device to charge at a variable charge rate, the variable charge rate including a slow charge rate, a medium charge rate, or a quick charge rate.
Example 12: The method as described in any of the previous examples, further comprising providing an audio message, the audio message including information pertaining to benefits or consequences associated with one of the one or more charging options for the rechargeable battery of the electronic device.
Example 13: The method as described in example 12, wherein the audio message further includes information pertaining to a battery state of health or a battery state of charge for the rechargeable battery of the electronic device.
Example 14: The method as described in any of the previous examples, further comprising maintaining, after charging the rechargeable battery of the electronic device to the desired charge level, the charge level based on a self-discharge rate.
Example 15: An electronic device comprising: one or more speaker configured to output audio; one or more microphones configured to capture audio; a network communication module configured to transmit data; and a processor configured to perform the method of any one of examples 1 to 14.
Example 16: The method as described in any of the previous examples, further comprising determining, prior to providing the audio inquiry, an intent to charge the rechargeable battery of the electronic device, the intent to charge based on one or more of an action of the user, a direction of travel of the user, a location of the user, a coupling of the electronic device to a charging unit, or a schedule of the user.
Example 17: The method as described in any of the previous examples, wherein the voice integration system comprises one or more electronic devices.
Example 18: The method as described in any of the previous examples, wherein the one or more electronic devices of the voice integration system each include a microphone or a speaker.
Example 19: The method as described in any of the previous examples, further comprising determining, prior to providing the audio inquiry, an identity of the user, the determination of the identity effective to determine an association with the electronic device.
Example 20: The method as described in any of the previous examples, wherein providing the audio inquiry is based on the determined identity of the user, and wherein providing the audio inquiry is provided to the user associated with the electronic device.
Example 21: The method as described in any of the previous examples, further comprising storing, in a database, the audio inquiry and the requested audio input indicative of one or more charging options for the rechargeable battery of the electronic device.
Example 22: The method as described in any of the previous examples, further comprising determining, in later use, one or more charging options based on the charging information stored in the database.
Example 23: The method as described in any of the previous examples, wherein directing causes the electronic device or the charging unit configured to charging the rechargeable battery of the electronic device to cease charging.
Example 24: The method as described in any of the previous examples, wherein the rechargeable battery is a lithium ion battery.
Example 25: An electronic device comprising: one or more speakers configured to output audio; one or more microphones configured to capture audio; a network communication module configured to transmit data; and a processor configured to perform the method of any one of the aforementioned examples.
Unless context dictates otherwise, use herein of the word “or” may be considered use of an “inclusive or,” or a term that permits inclusion or application of one or more items that are linked by the word “or” (e.g., a phrase “A or B” may be interpreted as permitting just “A,” as permitting just “B,” or as permitting both “A” and “B”). Also, as used herein, a phrase referring to “at least one of” a list of items refers to any combination of those items, including single members. For instance, “at least one of a, b, or c” can cover a, b, c, a-b, a-c, b-c, and a-b-c, as well as any combination with multiples of the same element (e.g., a-a, a-a-a, a-a-b, a-a-c, a-b-b, a-c-c, b-b, b-b-b, b-b-c, c-c, and c-c-c, or any other ordering of a, b, and c). Further, items represented in the accompanying Drawings and terms discussed herein may be indicative of one or more items or terms, and thus reference may be made interchangeably to single or plural forms of the items and terms in this written description.
Although aspects of battery management and optimization using voice integration systems have been described in language specific to features and/or methods, the subject of the appended claims is not necessarily limited to the specific features or methods described. Rather, the specific features and methods are disclosed as example implementations of the techniques for battery management and optimization using voice integration systems, and other equivalent features and methods are intended to be within the scope of the appended claims. Further, various different aspects are described, and it is to be appreciated that each described aspect can be implemented independently or in connection with one or more other described aspects.
Filing Document | Filing Date | Country | Kind |
---|---|---|---|
PCT/US2022/072078 | 5/3/2022 | WO |