The present description relates generally to acoustic devices including, for example, content sharing using sound-based locations of electronic devices.
Electronic devices often include geolocation circuitry, such as Global Positioning System (GPS) circuitry by which the device can determine its own location.
Certain features of the subject technology are set forth in the appended claims. However, for purpose of explanation, several aspects of the subject technology are set forth in the following figures.
The detailed description set forth below is intended as a description of various configurations of the subject technology and is not intended to represent the only configurations in which the subject technology may be practiced. The appended drawings are incorporated herein and constitute a part of the detailed description. The detailed description includes specific details for the purpose of providing a thorough understanding of the subject technology. However, it will be clear and apparent to those skilled in the art that the subject technology is not limited to the specific details set forth herein and may be practiced without these specific details. In some instances, well-known structures and components are shown in block diagram form in order to avoid obscuring the concepts of the subject technology.
In accordance with aspects of the subject technology, sound-based location detection for electronic devices is provided. As described in further detail hereinafter, in various examples, sound-based location detection of electronic devices can provide for locating a misplaced device using an audio output from the misplaced device, locating a device of a user that is lost or in distress, locating a device having a display for content sharing such as extended display operations, or locating a device for content sharing such as directional transmission of content or other data.
As described in further detail hereinafter, in various implementations, a device that is emitting an audio output can be located by a receiving device using multiple microphones of the receiving device, multiple audio outputs from multiple speakers of the emitting device, using time synchronization information provided along with the audio output, and/or using location information encoded into the audio output.
Illustrative electronic devices are shown in
In the example of
A speaker 114 may be configured to output audio in a human audible range (e.g., audio having one or more frequencies between approximately twenty Hertz and twenty kilohertz (kHz), and/or audio in an ultrasonic frequency range (e.g., audio having one or more frequencies above twenty kHz). For example, a speaker 114 may be operated (e.g., by a processor of the electronic device 100 or a processor of the electronic device 150) to output audio content in the human audible range for audio consumption by a user, such as music, spoken voice (e.g., recorded poetry, or livestreaming audio as part of a telephone call or audio or video conference) podcasts, audio content associated with video content. In one or more implementations, a speaker 114 that is configured to output audio content in the human audible range for audio consumption by a user may be operated (e.g., concurrently or at a different time) to output audio at one or more ultrasonic frequencies. A microphone 116 may be configured to detect audio in a human audible range (e.g., audio having one or more frequencies between approximately twenty Hertz and twenty kilohertz (kHz), and/or audio in an ultrasonic frequency range (e.g., audio having one or more frequencies above twenty kHz). For example, a microphone 116 may be operated (e.g., by a processor of the electronic device 100 or a processor of the electronic device 150) to generate microphone signals responsive to audio input received at the microphone in the human audible range, such as a voice of a person (e.g., a user of the electronic device or another person, such as for operation of a voice call application, a video conferencing application, an audio conferencing application, a recording application, a voice-assistant application, or any other application that operates based on audio inputs to the electronic device that includes the microphone). In one or more implementations, a microphone 116 that is configured to receive audio content in the human audible range may be operated (e.g., concurrently or at a different time) to receive audio (e.g., from another electronic device for locating the other electronic device) at one or more ultrasonic frequencies.
In the example of
As shown in
In various implementations, the housing 106 and/or the display 110 may also include other openings, such as openings for one or more microphones, one or more pressure sensors, one or more light sources, or other components that receive or provide signals from or to the environment external to the housing 106. Openings such as opening 108 and/or opening 112 may be open ports or may be completely or partially covered with a permeable membrane or a mesh structure that allows air and/or sound to pass through the openings. Although two openings (e.g., opening 108 and opening 112) are shown in
The configuration of electronic device 100 of
As shown in
In the example of
As shown in
As illustrated in
For example,
In various use cases, in order (for example) to facilitate distinguishing of the multiple audio outputs from multiple speakers of the electronic device 100, the electronic device 100 may emit the same audio content (e.g., AUDIO1) from the two (or more) speakers at two (or more) different (e.g., predetermined) times (e.g., a first one of the speakers 114 may emit the audio output at a first time and/or a cadence and a second one of the speakers 114 may emit the audio output at a second time or cadence that is offset from the first time or cadence by an offset time that is known to the electronic device 150 and/or a second frequency that is different from the first frequency) or may emit different audio content (e.g., AUDIO1 and AUDIO2, such as different patterns and/or different frequencies of sound) from different speakers 114. The content of the audio output (e.g., the emitted audio content) may include a patterned audio output, such as a series of chirps having predetermined durations, amplitudes, frequencies, and/or spacings between the chirps. Information indicating the predetermined durations, amplitudes, cadences, frequencies, and/or spacings between the chirps may have been previously provided to and/or stored at the electronic device 150 to be used to determine the location of the electronic device 100 from the audio output(s).
As shown in
As shown in
As shown in
In one illustrative example, the electronic device 100 may determine its own location in degrees, minutes, and seconds, of longitude and or latitude. However, since the electronic device 150 (and/or any other electronic devices that may receive the audio output from the speaker 114 of the electronic device 100), will be within an audible range of the audio output at the time that the audio output is received at the electronic device 150 (and/or other electronic devices that receive the audio output), the electronic device 100 may only encode a local portion of the location of the electronic device 100 in the audio output. For example, in one or more use cases, the entire audible range of the audio output from the speaker 114 may be disposed within a region defined by the degrees and minutes of the latitude and longitude of the electronic device 100. Accordingly, it may be unnecessary to provide the degrees and minutes of the latitude and longitude of the electronic device 100 to another electronic device that is located within a region having the same degrees and minutes of latitude and longitude. Accordingly, in one example, the electronic device 100 may encode only the seconds of latitude and longitude in the audio output from the speaker 114. This may improve the efficiency and reduce power and/or computing resource consumption by the encoding electronic device, which can be beneficial particularly, for example, in a battery powered and/or compact device in which power and/or computing resources are limited.
The location information of the electronic device 100 can be encoded into an audio output from a speaker 114 in various ways. For example, the local portion (e.g., the seconds of latitude and longitude) of the location of the electronic device 100 can be translated into Morse code or another coded language that can be encoded into modulations of amplitude, frequency, phase, and/or patterns of sound, and the Morse coded audio (or other coded audio) corresponding to the local portion of the location can be output from the speaker 114.
In some implementations and/or use cases, the electronic device 100 may have other constraints on the output of the speaker 114. For example, in one or more use cases, the output of from the speaker 114 may be designed as an audio distress signal that indicates a user of the electronic device 100 may be lost or in distress. An audio distress signal may have a frequency, a tone, a chirp pattern, or other aspects or features that are designed to be maximally audible to a human listener or another electronic device, and/or may be designed with a frequency or tone that corresponds to one or more resonant frequencies of the electronic device 100 (e.g., to maximize loudness of the output while reducing power usage). Accordingly, in some implementations, it may be desirable to encode some or all of the location of the electronic device 100 into the audio output from the speaker 114, without modifying one or more of the frequency, tone, chirp pattern, or other aspects of the audio output.
In one illustrative example, the audio output from the speaker 114 may consist of a pattern of chirps (e.g., short bursts of sound that are separated from each other in time) that are emitted at frequencies corresponding to one or more resonant frequencies of the electronic device 100. In this illustrative example, it may be undesirable to change the frequencies of the chirps in order to encode the location information. Accordingly, in this illustrative example, the location information may be encoded in the audio output by modifying the times at which the chirps are emitted. That is, in this illustrative example, the electronic device 150 may determine the location of the electronic device 100 by extracting location information for the location of the electronic device 100 from the relative arrival times of the chirps and/or the amounts of time between several of the emitted chirps of the audio output of the speaker 114 of the electronic device 100. In various implementations, the encoded location information may be repeated with multiple repeated chirp patterns, or may be encoded across multiple repeating chirp patterns.
Although the example of
As illustrated in
Whether or not time synchronization information is provided to the receiving device, the emitted chirps 700 of
Whether the location of the electronic device 100 is determined based on an audio output from the electronic device 100 using multiple speaker outputs (see, e.g.,
For example,
Although the visual indicator 802 is shown as an arrow in
In one or more implementations, the visual indicator 802 of
In one or more implementations, the visual indicator 802 of
In this example, in order to determine which portion of the desktop view of the electronic device 100 is displayed at the electronic device 100 and which portion of the desktop view of the electronic device 100 is displayed at the electronic device 150, the electronic device 100 and/or the electronic device 150 determines the location of the electronic device 100 relative to the location of the electronic device 150. In the example of
Accordingly, in this example, the electronic device 150 can determine an angular location of the electronic device 100 relative to the location of the electronic device 150 using the first audio output 900 and the second audio output 902. In one or more implementations, the electronic device 150 may transmit (e.g., using communications circuitry 115) the relative location determined based on the audio output from the electronic device 100 back to the electronic device 100, and the electronic device 100 may provide the display content A, display content B, or display content C to the electronic device 150 for display at the electronic device 150 based on the received relative location. In one or more other implementations, the electronic device 150 may receive the entire desktop view of the electronic device 100, and may determine locally which portion of the desktop view to display at the electronic device 150 (based on the determined location). The electronic device 100 may also adjust the display content A displayed at the electronic device 100 based on the received relative location.
In the example of
In the example of
In one or more implementations, an audio output may include a sound that is generated by one or more speakers of the electronic device. The sound may include audio content. The audio content may include media content, such a song, recorded voice content, voice content of a telephone call or audio or video conference, or other recorded and/or livestreaming media content, and/or may be audio content designed and generated primarily to assist in locating the emitting electronic device. For example, audio content that is designed and generated primarily to assist in locating the emitting electronic device may include a siren sound (e.g., a sound having one or more frequencies that rise and fall over time), a Morse code distress signal, one or more tones that are emitted at frequencies corresponding to one or more resonant frequencies of the electronic device (e.g., to enhance loudness and reduce power consumption), and/or one or more chirps (which may also be referred to as pings, ticks, beeps, blips, or the like). For example, as indicated in
In one or more implementations, the electronic device may receive a user request (e.g., via a touch interface such as a touch-based display or other touch sensor, via a voice input to one or more microphones, such as microphone(s) 116) to locate the other electronic device, or via any other input component), and may provide (e.g., using communication circuitry 115), responsive to the user request, a trigger signal to the other electronic device, the trigger signal including an instruction to emit the audio output. In this example, in one or more use cases, a user that is wearing a smart watch may input a request to the smart watch to locate their smartphone, tablet, or other device, and the smart watch may, responsively, send a trigger signal (e.g., a wireless radio signal) to the smartphone, tablet, or other device to output the audio output. In this example, in one or more other use cases, a user that is operating a smartphone, tablet, or other device may input a request to the smartphone, tablet, or other device to locate their watch, and the smartphone, tablet, or other device may, responsively, send a trigger signal (e.g., a wireless radio signal) to the smart watch to output the audio output. As another example, the audio output may be or include an audio distress sound emitted from the other electronic device and including at least a portion having one or more frequencies between approximately twenty Hertz and approximately twenty kilohertz (e.g., a portion that is audible to a typical human ear). In this other example, in one or more use cases, a user of the other electronic device may provide a user input to the other electronic device to emit the audio distress signal if the user of the other electronic device is lost, stuck, disabled, or otherwise in distress. In this other example, in one or more use cases, the audio distress sound may include a human audible siren sound, a human audible S.O.S. sound, and an ultrasonic locator sound.
At block 1004, the electronic device may determine, based on the audio output, a location of the other electronic device. For example, receiving the audio output may include receiving a first audio output from a first speaker of the other electronic device and receiving a second audio output from a second speaker of the other electronic device, and determining the location of the other electronic device may include determining the location based on a difference (e.g., in amplitude and/or arrival time(s)) between the first audio output and the second audio output. As another example, receiving the audio signal may include receiving a first portion of the audio output with a first microphone of the electronic device and receiving a second portion of the audio output with a second microphone of the electronic device, and determining the location of the other electronic device may include determining the location based on a difference (e.g., in amplitude and/or arrival time(s)) between the first portion of the audio output and the second portion of the audio output. In another example, determining the location may include decoding a portion of the audio output to extract location information for the other electronic device from the audio output; and determining the location of the other electronic device based on the location information (e.g., as described herein in connection with
At block 1006, the electronic device may provide, for display at the electronic device, a visual indicator (e.g., visual indicator 802) of a direction from the electronic device to the location of the other electronic device (e.g., as described herein in connection with
In the example of
At block 1104, the electronic device may generate, with a speaker (e.g., a speaker 114) of the electronic device (and/or one or more additional speakers of the electronic device), an audio output that encodes location information for the location of the electronic device (e.g., as described herein in connection with
In one or more implementations, encoding the location information into the audio output may include adding an additional “chirp” with coded (e.g., in Morse code or other code) GPS information to one or more chirps that are output as a distress signal. For example, an added chirp may be sent entirely separately from chirps and/or other audio patterns designed for detection by a human ear (e.g., a distress audio output, such as an S.O.S. message in Morse code), or can be overlapping with the chirps and/or other audio patterns designed for detection by a human ear. The location information can also, or alternatively, be encoded in the timing between the chirps and/or other audio patterns (e.g., by small variations in the emission times of the chirps, the small variations encoding the location information in a language other than Morse code, in some examples).
The encoded location information may also, or alternatively, include last known GPS location (e.g., in a use case in which a current GPS location cannot be obtained, such as due to lack of access to a GPS signal), a location confidence (e.g., based a quality and/or duration of inertial measurement unit (IMU) data that was used to track since a last known GPS location), and/or an identifier (e.g., a buddy identifier set up in advance with a buddy device of another user, such as by two hikers prior to embarking on a hike, etc.).
In one or more implementations, another electronic device (e.g., the electronic device 150) that receives the audio output including the encoded location information may emit an audio acknowledgement signal. For example, the audio acknowledgement signal may be an echo of the audio output including the encoded location information (or a portion thereof), may be output at a same or different tone as the audio output including the encoded location information, and/or may include different coded information (e.g., including location information of the receiving electronic device). As an example, relative direction information can be encoded into the audio acknowledgement signal (e.g., by increasing or decreasing an amplitude based on orientation of device). In various implementations, encoded location information can be distributed across multiple chirp cycles, and/or can be repeated with each chirp cycle or group of chirp cycles (e.g., so that a receiving device can determine a confidence metric using multiple repeated detections).
In the example of
At block 1204, the electronic device may determine, based on the audio output, a location of the electronic device relative to the other electronic device. In one or more implementations, receiving the audio output may include receiving a first audio output (e.g., a first audio output 900) from a first speaker (e.g., a first speaker 114) of the other electronic device and receiving a second audio output (e.g., a second audio output 902) from a second speaker (e.g., a second speaker 114) at the other electronic device, and determining the location may include determining an angular location of the electronic device relative to the other electronic device based on a difference (e.g., a difference in amplitude and/or arrival time) between the received first audio output and the received second audio output (e.g., as described herein in connection with
In one or more implementations, receiving the audio output may include receiving a first portion (e.g., a first portion 904) of the audio output at a first microphone (e.g., a first microphone 116) of the electronic device and receiving a second portion (e.g., a portion 906) of the audio output at a second microphone (e.g., a second microphone 116) of the electronic device, and determining the location may include determining an angular location of the electronic device relative to the other electronic device based on a difference (e.g., a difference in amplitude and/or arrival time) between the received first portion of the audio output and the received second portion of the audio output (e.g., as described herein in connection with
In one or more implementations, determining the location may include determining the location based on location information that is encoded in the audio output (e.g., as described herein in connection with
At block 1206, the electronic device may receive display content from the other electronic device, the display content based on the location of the electronic device relative to the other electronic device. For example, the display content may include some or all of a desktop view of the other electronic device (e.g., including a desktop background, one or more folder and/or application icons, one or more open user interface windows, etc.).
At block 1208, the electronic device may display the display content at the electronic device. For example, the display content that is displayed the electronic device may be an extension of display content displayed at the other electronic device. For example, the display content may include at least a portion of a desktop view displayed at the other electronic device. Displaying the display content may include displaying a particular portion of the desktop view of the other electronic device that is located in the same direction (relative to another portion of the desktop view that is displayed at the other electronic device) as the direction in which the electronic device is located relative to the other electronic device (e.g., as discussed herein in connection with
In one or more implementations, the process 1200 may also include providing, to the other electronic device, the location of the electronic device relative to the other electronic device. For example, the electronic device (e.g., electronic device 150) may provide the location of the electronic device relative to the other electronic device (e.g., as determined, at block 1204, using the audio output received at block 1202) to the other electronic device (e.g., electronic device 100) in a wireless electromagnetic signal (e.g., a WiFi signal or a Bluetooth signal transmitted using communications circuitry 115). In this way, the other electronic device can determine, in some examples, which portion of its own desktop view to display, and which portion is to be displayed at the electronic device based on the location information received from the electronic device. The other electronic device (e.g., electronic device 100) can then provide, to the electronic device for display at the electronic device, the portion determined, by the other electronic device, to be displayed at the electronic device. In this example, because the determination of which display content is to be displayed at which device is performed by the other electronic device, the electronic device can display, at block 1208, whichever portion of the display content is provided from the other electronic device.
In the example of
As another example, outputting the first audio content with the first speaker of the electronic device may include outputting the first audio content with the first speaker during a first period of time, outputting the second audio content with the second speaker of the electronic device may include outputting the second audio content with the second speaker of the electronic device during the first period of time, and the first audio content may be different from the second audio content. In this way, the other electronic device can distinguish which audio content is being received from which speaker of the electronic device based on the content itself.
In one or more implementations, the electronic device may provide a time synchronization signal for the audio from the electronic device to the other electronic device.
At block 1304, the electronic device may receive, from another electronic device (e.g., electronic device 100) responsive to outputting the audio: display content for display at the electronic device, the display content based on a relative location of the electronic device relative to the other electronic device, the relative location being based on the audio output. For example, the other electronic device may determine, based on the audio output from the electronic device, the location of the electronic device relative to the other electronic device (e.g., as described herein in connection with any of
At block 1306, the electronic device may display the display content at the electronic device at a location that corresponds to a location of additional display content displayed at the other electronic device. For example, displaying the display content may include displaying the received display content corresponding to the portion of the desktop view of the other electronic device that was determined for display at the electronic device. As another example, displaying the display content may include displaying the display content based on the location information (e.g., by determining, at the electronic device, which portion of the received display content to display based on the location information, and displaying that portion of the received display content).
In the example of
At block 1404, the electronic device may determine, based on the audio output, a location of the other electronic device. Determining the location may include determining the location as described in connection with any of
At block 1406, the electronic device may provide content to the other electronic device based on the location of the other electronic device. For example, providing the content to the other electronic device based on the location of the other electronic device may include: identifying the other electronic device as a target for the content based on the location and based on a user gesture corresponding to the location; and providing the content to the other electronic device identified as the target. For example, identifying the other electronic device as the target may include receiving a user input having a direction (e.g., a swipe in the direction, a hand gesture in the direction, an orientation of the electronic device toward the direction, or a motion of the electronic device in the direction), and identifying the other electronic device as the target by determining that the determined location of the other electronic device is in the direction of the user input. The content that is provided based on the location of the other electronic device may include photos, videos, media content, or other data.
As another example, providing the content to the other electronic device based on the location of the other electronic device may include providing display content to the other electronic device based on the location of the electronic device relative to the other electronic device (e.g., as described herein in connection with
In the example of
In one or more implementations, the process 1500 may also include providing a time synchronization signal for the audio output from the electronic device to the other electronic device (e.g., as described herein in connection with
At block 1504, the electronic device may display a first portion (e.g., display content A of
Various examples are described herein in which the electronic device 100 outputs audio that can be used by another electronic device (e.g., the electronic device 150) to locate the electronic device 100 and in which the electronic device 100 is implemented as a computer such as a computer that is integrated into a display such as a computer monitor, a laptop computer, a media player, a gaming device, a navigation device, a computer monitor, a television, a headphone, an earbud, or a wearable device such as a smart watch. In one or more other implementations, the electronic device 100 may be implemented in or as a vehicle, such as a car, a bus, a train, a bicycle, a scooter, or the like that may include one or more speakers and be configured to output audio that can be used by another electronic device to locate the vehicle (e.g., to determine a range and/or angular location of the vehicle relative to the other electronic device) and/or that includes one or more microphones that can be used to determine a location of another electronic device based on audio received from the other electronic device. In one or more implementations, the electronic device 100 may be a personal electronic device that is associated with a particular user account and/or user. In one or more other implementations, the electronic device 100 may be implemented as a public device such as a traffic signal, a crosswalk signal, an alarm or alert device, or any other public device having one or more speakers configured to output audio that can be used by another electronic device to locate the public device (e.g., determine a range and/or angular location of the public device relative to the other electronic device) and/or that includes one or more microphones that can be used to determine a location of another electronic device based on audio received from the other electronic device. For example, such public devices may be distributed throughout a city or a building. As examples, public devices that can provide audio for determining a location of the public device and/or that can determine a location of another device using received audio can be used to provide location-specific safety alerts, evacuation directions, advertising, traffic control, and/or other location-specific services to one or more other electronic devices that are within an audio range of the public devices.
As described above, aspects of the present technology may include the gathering and use of data available from specific and legitimate sources for providing user information in association with processing audio and/or non-audio signals. The present disclosure contemplates that in some instances, this gathered data may include personal information data that uniquely identifies or can be used to identify a specific person. Such personal information data can include voice data, demographic data, location-based data, online identifiers, telephone numbers, email addresses, home addresses, data or records relating to a user's health or level of fitness (e.g., vital signs measurements, medication information, exercise information), date of birth, or any other personal information.
The present disclosure recognizes that the use of such personal information data, in the present technology, can be used to the benefit of users. For example, the personal information data can be used for determining a location of an electronic device. Further, other uses for personal information data that benefit the user are also contemplated by the present disclosure.
For instance, health and fitness data may be used, in accordance with the user's preferences to provide insights into their general wellness, or may be used as positive feedback to individuals using technology to pursue wellness goals.
The present disclosure contemplates that those entities responsible for the collection, analysis, disclosure, transfer, storage, or other use of such personal information data will comply with well-established privacy policies and/or privacy practices. In particular, such entities would be expected to implement and consistently apply privacy practices that are generally recognized as meeting or exceeding industry or governmental requirements for maintaining the privacy of users. Such information regarding the use of personal data should be prominently and easily accessible by users, and should be updated as the collection and/or use of data changes. Personal information from users should be collected for legitimate uses only. Further, such collection/sharing should occur only after receiving the consent of the users or other legitimate basis specified in applicable law. Additionally, such entities should consider taking any needed steps for safeguarding and securing access to such personal information data and ensuring that others with access to the personal information data adhere to their privacy policies and procedures. Further, such entities can subject themselves to evaluation by third parties to certify their adherence to widely accepted privacy policies and practices. In addition, policies and practices should be adapted for the particular types of personal information data being collected and/or accessed and adapted to applicable laws and standards, including jurisdiction-specific considerations which may serve to impose a higher standard. For instance, in the US, collection of or access to certain health data may be governed by federal and/or state laws, such as the Health Insurance Portability and Accountability Act (HIPAA); whereas health data in other countries may be subject to other regulations and policies and should be handled accordingly.
Despite the foregoing, the present disclosure also contemplates embodiments in which users selectively block the use of, or access to, personal information data. That is, the present disclosure contemplates that hardware and/or software elements can be provided to prevent or block access to such personal information data. For example, in the case of determining a location of an electronic device, the present technology can be configured to allow users to select to “opt in” or “opt out” of participation in the collection of personal information data during registration for services or anytime thereafter. In addition to providing “opt in” and “opt out” options, the present disclosure contemplates providing notifications relating to the access or use of personal information. For instance, a user may be notified upon downloading an app that their personal information data will be accessed and then reminded again just before personal information data is accessed by the app.
Moreover, it is the intent of the present disclosure that personal information data should be managed and handled in a way to minimize risks of unintentional or unauthorized access or use. Risk can be minimized by limiting the collection of data and deleting data once it is no longer needed. In addition, and when applicable, including in certain health related applications, data de-identification can be used to protect a user's privacy. De-identification may be facilitated, when appropriate, by removing identifiers, controlling the amount or specificity of data stored (e.g., collecting location data at city level rather than at an address level), controlling how data is stored (e.g., aggregating data across users), and/or other methods such as differential privacy.
Therefore, although the present disclosure broadly covers use of personal information data to implement one or more various disclosed embodiments, the present disclosure also contemplates that the various embodiments can also be implemented without the need for accessing such personal information data. That is, the various embodiments of the present technology are not rendered inoperable due to the lack of all or a portion of such personal information data.
The bus 1608 collectively represents all system, peripheral, and chipset buses that communicatively connect the numerous internal devices of the electronic system 1600. In one or more implementations, the bus 1608 communicatively connects the one or more processing unit(s) 1612 with the ROM 1610, the system memory 1604, and the permanent storage device 1602. From these various memory units, the one or more processing unit(s) 1612 retrieves instructions to execute and data to process in order to execute the processes of the subject disclosure. The one or more processing unit(s) 1612 can be a single processor or a multi-core processor in different implementations.
The ROM 1610 stores static data and instructions that are needed by the one or more processing unit(s) 1612 and other modules of the electronic system 1600. The permanent storage device 1602, on the other hand, may be a read-and-write memory device. The permanent storage device 1602 may be a non-volatile memory unit that stores instructions and data even when the electronic system 1600 is off. In one or more implementations, a mass-storage device (such as a magnetic or optical disk and its corresponding disk drive) may be used as the permanent storage device 1602. ROM 1610, storage 1602, and/or system memory 1604 may store executable code (e.g., executable by the processor(s) 1612 for one or more applications, such as a telephony application, a mail application, a browser application, a media player application, a video conferencing application, a recording application, a messaging application, a calendar application, a fitness application, a mapping application, a payment processing application, a device location application, a word processing application, a presentation application, and/or any other end-user application.
In one or more implementations, a removable storage device (such as a flash drive, and its corresponding disk drive) may be used as the permanent storage device 1602. Like the permanent storage device 1602, the system memory 1604 may be a read-and-write memory device. However, unlike the permanent storage device 1602, the system memory 1604 may be a volatile read-and-write memory, such as random access memory. The system memory 1604 may store any of the instructions and data that one or more processing unit(s) 1612 may need at runtime. In one or more implementations, the processes of the subject disclosure are stored in the system memory 1604, the permanent storage device 1602, and/or the ROM 1610. From these various memory units, the one or more processing unit(s) 1612 retrieves instructions to execute and data to process in order to execute the processes of one or more implementations.
The bus 1608 also connects to the input and output device interfaces 1614 and 1606. The input device interface 1614 enables a user to communicate information and select commands to the electronic system 1600. Input devices that may be used with the input device interface 1614 may include, for example, microphones, alphanumeric keyboards and pointing devices (also called “cursor control devices”). The output device interface 1606 may enable, for example, the display of images generated by electronic system 1600. Output devices that may be used with the output device interface 1606 may include, for example, printers and display devices, such as a liquid crystal display (LCD), a light emitting diode (LED) display, an organic light emitting diode (OLED) display, a flexible display, a flat panel display, a solid state display, a projector, a speaker or speaker module, or any other device for outputting information. One or more implementations may include devices that function as both input and output devices, such as a touchscreen. In these implementations, feedback provided to the user can be any form of sensory feedback, such as visual feedback, auditory feedback, or tactile feedback; and input from the user can be received in any form, including acoustic, speech, or tactile input.
Finally, as shown in
Implementations within the scope of the present disclosure can be partially or entirely realized using a tangible computer-readable storage medium (or multiple tangible computer-readable storage media of one or more types) encoding one or more instructions. The tangible computer-readable storage medium also can be non-transitory in nature.
The computer-readable storage medium can be any storage medium that can be read, written, or otherwise accessed by a general purpose or special purpose computing device, including any processing electronics and/or processing circuitry capable of executing instructions. For example, without limitation, the computer-readable medium can include any volatile semiconductor memory, such as RAM, DRAM, SRAM, T-RAM, Z-RAM, and TTRAM. The computer-readable medium also can include any non-volatile semiconductor memory, such as ROM, PROM, EPROM, EEPROM, NVRAM, flash, nvSRAM, FeRAM, FeTRAM, MRAM, PRAM, CBRAM, SONOS, RRAM, NRAM, racetrack memory, FJG, and Millipede memory.
Further, the computer-readable storage medium can include any non-semiconductor memory, such as optical disk storage, magnetic disk storage, magnetic tape, other magnetic storage devices, or any other medium capable of storing one or more instructions. In one or more implementations, the tangible computer-readable storage medium can be directly coupled to a computing device, while in other implementations, the tangible computer-readable storage medium can be indirectly coupled to a computing device, e.g., via one or more wired connections, one or more wireless connections, or any combination thereof.
Instructions can be directly executable or can be used to develop executable instructions. For example, instructions can be realized as executable or non-executable machine code or as instructions in a high-level language that can be compiled to produce executable or non-executable machine code. Further, instructions also can be realized as or can include data. Computer-executable instructions also can be organized in any format, including routines, subroutines, programs, data structures, objects, modules, applications, applets, functions, etc. As recognized by those of skill in the art, details including, but not limited to, the number, structure, sequence, and organization of instructions can vary significantly without varying the underlying logic, function, processing, and output.
While the above discussion primarily refers to microprocessor or multi-core processors that execute software, one or more implementations are performed by one or more integrated circuits, such as ASICs or FPGAs. In one or more implementations, such integrated circuits execute instructions that are stored on the circuit itself.
Various functions described above can be implemented in digital electronic circuitry, in computer software, firmware or hardware. The techniques can be implemented using one or more computer program products. Programmable processors and computers can be included in or packaged as mobile devices. The processes and logic flows can be performed by one or more programmable processors and by one or more programmable logic circuitry. General and special purpose computing devices and storage devices can be interconnected through communication networks.
Some implementations include electronic components, such as microprocessors, storage and memory that store computer program instructions in a machine-readable or computer-readable medium (alternatively referred to as computer-readable storage media, machine-readable media, or machine-readable storage media). Some examples of such computer-readable media include RAM, ROM, read-only compact discs (CD-ROM), recordable compact discs (CD-R), rewritable compact discs (CD-RW), read-only digital versatile discs (e.g., DVD-ROM, dual-layer DVD-ROM), a variety of recordable/rewritable DVDs (e.g., DVD-RAM, DVD-RW, DVD+RW, etc.), flash memory (e.g., SD cards, mini-SD cards, micro-SD cards, etc.), magnetic and/or solid state hard drives, ultra density optical discs, and any other optical or magnetic media. The computer-readable media can store a computer program that is executable by at least one processing unit and includes sets of instructions for performing various operations. Examples of computer programs or computer code include machine code, such as is produced by a compiler, and files including higher-level code that are executed by a computer, an electronic component, or a microprocessor using an interpreter.
While the above discussion primarily refers to microprocessor or multi-core processors that execute software, some implementations are performed by one or more integrated circuits, such as application specific integrated circuits (ASICs) or field programmable gate arrays (FPGAs). In some implementations, such integrated circuits execute instructions that are stored on the circuit itself.
In accordance with aspects of the subject technology, an electronic device is provided that includes a handheld or wearable housing; one or more microphones disposed in the handheld or wearable housing; a display mounted to the handheld or wearable housing; a memory disposed in the handheld or wearable housing; and one or more processors disposed in the handheld or wearable housing and configured to: receive, using the one or more microphones, an audio output from another electronic device; determine, based on the audio output, a location of the other electronic device; and display, using the display of the electronic device, a visual indicator of a direction from the electronic device to the location of the other electronic device.
In one or more implementations, the one or more processors are further configured to receive, at the electronic device, a user request to locate the other electronic device; and provide, responsive to the user request, a trigger signal to the other electronic device, the trigger signal comprising an instruction to emit the audio output. The audio output may include an audio distress sound emitted from the other electronic device and including at least a portion having one or more frequencies between approximately twenty Hertz and approximately twenty kilohertz. The one or more processors may also be configured to: receive the audio output by receiving a first audio output from a first speaker of the other electronic device and receiving a second audio output from a second speaker of the other electronic device, and determine the location of the other electronic device based on a difference between the first audio output and the second audio output. The one or more processors may also be configured to: receive the audio output by receiving a first portion of the audio output with a first microphone of the electronic device and receiving a second portion of the audio output with a second microphone of the electronic device, and determine the location of the other electronic device based on a difference between the first portion of the audio output and the second portion of the audio output. The one or more processors may be configured to determine the location by: decoding a portion of the audio output to extract location information for the other electronic device from the audio output; and determining the location of the other electronic device based on the location information. The one or more processors may also be configured to: receive time synchronization information from the other electronic device; and determine the location of the other electronic device based on the audio output and the time synchronization information. The one or more processors may also be configured to: receive the audio output while operating the electronic device in a low power mode of operation; and responsive to receiving the audio output, switch the electronic device to a higher power mode of operation for determining the location of the other electronic device.
In accordance with other aspects of the subject technology, an electronic device is provided that includes a speaker; a memory; and one or more processors configured to: obtain a location of the electronic device; and generate, with the speaker, an audio output that encodes location information for the location of the electronic device. The one or more processors may also be configured to generate the audio output that encodes the location information by modulating emission times of multiple portions of the audio output. The multiple portions of the audio output may be emitted with one or more frequencies that are determined based on a resonance feature of the electronic device. The location information may include a local portion the location of the electronic device. The location may include the local portion and a regional portion that is omitted from the location information that is encoded in the audio output. An audio range of the audio output may be within the regional portion of the location of the electronic device.
In accordance with other aspects of the subject technology, a method is provided that includes obtaining, by an electronic device, a location of the electronic device; and generating, with a speaker of the electronic device, an audio output that encodes location information for the location of the electronic device. Generating the audio output that encodes the location information may include modulating emission times of multiple portions of the audio output. The multiple portions of the audio output may be emitted with one or more frequencies that are determined based on a resonance feature of the electronic device. The location information may include a local portion the location of the electronic device. The location may include the local portion and a regional portion that is omitted from the location information that is encoded in the audio output. An audio range of the audio output may be within the regional portion of the location of the electronic device.
As used in this specification and any claims of this application, the terms “computer”, “processor”, and “memory” all refer to electronic or other technological devices. These terms exclude people or groups of people. For the purposes of the specification, the terms “display” or “displaying” means displaying on an electronic device. As used in this specification and any claims of this application, the terms “computer readable medium” and “computer readable media” are entirely restricted to tangible, physical objects that store information in a form that is readable by a computer. These terms exclude any wireless signals, wired download signals, and any other ephemeral signals.
Many of the above-described features and applications are implemented as software processes that are specified as a set of instructions recorded on a computer readable storage medium (also referred to as computer readable medium). When these instructions are executed by one or more processing unit(s) (e.g., one or more processors, cores of processors, or other processing units), they cause the processing unit(s) to perform the actions indicated in the instructions. Examples of computer readable media include, but are not limited to, CD-ROMs, flash drives, RAM chips, hard drives, EPROMs, etc. The computer readable media does not include carrier waves and electronic signals passing wirelessly or over wired connections.
In this specification, the term “software” is meant to include firmware residing in read-only memory or applications stored in magnetic storage, which can be read into memory for processing by a processor. Also, in some implementations, multiple software aspects of the subject disclosure can be implemented as sub-parts of a larger program while remaining distinct software aspects of the subject disclosure. In some implementations, multiple software aspects can also be implemented as separate programs. Finally, any combination of separate programs that together implement a software aspect described here is within the scope of the subject disclosure. In some implementations, the software programs, when installed to operate on one or more electronic systems, define one or more specific machine implementations that execute and perform the operations of the software programs.
A computer program (also known as a program, software, software application, script, or code) can be written in any form of programming language, including compiled or interpreted languages, declarative or procedural languages, and it can be deployed in any form, including as a standalone program or as a module, component, subroutine, object, or other unit suitable for use in a computing environment. A computer program may, but need not, correspond to a file in a file system. A program can be stored in a portion of a file that holds other programs or data (e.g., one or more scripts stored in a markup language document), in a single file dedicated to the program in question, or in multiple coordinated files (e.g., files that store one or more modules, sub programs, or portions of code). A computer program can be deployed to be executed on one computer or on multiple computers that are located at one site or distributed across multiple sites and interconnected by a communication network.
It is understood that any specific order or hierarchy of blocks in the processes disclosed is an illustration of example approaches. Based upon design preferences, it is understood that the specific order or hierarchy of blocks in the processes may be rearranged, or that all illustrated blocks be performed. Some of the blocks may be performed simultaneously. For example, in certain circumstances, multitasking and parallel processing may be advantageous. Moreover, the separation of various system components in the implementations described above should not be understood as requiring such separation in all implementations, and it should be understood that the described program components and systems can generally be integrated together in a single software product or packaged into multiple software products.
The previous description is provided to enable any person skilled in the art to practice the various aspects described herein. Various modifications to these aspects will be readily apparent to those skilled in the art, and the generic principles defined herein may be applied to other aspects. Thus, the claims are not intended to be limited to the aspects shown herein, but are to be accorded the full scope consistent with the language claims, wherein reference to an element in the singular is not intended to mean “one and only one” unless specifically so stated, but rather “one or more.” Unless specifically stated otherwise, the term “some” refers to one or more. Pronouns in the masculine (e.g., his) include the feminine and neuter gender (e.g., her and its) and vice versa. Headings and subheadings, if any, are used for convenience only and do not limit the subject disclosure.
The predicate words “configured to”, “operable to”, and “programmed to” do not imply any particular tangible or intangible modification of a subject, but, rather, are intended to be used interchangeably. For example, a processor configured to monitor and control an operation or a component may also mean the processor being programmed to monitor and control the operation or the processor being operable to monitor and control the operation. Likewise, a processor configured to execute code can be construed as a processor programmed to execute code or operable to execute code
A phrase such as an “aspect” does not imply that such aspect is essential to the subject technology or that such aspect applies to all configurations of the subject technology. A disclosure relating to an aspect may apply to all configurations, or one or more configurations. A phrase such as an aspect may refer to one or more aspects and vice versa. A phrase such as a “configuration” does not imply that such configuration is essential to the subject technology or that such configuration applies to all configurations of the subject technology. A disclosure relating to a configuration may apply to all configurations, or one or more configurations. A phrase such as a configuration may refer to one or more configurations and vice versa.
The word “example” is used herein to mean “serving as an example or illustration.” Any aspect or design described herein as “example” is not necessarily to be construed as preferred or advantageous over other aspects or design.
In one aspect, a term coupled or the like may refer to being directly coupled. In another aspect, a term coupled or the like may refer to being indirectly coupled.
Terms such as top, bottom, front, rear, side, horizontal, vertical, and the like refer to an arbitrary frame of reference, rather than to the ordinary gravitational frame of reference. Thus, such a term may extend upwardly, downwardly, diagonally, or horizontally in a gravitational frame of reference.
All structural and functional equivalents to the elements of the various aspects described throughout this disclosure that are known or later come to be known to those of ordinary skill in the art are expressly incorporated herein by reference and are intended to be encompassed by the claims. Moreover, nothing disclosed herein is intended to be dedicated to the public regardless of whether such disclosure is explicitly recited in the claims. No claim element is to be construed under the provisions of 35 U.S.C. § 112(f), unless the element is expressly recited using the phrase “means for” or, in the case of a method claim, the element is recited using the phrase “step for.” Furthermore, to the extent that the term “include,” “have,” or the like is used in the description or the claims, such term is intended to be inclusive in a manner similar to the term “comprise” as “comprise” is interpreted when employed as a transitional word in a claim.
This application claims the benefit of priority to U.S. Provisional Patent Application No. 63/435,218, entitled, “Content Sharing Using Sound-Based Locations of Electronic Devices”, filed on Dec. 23, 2022, the disclosure of which is hereby incorporated herein in its entirety.
Number | Date | Country | |
---|---|---|---|
63435218 | Dec 2022 | US |