Aspects related to audio systems, are disclosed. More particularly, aspects related to audio systems used to render spatial audio, are disclosed.
Spatial audio can be rendered using headphones that are worn by a user. For example, the headphones can reproduce a spatial audio signal that simulates a soundscape around the user. An effective binaural audio reproduction can convince the user that the user is not wearing headphones. More particularly, sounds are ideally rendered such that the user perceives sound sources within the soundscape external to the user's head, just as the user would experience the sounds if encountered in the real world.
When a sound travels to a listener from a surrounding environment, the sound propagates along a direct path, e.g., through air to the listeners ear canal entrance, and along one or more indirect paths, e.g., by reflecting and diffracting around the listeners head or shoulders. As the sound travels along the indirect paths, artifacts can be introduced into the acoustic signal that the ear canal entrance receives. These artifacts are anatomy dependent, and accordingly, are user-specific. The user therefore perceives the artifacts as natural. Accordingly, the artifacts can be reproduced when rendering a soundscape to provide an accurate binaural reproduction to the user.
User-specific artifacts can be incorporated into binaural audio by signal processing algorithms that use spatial audio filters. The audio filters can be applied to an input signal to shape a frequency response in a way that simulates a sound traveling to the user from a surrounding environment. For example, a head-related transfer function (HRTF) contains all of the acoustic information required to describe how sound reflects or diffracts around a listener's head before entering their auditory system. Similarly, a headphone equalization (HpEQ) filter contains acoustic information to compensate for effects of a listener's outer ear, e.g., a pinna of the listener, on sound generated by a headphone driver worn by the listener. The HRTF and HpEQ filter can be encapsulated in respective datasets. A listener can use simple stereo headphones to create the illusion of a sound source somewhere in a listening environment by applying the HRTF and the HpEQ filter to a binaural recording of the sound source.
Existing binaural reproduction methods include choosing a head-related transfer function (HRTF) or a headphone equalization (HpEQ) filter for a listener based on similarities between the listener and one or more other people. For example, an HpEQ filter may be integrated in circuitry of a pair of headphones to shape a frequency response toward a general target for all listeners. In reality, HRTFs and HpEQ filters are highly individualized, however, and binaural simulation using general, non-individualized, HRTFs or HpEQ filters (for example when a listener auditions a simulation using the audio filter dataset of another person or a population) can cause audible problems in both the perceived position and quality (timbre) of the virtual sound. As such, an HRTF or an HpEQ filter that effectively simulates a sound source at a location relative to a first user may not effectively simulate the sound source at the same relative location to a second user. That is, the first user may experience the simulation as a realistic rendering, but the second user may not.
An audio system and a method of using the audio system to determine one or more of an HRTF or an HpEQ filter, which are personalized for a user, are described. By applying a personalized HRTF and HpEQ filter of the user to an input audio signal, a user-specific spatial audio signal can be generated and played for the user. When reproduced, the spatial audio can provide an accurate binaural reproduction to the user.
The method of using the audio system to determine one or more of an HRTF or an HpEQ filter can include driving a speaker of a headphone, e.g., an earcup or an earbud of a pair of headphones, with a first input signal while a user is wearing the headphones. The speaker can generate a sound corresponding to the input signal, and a microphone of the headphone can detect the sound after the sound reflects from a pinna of the user. The sound is detected as an output signal, and the audio system can determine an acoustic transfer function that relates the output signal detected by the microphone to the input signal played by the speaker. More particularly, the acoustic transfer function can correspond to an impulse response of the headphone and pinna system, and accordingly, the acoustic transfer function can be specific to the user. The audio system can use the user-specific acoustic transfer function to determine one or more audio filters for the headphone. For example, a head-related transfer function (HRTF) or a headphone equalization (HpEQ) filter can be determined for the user based on the acoustic transfer function. The audio filters can be generated or selected. For example, the acoustic transfer function can be inverted to generate the HpEQ filter. Alternatively, the acoustic transfer function can be matched to a predetermined transfer function of others, and an HRTF or HpEQ corresponding to the matching transfer function of others can be selected as an audio filter for the user.
The audio filter(s) that are determined based on the user-specific acoustic transfer function, e.g., one or more of the HRTF or the HpEQ, can be applied to a second input signal. The second input signal may be a different portion of a user content signal than the first input signal. Application of the audio filters to the second input signal can generate a spatial input signal. The spatial input signal can be specific to the user. Accordingly, the speaker can be driven with the spatial input signal to generate a spatialized sound that realistically simulates an externalized sound source in a soundscape around the user.
The above summary does not include an exhaustive list of all aspects of the present invention. It is contemplated that the invention includes all systems and methods that can be practiced from all suitable combinations of the various aspects summarized above, as well as those disclosed in the Detailed Description below and particularly pointed out in the claims filed with the application. Such combinations have particular advantages not specifically recited in the above summary.
Aspects describe an audio system and a method of using the audio system to determine one or more of a head-related transfer function (HRTF) or a headphone equalization (HpEQ) filter that is personalized for a user. The audio system can incorporate a pair of headphones, and each headphone can have a speaker and a microphone. For example, the headphones can be circumaural headphones having the microphone in an interior of an earcup. The audio system may, however, include other devices for rendering audio to the user, including: other types of headphones such as earbuds or, a headset; or another head-mounted consumer electronics product, such as a motorcycle helmet, to name only a few possible applications.
In various aspects, description is made with reference to the figures. However, certain aspects may be practiced without one or more of these specific details, or in combination with other known methods and configurations. In the following description, numerous specific details are set forth, such as specific configurations, dimensions, and processes, in order to provide a thorough understanding of the aspects. In other instances, well-known processes and manufacturing techniques have not been described in particular detail in order to not unnecessarily obscure the description. Reference throughout this specification to “one aspect,” “an aspect,” or the like, means that a particular feature, structure, configuration, or characteristic described is included in at least one aspect. Thus, the appearance of the phrase “one aspect,” “an aspect,” or the like, in various places throughout this specification are not necessarily referring to the same aspect. Furthermore, the particular features, structures, configurations, or characteristics may be combined in any suitable manner in one or more aspects.
The use of relative terms throughout the description may denote a relative position or direction. For example, “in front of” may indicate a first direction away from a reference point. Similarly, “behind” may indicate a location in a second direction away from the reference point and opposite to the first direction. Such terms are provided to establish relative frames of reference, however, and are not intended to limit the use or orientation of an audio system to a specific configuration described in the various aspects below.
In an aspect, an audio system includes headphones having internally and/or externally mounted microphone(s) that are used to personalize an HRTF and/or HpEQ filter for a user. By driving a speaker of the headphones with an input signal, a sound can be generated that reflects from a pinna of a user wearing the headphones. The microphone(s) can detect the reflected sound as an output signal that correlates to the input signal. The audio system can determine an acoustic transfer function that relates the output signal to the input signal. Given that the acoustic transfer function is specific to an anatomy of the user, e.g., the pinna shape, the acoustic transfer function is user-specific. The user-specific acoustic transfer function can be used by the audio system to determine one or more spatial audio filters, such as an HRTF or an HpEQ filter, for the user. For example, the spatial audio filters can be selected or constructed by the audio system based on the user-specific acoustic transfer function. Accordingly, the personalized HRTF and HpEQ filter can be applied to the input signal to generate a spatial input signal that accurately renders spatial audio to the user.
Referring to
Audio system 102 can include one or more microphones 106. Microphone(s) 106 may be built into headphones 104 to detect sounds internal to and/or external to the earphone. For example, headphones 104 may be circumaural headphones having a pair of earcups 108 (or earbuds having a first earbud and a second earbud). One or more microphones 106 can be mounted on each earcup 108 facing a surrounding environment. Similarly, one or more microphones 106 may be mounted within an interior of earcup 108, e.g., on an internal surface of earcup 108. Microphone(s) 106 can generate microphone output signals based on the detected sounds. For example, microphones 106 on an exterior of earcup 108 can generate microphone output signals corresponding to sounds traveling toward user 100 from a surrounding environment. Similarly, microphones 106 contained within an interior of earcup 108 can generate microphone output signals corresponding to sounds traveling within the interior between earcup 108 and an ear of user 100. Microphone output signals generated by microphones 106 can be used to model an acoustic path from a sound source, such as a voice, to microphones 106. For example, such modeling can be used in echo-cancellation algorithms for active noise canceling applications. As described below, the information from microphones 106 can also be used to determine acoustic transfer functions specific to user 100, which can inform a selection or construction of audio filters for audio rendering.
In an aspect, audio system 102 includes a device 110, such as a mobile device, laptop, home stereo, etc. Device 110 can include circuitry to perform the functions described below. For example, device 110 can generate or transmit an input signal that is played by headphones 104. Furthermore, device 110 can receive signals from headphones 104, such as the microphone output signals from microphones 106, and use the signals to determine user-specific acoustic transfer functions and/or personalized audio filters for user 100. Accordingly, device 110 and headphones 104 can be connected wirelessly or by a wired connection to communicate signals used for audio rendering, e.g., binaural audio reproduction.
Referring to
Device processor(s) 202 can retrieve audio data 206 from device memory 204. Audio data 206 may be associated with one or more audio sources 207, including phone and/or music playback functions controlled by telephony or music application programs that run on top of the operating system. Similarly, audio data 206 may be associated with an augmented reality (AR) or virtual reality (VR) application program that runs on top of the operating system. The audio sources 207 can output user content signals 208 for playback by headphones 104.
In an aspect, device memory 204 stores audio filter data. For example, device memory 204 can store a transfer function database 210. Transfer function database 210 can include a dataset of generic or individualized transfer functions or filters, such as HRTFs or HpEQ filters. For example, a dataset of HRTFs can include several HRTFs that correspond to specific locations relative to user 100. A single HRTF of the dataset can be a pair of acoustic filters (one for each ear) that characterize the acoustic transmission from the particular location in a reflection-free environment to an entrance to an ear canal of user 100. The dataset of HRTFs encapsulate the fundamentals of spatial hearing of user 100. The dataset can include an HRTF filter that corresponds to an acoustic transfer function that is specific to user 100. Similarly, a dataset of HpEQ filters can include several HpEQ filters that correspond to acoustic transfer functions of one or more users. The acoustic transfer functions can correspond to anatomical characteristics of the user(s). The dataset can include an HpEQ filter that corresponds to an acoustic transfer function that is specific to user 100. Accordingly, determination of the user-specific acoustic transfer function can be used by device processor 202 to determine or select one or more of an HRTF or an HpEQ filter personalized for user 100.
To perform the various functions, device processor(s) 202 may directly or indirectly implement control loops and receive input signals from, and/or provide output signals to, other electronic components. For example, device 110 may receive input signals from microphone(s) or menu buttons of device 110, including through input selections of user interface elements displayed on a display 212. Device 110 can communicate system signals, such as user content signal 208, to headphone 104. More particularly, device 110 and headphones 104 can communicate wirelessly via respective RF circuitry, or through a wired connection.
Headphones 104 can include one or more headphone, e.g., a first headphone 220A and a second headphone 220B. Each headphone (earphone) 220, can be physically connected by a headband, a neck cord, or another physical connector (shown in phantom). Each earphone 220 may include a headphone processor 224 to perform one or more of the various functions described below. For example, headphone processor 224 can communicate with a headphone memory 226, which stores audio data 206, e.g., a cached portion of user content signal 208 received from device 110, an HRTF filter, and/or an HpEQ filter for a respective earphone 220. Headphone processor 224 can apply the HRTF filter and the HpEQ filter to the cached portion when rendering binaural playback to user 100 through the respective headphones. In an aspect, all functionality of system 102 can be performed by the components in headphones 104.
Each headphone 220 of headphones 104 can include a speaker 228 to output a sound 230 to user 100. More particularly, speakers 228 can receive an input signal 232 from device processor 202 and/or headphone processor 224. Input signal 232 can be a portion of user content signal 208. Input signal 232 can drive speaker 228 to generate sound 230, and emit sound 230 toward the ears of user 100. The ears may be contained within earcup 108 when user 100 is wearing circumaural headphones. At least a portion of sound 230 can reflect or diffract from a pinna of the ear (as represented by the curving path) and the reflected/diffracted sound 230 may be detected by microphone 106 mounted on headphones 104. For example, microphones 106 can detect the reflected sound 230 as an output signal 234. Output signal 234 can be a pressure at microphone 106, and can correspond to input signal 232. As described below, one or more processors of audio system 102, e.g., device processor 202 or headphone processor 224, can be configured to process the outputs of the one or more microphones 106 to determine an acoustic transfer function that relates output signal 234 to input signal 232. The acoustic transfer function can be used to inform the selection or construction of audio filters for user 100.
Referring to
At operation 302, speaker 228 of headphones 104 worn by user 100 is driven with input signal 232 to generate sound 230. Input signal 232 can be a first input signal provided to speaker 228 by headphone processor 224. The first input signal can be a portion of user content signal 208, e.g., music or binaural audio, intended for listening by user 100. Alternatively, the first input signal can be a test signal intended specifically for the purpose of determining a user-specific acoustic transfer function of user 100. In either case, the first input signal can have predetermined frequency content.
Referring to
Referring to
At operation 304, microphone 106 can detect sound 230 reflected from pinna 406 as output signal 234. As described above, microphone 106 can receive sound 230 from a speaker placed internal to or external to pinna 406. Output signal 234 is an impulse response of the headphone/anatomical system of user 100 when input signal 232 is reproduced as sound 230. More particularly, the impulse response is dependent on the combination of headphones 104, geometry of pinna 406, and the interactions of headphones 104 and pinna 406 such as how well earcup 108 seals against a head of user. For example, when input signal 232 is played by speaker 228, sound 230 can transmit toward or through pinna 406 of user 100, and a portion of sound 230 can be reflected or diffracted from pinna 406 toward microphone 106. Similarly, a portion of sound 230 can be reflected or diffracted (as well as absorbed) by earcup 108 or earbud 502. The reflected/diffracted portion of sound 230 can be output signal 234 that is an input to microphone 106, and microphone 106 can generate a corresponding microphone output signal 222. Since output signal 234 defines the response of earcup 108 and/or pinna 406 to input signal 232, the impulse response depends on the unique anatomy of user 100 as is user-specific. The impulse response can be a signature of user 100 when wearing headphones 104.
Referring to
Acoustic transfer function 602 defines the change to sound 230 as the sound propagates through interior 404 to microphone 106. Microphone 106 may not be located in an entrance to the ear canal of user 100, however, microphone 106 may located near the entrance. Therefore, output signal 234 detected by microphone 106 can serve as a proxy for the sound actually heard at the ear canal entrance. That is, the location of microphone 106 can be correlated to the ear canal, and acoustic transfer function 602 can serve as a proxy for a true acoustic transfer function relating input signal 232 to a pressure generated at the ear canal entrance.
In an aspect, the one or more processors determine acoustic transfer function 602 in response to detecting that user 100 has donned or repositioned headphones 104. By way of example, headphones 104 can incorporate a proximity sensor that detects when earcup 108 or earbuds 502 are placed against the head or within the ear of user 100. Similarly, headphones 104 can incorporate an accelerometer that detects when earcup 108 or earbuds 502 are moved, e.g., during adjustment of headphones 104, while on user's head. Detection of the placement or repositioning can trigger monitoring of output signal 234 that is generated by speaker 228 via playback of input signal 232. The monitored output signal 234 and the predetermined input signal 232, e.g., audio data 206, can be used to determine acoustic transfer function 602 in real-time. The real-time determination can allow acoustic transfer function 602 to be accurately estimated based on user/headphone interactions, e.g., based on a current placement of headphones 104 on the user's head.
At operation 308, device processor 202 or headphone processor 224 can determine an HRTF or an HpEQ filter based on acoustic transfer function 602. The acoustic path between speaker 228 and microphone 106 of circumaural headphones 402 can be specific to a geometry of user's ears, given that the reflections and diffractions of sound 230 along the path depends on the unique shape of pinna 406. Similarly, the acoustic path between speaker 228 and microphone 106 of earbuds 502 can be specific to the geometry of user's ears, given that the reflections and diffractions of sound 230 along the path depends on the unique shape of pinna 406. In either case, acoustic transfer function 602 between the speaker input and the microphone output will contain information specific to the outer ear of user 100. That user-specific information can be used to select or design audio filters that are personalized for user 100, such as a personalized HRTF or a personalized HpEQ filter, as described below
In an aspect, an audio filter can be applied to input signal 232 delivered to speaker 228 to render binaural audio to user 100. When reproduced sound 230 is played to user 100, errors can be introduced by headphones 104 and/or the interaction between the headphones 104 and the ear of user 100. For example, reflections and reverberations of sounds 230 within interior 404 of earcup 108 can negatively impact the reproduced sound 230. Typically, headphones 104 can have a general equalization filter that is preset. The audio filter can be an HpEQ filter that modifies input signal 232 such that a pressure at an entrance to the ear canal of user 100 is approximately matched to a sound that the binaural rendering is intended to reproduce. In reality, however, generalized filters are not user-specific and do not sufficiently compensate for the effect of the combined user anatomy/headphones system on sound 230. In other words, generalized filters may not convince user 100 that no headphones are being worn while auditioning a binaural audio rendering.
Rather than apply a general equalization filter to compensate for an acoustic transfer function of a listener, audio system 102 can apply a personalized HpEQ to input signal 232. The personalized HpEQ can modify the input signal to counteract the effect of acoustic transfer function 602. Accordingly, the intended sound can be faithfully reproduced the ear canal entrance.
In an aspect, the one or more processors of audio system 102 generate the HpEQ filter based on acoustic transfer function 602. Still referring to operation 308, determining HpEQ filter can include determining an inverse 604 of acoustic transfer function 602. More particularly, acoustic transfer function 602 can be inverted to generate an inverse acoustic transfer function 604. Inverse acoustic transfer function 604 may be an inverse 604 function that reverses acoustic transfer function 602. For example, as shown in
Still referring to operation 308, a personalized HpEQ filter or a personalized HRTF for user 100 can be determined by selecting one or more of the audio filters from a dataset of predetermined transfer functions. Acoustic transfer function 602, the HRTF for user 100, and the HpEQ filter for user 100 are all anatomy dependent, and accordingly, can be correlated to each other. More particularly, users having a particular acoustic transfer function are also likely to have a particular HRTF and/or HpEQ filter. An audio filter can be picked by matching the acoustic transfer function 602, which is specific to user 100, to predetermined acoustic transfer functions of other users. More particularly, if the acoustic transfer function of user 100 matches an acoustic transfer function of another user 100, an acoustic transfer function of several other users, or a statistical measure of acoustic transfer functions of several other users (e.g., a mean or median acoustic transfer function of several other users), then an audio filter of the other user(s) having the matching acoustic transfer function can be selected as the audio filter for user 100. The selected audio filter is likely to fit user 100 because the acoustic transfer function and audio filter are correlated through shared anatomical characteristics of user 100 and the other users.
Referring to
Predetermined transfer functions 702 can include graphical representations of acoustic transfer functions stored in a database. For example, transfer function database 210 stored in device memory 204 or headphone memory 226 can include a dataset of acoustic transfer functions that describe an impulse response of the other users. The stored transfer functions are dependent on the pinnas 406 of the other users, and accordingly, transfer functions similar to acoustic transfer function 602 are likely to be for users having similar anatomy to user 100. Device processor 202 or headphone processor 224 can perform a signal matching algorithm to compare acoustic transfer function 602 to each of predetermined transfer functions 702 to find a matching transfer function 704. For example, matching transfer function 704 may be an acoustic transfer function 602 of the other users that is a closest fit to acoustic transfer function of user 100.
Matching transfer function 704 can correspond to an audio filter, e.g., an HRTF 706 or an HpEQ filter 708 of the other user(s) that have matching transfer function 704. The audio filters can be, for example, generated in a laboratory using known techniques to develop the database of audio filters that correspond to anatomical characteristics of the users. Accordingly, when matching transfer function 704 is determined, the audio filter corresponding to matching transfer function 704, e.g., HRTF 706 or HpEQ filter 708, can be selected. Similarly, when matching transfer function 704 is determined, any other perceptual parameter can be extracted from the transfer function, and a corresponding audio filter can be selected based on the parameter. For example, the extracted parameter of the matching transfer function 704 can be used to determine HRTF 706 or HpEQ filter 708.
Operation 308 may include determining only one of HRTF 706 or HpEQ filter 708. By way of example, when headphones 104 are earbuds, speaker 228 delivers sound 230 directly into ear canal, and accordingly, earbuds 502 do not introduce headphone-related effects into the transmitted sound 230. More particularly, earbuds 502 have an auditioned sound output that is effectively equal to the input signal 232. Thus, there is no need to determine or apply an HpEQ filter 708 to the input signal 232 in the case of earbuds 502. Acoustic transfer function 602 may nonetheless be used to determine HRTF 706 of user 100. For example, acoustic transfer function 602 relating output signal 234 detected at stem 506 to input signal 232 can be used to select HRTF 706 specific to user 100. The measured acoustic transfer function 602 is anatomy-dependent because sound 230 reflects from pinna 406 as it travels out of the ear toward microphone 106 mounted on stem 506. This anatomical dependence provides that acoustic transfer function 602 corresponds to a particular HRTF 706 for the pinna shape. Accordingly, acoustic transfer function 602 represents the effect of pinna 406 on acoustics, which can be used to select HRTF 706 from transfer function database 210.
At operation 310, one or more processors of audio system 102 can apply one or more of the determined HRTF 706 or HpEQ filter 708 to input signal 232. When an optimized HRTF 706 and/or HpEQ filter 708 is selected and/or personalized to generate the individualized audio filters of user 100, audio system 102 can use the individualized audio filters to render binaural audio to user 100. Binaural rendering of audio to user 100 can include applying the individualized audio filters to a second input signal. The second input signal can be a portion of user content signal 208, e.g., a different portion of user content signal 208 than the first input signal that was played at operation 302. HRTF 706 and HpEQ filter 708 can be combined to achieve a single linear transfer function that generates a spatial input signal based on second input signal. More particularly, the spatial input signal can be generated by applying a selected HRTF 706 of user 100 to second input signal to generate an intended binaural audio signal. The intended binaural audio signal may, however, not account for the effect of acoustic transfer function 602 when the intended binaural audio signal is played by speaker 228. Accordingly, a selected or generated HpEQ filter 708, which is an inverse function of acoustic transfer function 602 of user 100, can be applied to the intended binaural audio signal to generate a modified binaural audio signal. The modified binaural audio signal can be the spatial input signal.
At operation 312, audio system processor(s) can drive speaker 228 with the spatial input signal to generate a spatialized sound. The spatial input signal can be transmitted to headphones 104 for playback to user 100. More particularly, when speakers 228 of headphones 104 play the spatial input signal, the spatialized sound can be emitted to render binaural audio to user 100.
The reproduced audio, which is based on the individualized audio filters that are specific to an anatomy of user 100, can improve an illusion of external sound sources in spatial audio and improve the overall sound quality experienced by user 100. The improvement can be transparent to user 100 because the determination of acoustic transfer function 602, HRTF 706, or HpEQ filter 708 can take place while delivering audio to user 100 without requiring input by user 100. The determination of acoustic transfer function 602 and derivation of audio filters by selection or generation can occur in an uncontrolled environment, and thus, can be performed seamlessly and with relative ease as compared to developing user-specific HRTF and HpEQ filters for a user in a controlled laboratory setting.
As described above, one aspect of the present technology is the gathering and use of data available from various sources to determine an HRTF or an HpEQ filter for a user. The present disclosure contemplates that in some instances, this gathered data may include personal information data that uniquely identifies or can be used to contact or locate a specific person. Such personal information data can include demographic data, location-based data, telephone numbers, email addresses, TWITTER ID's, home addresses, data or records relating to a user's health or level of fitness (e.g., vital signs measurements, medication information, exercise information), date of birth, or any other identifying or personal information.
The present disclosure recognizes that the use of such personal information data, in the present technology, can be used to the benefit of users. For example, the personal information data can be used to determine the HRTF or the HpEQ filter for the user. Accordingly, use of such personal information data provides an improved spatial audio experience to the user. Further, other uses for personal information data that benefit the user are also contemplated by the present disclosure. For instance, health and fitness data may be used to provide insights into a user's general wellness, or may be used as positive feedback to individuals using technology to pursue wellness goals.
The present disclosure contemplates that the entities responsible for the collection, analysis, disclosure, transfer, storage, or other use of such personal information data will comply with well-established privacy policies and/or privacy practices. In particular, such entities should implement and consistently use privacy policies and practices that are generally recognized as meeting or exceeding industry or governmental requirements for maintaining personal information data private and secure. Such policies should be easily accessible by users, and should be updated as the collection and/or use of data changes. Personal information from users should be collected for legitimate and reasonable uses of the entity and not shared or sold outside of those legitimate uses. Further, such collection/sharing should occur after receiving the informed consent of the users. Additionally, such entities should consider taking any needed steps for safeguarding and securing access to such personal information data and ensuring that others with access to the personal information data adhere to their privacy policies and procedures. Further, such entities can subject themselves to evaluation by third parties to certify their adherence to widely accepted privacy policies and practices. In addition, policies and practices should be adapted for the particular types of personal information data being collected and/or accessed and adapted to applicable laws and standards, including jurisdiction-specific considerations. For instance, in the US, collection of or access to certain health data may be governed by federal and/or state laws, such as the Health Insurance Portability and Accountability Act (HIPAA); whereas health data in other countries may be subject to other regulations and policies and should be handled accordingly. Hence different privacy practices should be maintained for different personal data types in each country.
Despite the foregoing, the present disclosure also contemplates embodiments in which users selectively block the use of, or access to, personal information data. That is, the present disclosure contemplates that hardware and/or software elements can be provided to prevent or block access to such personal information data. For example, in the case of spatial audio rendering, the present technology can be configured to allow users to select to “opt in” or “opt out” of participation in the collection of personal information data during registration for services or anytime thereafter. In addition to providing “opt in” and “opt out” options, the present disclosure contemplates providing notifications relating to the access or use of personal information. For instance, a user may be notified upon downloading an app that their personal information data will be accessed and then reminded again just before personal information data is accessed by the app.
Moreover, it is the intent of the present disclosure that personal information data should be managed and handled in a way to minimize risks of unintentional or unauthorized access or use. Risk can be minimized by limiting the collection of data and deleting data once it is no longer needed. In addition, and when applicable, including in certain health related applications, data de-identification can be used to protect a user's privacy. De-identification may be facilitated, when appropriate, by removing specific identifiers (e.g., date of birth, etc.), controlling the amount or specificity of data stored (e.g., collecting location data a city level rather than at an address level), controlling how data is stored (e.g., aggregating data across users), and/or other methods.
Therefore, although the present disclosure broadly covers use of personal information data to implement one or more various disclosed embodiments, the present disclosure also contemplates that the various embodiments can also be implemented without the need for accessing such personal information data. That is, the various embodiments of the present technology are not rendered inoperable due to the lack of all or a portion of such personal information data. For example, the HRTF or the HpEQ filter can be determined by inferring preferences based on non-personal information data or a bare minimum amount of personal information, such as the content being requested by the device associated with a user, other non-personal information available to the device processors, or publicly available information.
To aid the Patent Office and any readers of any patent issued on this application in interpreting the claims appended hereto, applicants wish to note that they do not intend any of the appended claims or claim elements to invoke 35 U.S.C. 112(f) unless the words “means for” or “step for” are explicitly used in the particular claim.
In the foregoing specification, the invention has been described with reference to specific exemplary aspects thereof. It will be evident that various modifications may be made thereto without departing from the broader spirit and scope of the invention as set forth in the following claims. The specification and drawings are, accordingly, to be regarded in an illustrative sense rather than a restrictive sense.
This application claims the benefit of priority of U.S. Provisional Patent Application No. 62/737,728, filed on Sep. 27, 2018, and incorporates herein by reference that provisional patent application.
Number | Name | Date | Kind |
---|---|---|---|
6256386 | Phillips | Jul 2001 | B1 |
7406179 | Ryan | Jul 2008 | B2 |
9578415 | Zhou | Feb 2017 | B1 |
9747887 | O'Connell | Aug 2017 | B2 |
9918177 | Horbach | Mar 2018 | B2 |
10142733 | Fujii | Nov 2018 | B2 |
10382864 | Alderson | Aug 2019 | B2 |
10652686 | Wardle | May 2020 | B2 |
20140044275 | Goldstein | Feb 2014 | A1 |
20140072135 | Bajic | Mar 2014 | A1 |
20140311499 | Lee | Oct 2014 | A1 |
20140341388 | Goldstein | Nov 2014 | A1 |
20150223002 | Mehta | Aug 2015 | A1 |
20150243271 | Goldstein | Aug 2015 | A1 |
20150289064 | Jensen | Oct 2015 | A1 |
20160365082 | Poulsen | Dec 2016 | A1 |
20170094386 | Trainer | Mar 2017 | A1 |
20170156017 | Gamper | Jun 2017 | A1 |
20170195793 | Vilermo | Jul 2017 | A1 |
20170245040 | Hankey | Aug 2017 | A1 |
20170272867 | Zisapel | Sep 2017 | A1 |
20170295445 | Christoph | Oct 2017 | A1 |
20180077511 | Mehta et al. | Mar 2018 | A1 |
20180132764 | Jain | May 2018 | A1 |
20180324516 | Campbell | Nov 2018 | A1 |
20190069068 | Villasenor | Feb 2019 | A1 |
20190246217 | Miller | Aug 2019 | A1 |
20190373357 | Khaleghimeybodi | Dec 2019 | A1 |
20190394564 | Mehra | Dec 2019 | A1 |
20200184057 | Mukund | Jun 2020 | A1 |
Number | Date | Country |
---|---|---|
WO2017161383 | Sep 2017 | WO |
Entry |
---|
HRTF Wikipedia page available 8/18 (Year: 2018). |
Number | Date | Country | |
---|---|---|---|
62737728 | Sep 2018 | US |