The present disclosure relates generally to the sound management in connection with content presentation. Sound forms a significant part of the enjoyment of content consumption. This is especially the case as relating to restricting or eliminating the perceptibility of sounds that do not originate from the content being consumed. In the case of theaters such as movie theaters, this can include attempts at soundproofing of the theater.
Soundproofing of a theater can be accomplished via a number of different techniques. These techniques can include, for example: building the theater as a room within a room to better acoustically isolate the theater; the inclusion of sound baffles along the walls and/or ceiling of the theater to reflect sound away from the walls and/or ceiling; and/or construction of the theater with specialized materials including, for example, soundproof drywall, soundproof insulation, and/or acoustic panels.
While these common soundproofing techniques can effectively dampen sound, they also present several drawbacks. Many of these techniques are bulky and take up significant amount of space. Further, these techniques may force certain appearances of the theater which may limit the creative ability of a theater designer to create a desired consumer experience. Further, these techniques are not always effective, especially for loud sounds. In light of these shortcomings in current soundproofing techniques for theaters, new systems and methods for active acoustic compensation in proximate theaters are desired.
Some embodiments of the present disclosure relate to audio-active noise cancellation or masking. Specifically, this can include the identification of a coming first acoustic event in a first theater, the determination of the effect of that acoustic event in a second theater, and the generation of second acoustic event in the second theater to mask or cancel the first acoustic event. This second acoustic event in the second theater can include a video portion or an audio portion, and this second acoustic event can be integrated into a plot or narrative of content presented in the second theater.
This second acoustic event can be selected and controlled according to one or several attributes of the first acoustic event including, for example, the magnitude of the first acoustic event including, for example, the amplitude of sound waves forming the first acoustic event, the frequency of some or all of the sound waves forming the first acoustic event, or the phase of some or all of the sound waves forming the first acoustic event. The second acoustic event can further, in some embodiments, be selected and controlled according to the attenuation of the first acoustic event before reaching the second theater. In some embodiments, data indicative of the one or several attributes of the first acoustic event can be modified by one or several values or parameters representative of sound attenuation between the first and second theaters. The second acoustic event can be selected and controlled according these one or several modified attributes of the first acoustic event.
One aspect of the present disclosure relates to a system for active acoustic compensation in proximate theaters. The system includes: a first theater having a first content presentation system; a second theater having a second content presentation system; and a processor in communicating connection with the first content presentation system and the second content presentation system. The processor can: direct presentation of first content in the first theater via the first content presentation system; direct presentation of second content in the second theater via the second content presentation system; identify an impending first acoustic event in the first theater, which first acoustic event has a level sufficient to be detectable in the second theater; and control the second content presentation system to generate a second acoustic event at least partially coinciding with the first acoustic event, which second acoustic event mitigates the detectability of the first acoustic event in the second theater.
In some embodiments, controlling the second content presentation system to generate a second acoustic event includes controlling the second content presentation system to generate a second visual event corresponding the second acoustic event. In some embodiments, the second acoustic event and the second visual event are integrated into a narrative formed by content presented before and after the second acoustic event and the second visual event. In some embodiments, the second acoustic event masks the first acoustic event. In some embodiments, the second acoustic event mitigates the first acoustic event via destructive cancellation.
In some embodiments, the processor can determine a magnitude of the second acoustic event. In some embodiments, determining the magnitude of second acoustic event includes: identifying the magnitude of the first acoustic event in the second theater; and matching the magnitude of the second acoustic event to the determined magnitude of the first acoustic event. In some embodiments, identifying the magnitude of the first acoustic event in the second theater includes: identifying an amplitude of the first acoustic event in the first theater; identifying transmission losses from the first theater to the second theater; and calculating the amplitude of the first acoustic event in the second theater based on the identified amplitude of the first acoustic event in the first theater and the transmission losses.
In some embodiments, identifying the amplitude of the first acoustic event in the first theater includes retrieving data for generating the first acoustic event from a memory. In some embodiments, the data characterizes the amplitude of the first acoustic event in the first theater. In some embodiments, determining the magnitude of second acoustic event includes: identifying an attribute of the first acoustic event in the second theater; and identifying a cancellation attribute for the second acoustic event in the second theater. In some embodiments, identifying the cancellation attribute includes: identifying at least one initial wave property of the first acoustic event in the first theater; identifying a filter value affecting acoustic transmission from the first theater to the second theater; and calculating at least one wave property of the first acoustic event in the second theater based on the at least one initial wave property and the filter value. In some embodiments, the at least one wave property includes at least one of: a frequency; a phase; and an amplitude.
One aspect of the present disclosure relates to a method for active acoustic compensation in proximate theaters. The method includes: directing presentation of first content in the first theater via the first content presentation system; directing presentation of second content in the second theater via the second content presentation system; identifying an impending first acoustic event in the first theater, which first acoustic event has a level sufficient to be detectable in the second theater; and controlling the second content presentation system to generate a second acoustic event at least partially coinciding with the first acoustic event, which second acoustic event mitigates the detectability of the first acoustic event in the second theater.
In some embodiments, the second acoustic event masks the first acoustic event. In some embodiments, directing presentation of second content in the second theater via the second content presentation system includes controlling the second content presentation system to present video and audio content forming a narrative. In some embodiments, controlling the second content presentation system to generate a second acoustic event includes controlling the second content presentation system to generate a second visual event corresponding the second acoustic event. In some embodiments, the second acoustic event and the second visual event generated are integrated into the narrative.
In some embodiments, the second acoustic event mitigates the first acoustic event via destructive cancellation. In some embodiments, the method includes determining a magnitude of the second acoustic event. In some embodiments, determining the magnitude of second acoustic event includes: identifying the magnitude of the first acoustic event in the second theater; and matching the magnitude of the second acoustic event to the determined magnitude of the first acoustic event. In some embodiments, identifying the magnitude of the first acoustic event in the second theater includes: identifying an amplitude of the first acoustic event in the first theater; identifying transmission losses from the first theater to the second theater; and calculating the amplitude of the first acoustic event in the second theater based on the identified amplitude of the first acoustic event in the first theater and the transmission losses.
In some embodiments, identifying the amplitude of the first acoustic event in the first theater comprises retrieving data for generating the first acoustic event from a memory. In some embodiments, the data characterizes the amplitude of the first acoustic event in the first theater. In some embodiments, determining the magnitude of second acoustic event includes: identifying an attribute of the first acoustic event in the second theater; and identifying a cancellation attribute for the second acoustic event in the second theater. In some embodiments, identifying the cancellation attribute includes: identifying at least one initial wave property of the first acoustic event in the first theater; identifying a filter value affecting acoustic transmission from the first theater to the second theater; and calculating at least one wave property of the first acoustic event in the second theater based on the at least one initial wave property and the filter value. In some embodiments, the at least one wave property includes at least one of: a frequency; a phase; and an amplitude.
The ensuing description provides illustrative embodiment(s) only and is not intended to limit the scope, applicability or configuration of the disclosure. Rather, the ensuing description of the illustrative embodiment(s) will provide those skilled in the art with an enabling description for implementing a preferred exemplary embodiment. It is understood that various changes can be made in the function and arrangement of elements without departing from the spirit and scope as set forth in the appended claims.
I. Introduction
Storytelling can occur in a number of locations and can be performed using a number of techniques. Modern storytelling includes the presentation of content, which can be digital content, to a content consumer. This content can include video content or audio content. Such storytelling can use sound in the form of character monologue or dialogue, sound effects corresponding to a depicted action or event such as, for example, the sound of an automobile engine accompanying video of an automobile race, and/or music. This sound can develop a plot of a storyline being conveyed, develop one or several characters, link events, or convey emotion. Aspects of the sound can be modified throughout the presentation of content to enhance storytelling. These modifications can include, changes to the volume, frequency, or tempo of generated sounds. In many instances, the conveyed sound can be very quiet, or can be very loud.
Storytelling can occur in a number of locations including within a movie theater, or on an amusement ride. In many instances, one or several theaters, which can include a room or area in which content is presented such as a movie theater or an amusement ride, may be proximate to each other such that some sound from one theater may travel to another theater. This travel of sound between theaters can adversely affect the experience of content consumers. In addition to this, an acoustic event, which can include either sound or vibrations can originate from a first theater and may travel to a second theater, adversely affecting the experience of content consumer in the second theater.
Traditionally, acoustics traveling between theaters, or specifically from a first theater to a second theater has been managed via passive measures. While these passive measures have provided some benefit, they have not been able to adequately address or eliminate this acoustic pollution. Active sound compensation addresses many of the shortcomings of traditional sound isolation techniques. The active acoustic compensation can include generating one or several acoustic events in the second theater. These one or several acoustic events in the second theater can mask sound traveling from the first theater or can eliminate that sound traveling from the first theater via destructive interference.
II. Active Acoustic Compensation System
With reference now to
The system 100 can include memory 106. The memory 106 can represent one or more storage media and/or memories for storing data, including read only memory (ROM), random access memory (RAM), magnetic RAM, core memory, magnetic disk storage mediums, optical storage mediums, flash memory devices and/or other machine readable mediums for storing information. The term “machine-readable medium” includes, but is not limited to portable or fixed storage devices, optical storage devices, and/or various other storage mediums capable of storing that contain or carry instruction(s) and/or data. The memory 106 can be an integral part of the processor 102 and/or can be separate from the processor 102. In embodiments in which the memory 106 is separate from the processor 102, the memory 106 and the processor 102 can be communicatingly linked via, for example, communications network 130. In some embodiments, the communications network 130 can comprise any wired or wireless communication connection between the components of the simulation system 100.
The memory 106 can include software code and/or instructions for directing the operation of the processor 102 and/or one or several databases 106 containing information used by the processor 102 and/or generated by the processor 102. These databases include, for example, a content database 106-A, a filter database 106-B, and a compensation database 106-C.
The content database 106-A can include content for presentation in one or several theaters. In some embodiments, this content can comprise video content, audio content, combined video and audio content, or the like. This content can be in the form of one or several films, movies, shows, simulations, interactive stories, or video games. In some embodiments, this content can include a storyline, a plot, or narrative the may be static or that may be dynamic based on received user inputs. In some embodiments, the content database 106-A can include information identifying one or several acoustic events in the content. This information can identify the time within the content presentation of the occurrence of the one or several acoustic events. In some embodiments, this information in the content database can include attributes of those one or several acoustic events, which attributes can include, for example, the duration of one or several acoustic events, the volume of one or several acoustic events, the frequency or frequencies forming the one or several acoustic events, and/or the phase of sound waves forming the one or several acoustic events.
The filter database 106-B can include information relating to the transmission of sound and/or vibrations between theaters. In some embodiments, this information can be specific to a pair of theaters such as, for example, the pair including the first theater and the second theater, and in some embodiments, this information can be generic. The information relating to the transmission of sound and/or vibrations between theaters can, for example, specify transmission losses to sound and/or vibrational waves moving between theaters, and can specifically characterize the degree of sound and/or vibrational dampening occurring between theaters, phase shift in sound and/or vibrational waves of the acoustic event occurring between theaters, change in the frequency of the acoustic event between theaters, or the like. In some embodiments, this information can be stored in one or several filter values within the filter database 106-B.
The compensation database 106-C can include information identifying one or several compensating acoustic events and/or associated video events or visual events. In some embodiments, a compensating acoustic event can be an acoustic event selected and/or generated to decrease the detectability of sound and/or vibrations coming from another theater. This compensating acoustic event can compensate for the sound and/or vibrations from the other theater by, for example, masking the sound and/or vibrations from the other theater, or mitigating the sound and/or vibrations from the other theater via, for example, destructive interference. This acoustic event and the accompanying visual event can be integrated into the narrative which can include content presented before the acoustic event and content presented after the acoustic event.
The system 100 can include a plurality of theaters 108, which can include, a first theater 108-A, a second theater 108-B, and up to an Nth theater, 108-N. The theaters 108 can be any room or area in which content can be provided to a content consumer. The theaters 108 can include, for example, one or several movie theaters, one or several amusement rides, or the like. The theaters 108 can each include a content presentation system 110 such that the first theater 108-A includes a first content presentation system 110-A, the second theater 108-B includes a second content presentation system 110-B, and the Nth theater 108-N includes an Nth content presentation system 110-N.
The content presentation system 110 of the theaters 108 can provide and/or present audio content via the audio presentation system 112 and/or video content via the video presentation system 114. Specifically, the audio presentation system 112 can include a first audio presentation system 112-A in the first theater 108-A and second audio presentation system 112-B in the second theater 108-B. Similarly, the video presentation system 114 can include a first video presentation system 114-A in the first theater 108-A and a second video presentation system 114-B in the second theater 108-B.
III. Sound Transmission Between Theaters
With reference now to
Each of the first and second theaters 108-A, 108-B, and specifically each of the first and second content presentation systems 110-A, 110-B, is communicatingly connected to the processor 102. The processor 102 controls and/or can control the first and second content presentation systems 110-A, 110-B. Specifically, the processor can direct the first content presentation system 110-A to present first content in the first theater 108-A and can direct the second content presentation system 110-B to present second content in the second theater 108-B.
During the presentation of the first content in the first theater 108-A, a first acoustic event is generated. Specifically, the processor 102 controls the first content presentation system 110-A to generate sound and/or vibrations associated with a first acoustic event. The sound and/or vibrations generated by the first acoustic event travels to the second theater 108-B via sound and/or vibration waves 200. If the sound and/or vibrations generated by the first acoustic event is sufficiently loud, then the sound and/or vibrations can interfere with the presentation of content in the second theater 108-B.
IV. Active Acoustic Compensation
With reference now to
The process 300 begins at block 302 wherein presentation of first content in the first theater 108-A is directed. In some embodiments, this step can include the generation and sending of one or several control signals from the processor 102 to the first theater 108-A, and specifically to the first content presentation system 110-A of the first theater 108-A. In some embodiments, for example, the processor 102 can retrieve information from the memory 106, and specifically from the content database 106-A. This information can include all or portions of the first content including, for example, data relating to one or several acoustic events included in the content. The processor 102 can then provide all or portions of the first content to the first theater 108-A and/or to the first content presentation system 110-A via the network 130.
At block 304 of process 300, presentation of second content in the second theater 108-B is directed. In some embodiments, this step can include the generation and sending of one or several control signals from the processor 102 to the second theater 108-B, and specifically to the second content presentation system 110-B of the second theater 108-B. In some embodiments, for example, the processor 102 can retrieve information from the memory 106, and specifically from the content database 106-A. This information can include all or portions of the second content including, for example, data relating to one or several acoustic events included in the second content. The processor 102 can then provide all or portions of the second content to the second theater 108-B and/or to the second content presentation system 110-B via the network 130.
At block 306 of the process 300 an impending first acoustic event is identified. In some embodiments, the impending first acoustic event can occur in the first content presented in the first theater 108-B. In some embodiments, the first acoustic event can be one or several sounds and/or vibrations identified as affecting viewing in other theaters 108, and, in some embodiments, an acoustic event can be impending when it is set to occur within the next 10 minutes, within the next 5 minutes, within the next 1 minute, within the next 45 seconds, within the next 30 seconds, within the next 15 seconds, within the next 5 seconds, or within the next 1 second. In some embodiments, one or several sounds and/or vibrations can be identified as an acoustic event based on, for example, a volume of these one or several sounds and/or vibrations exceeding a volume threshold. This comparison between the volume of one or several sounds and/or vibrations to a volume threshold can be performed simultaneous with the presentation of content as described in block 302, or can be performed before the presentation of content as described in block 302. In some embodiments, for example, this comparison can be performed in advance of the presentation of content as described in block 302 and the result of this comparison can be stored in the memory 106 and specifically within the content database 106-A.
In some embodiments, an acoustic event can be identified within the content database 106-A and be associated with the content of which the acoustic event is a part. In some embodiments, the designation of an acoustic event within the content database 106-A can include the association of one or several values with one or several sounds and/or vibrations and/or the storing of one or several values associated with one or several sounds and/or vibrations in the content database 106-A. In some embodiments in which the acoustic event is designated within the content database 106-A, the identification of an impending first acoustic event can include the determination of the presence of one or several values designating one or several sounds and/or vibrations as an acoustic event. This determination can be made by the processor 102.
At block 308 of the process 300, the magnitude of a second acoustic event is determined. In some embodiments, this can include determining one or several properties of the second acoustic event to mitigate the effect of the first acoustic event on content consumers in the second theater 108-B. In some embodiments, this can include selecting a volume of the second acoustic event, a frequency of the second acoustic event, a phase of the second acoustic event, and/or a duration of the second acoustic event to mask the first acoustic event and/or destructively interfere with the first acoustic event. In some embodiments, the magnitude of the second acoustic event can be determined based on the magnitude of the first acoustic event. The magnitude of the second acoustic event can be determined by the processor 102.
At block 310 of the process 300, the second acoustic event is generated. In some embodiments, this can include the directing of the generation of the second acoustic event by the processor 102. This can include the generation and sending of one or several control signals from the processor 102 to the second theater 108-B, and specifically to the second content presentation system 110-B, which one or several control signals cause the generation of the second acoustic event. In some embodiments, the second acoustic event can be generated at the same time as the generation of the first acoustic event and/or at the time the first acoustic event is heard in the second theater 108-B, and in some embodiments, the second acoustic event can be generated to be at least partially coinciding with the first acoustic event, such that the second acoustic event mitigates the detectability of the first acoustic event in the second theater 108-B.
In embodiments in which the presented content comprises static content that does not change based on user inputs, the generation of the second acoustic event can include the modification of one or several existing sounds and/or vibrations within the content being presented in the second theater 108-B. This can include, for example, the increasing of the volume of one or several sounds and/or vibrations in the second theater to mask the sound and/or vibrations from the first theater, or the changing of the frequency of one or several sounds and/or vibrations in the second theater. In embodiments in which the presented content comprises dynamic content, the generation of the second acoustic event can include the addition of one or several second acoustic event, and in some embodiments, associated video content to the content being presented in the second theater 108-B.
With reference now to
After the magnitude of the first acoustic event in the second theater 108-B has been determined, the process 400 proceeds to block 404 wherein a magnitude of the second acoustic event is matched to the identified magnitude of the first acoustic event in the second theater. In some embodiments, this can include selecting a second acoustic event from the event database 106-C that has magnitude equal to the magnitude of the first acoustic event in the second theater 108-B or that has a magnitude greater than the magnitude of the first acoustic event in the second theater 108-B. In some embodiments, matching the magnitude of the second acoustic event to the identified magnitude can include the modification of the magnitude of a second acoustic event so that the magnitude of the second acoustic event is greater than or equal to the magnitude of the first acoustic event in the second theater 108-B. The matching of the magnitude of the second acoustic event to the identified magnitude of the first acoustic event in the second theater 108-B can be performed by the processor 102.
After the magnitude of the second acoustic event has been matched to the identified magnitude of the first acoustic event in the second theater 108-B, the process 400 can advance to block 310 of the process 300 shown in
With reference now to
The process 450 begins at block 452 wherein the volume of the first acoustic event is identified. In some embodiments, this volume can be the volume of the first acoustic event in the first theater 108-A. The volume of the first acoustic event can be determined based on information stored in the memory 106 and specifically in the content database 106-A. This information stored in the memory 106 or in the content database 106-A can, in some embodiments, be used to generate the first acoustic event. In some embodiments, information specifying the volume in the memory 106 or in the content database 106-A can, for example, specify an amplitude of acoustic waves forming the first acoustic event, a decibel level of the first acoustic event, a power consumption level of the first audio presentation system 112-A to generate the first acoustic event, and/or a normalized value indicative of the volume of the first acoustic event. In embodiments in which information specifying the volume of the first acoustic event is contained in the memory 106 and specifically in the content database 106-A, identifying the volume of the first acoustic event can include retrieving this information from the memory 106 and/or the content database 106-A with the server 102.
After the volume of the first acoustic event has been identified, the process 450 proceeds to block 454 wherein one or several transmission losses are identified. In some embodiments, the transmission losses can specify or indicate the dampening or degree of dampening of sounds and/or vibrations traveling from the first theater 108-A to the second theater 108-B. In some embodiments the transmission losses can be identified via one or several values which can be stored in the memory 106 and specifically in the filter database 106-B. These one or several values can indicate, for example, a degree to which the volume of a sound and/or vibration decreases when traveling from the first theater 108-A to the second theater 108-B, or the number of decibels lost in transmission of a sound and/or vibration from the first theater 108-A to the second theater 108-B. In some embodiments, the identification of transmission losses in step 454 can include the retrieval of these one or several values from the memory 106 and specifically from the filter database 106-B by the server 102.
After the transmission losses have been identified, the process 450 proceeds to block 456 wherein a predicted volume or amplitude of the first acoustic event in the second theater 108-B is identified, or in some embodiments, is calculated. The identification of the predicted volume of the first acoustic event in the second theater 108-B can include the application of the transmission losses identified in block 454 to the volume of the first acoustic event identified in block 452. This can include, for example, performing one or several mathematical operations on the identified volume of the first acoustic event based on the one or several values characterizing transmission losses of sound and/or vibration from the first theater 108-A to the second theater 108-B. These operations can include, for example, multiplying or dividing the identified volume by the one or several values characterizing transmission losses, or subtracting the one or several values characterizing transmission losses from the identified volume.
After the predicted volume of the first acoustic event in the second theater 108-B has been calculated, the process 450 can advance to block 404 of the process 400 shown in
With reference now to
After the attribute of the first acoustic event has been identified, the process 470 proceeds to block 474 wherein a filter value is identified. In some embodiments, the filter value can specify or indicate how, or the degree to which one or several attributes of sounds and/or vibrations traveling from the first theater 108-A to the second theater 108-B change. In some embodiments the filter value can comprise one or several values which can be stored in the memory 106 and specifically in the filter database 106-B. These one or several values can indicate, for example, how or the degree to which some or all of the attributes of the first acoustic event change while traveling from the first theater 108-A to the second theater 108-B, including, for example, how or the degree to which the frequency, the amplitude, the volume, or the phase of the sound and/or vibration change while traveling from the first theater 108-A to the second theater 108-B. In some embodiments, the identification of filter value in step 484 can include the retrieval of these one or several values from the memory 106 and specifically from the filter database 106-B by the server 102.
After the filter value has been identified, the process 470 proceeds to block 476 wherein a predicted attribute, which can be one or several wave properties, of the first acoustic event in the second theater 108-B is identified or, in some embodiments, is calculated. In some embodiments, the identification of the predicted attribute of the first acoustic event in the second theater 108-B can include the application of the filter value identified in block 474 to the attribute of the first acoustic event identified in block 472. This can include, for example, performing one or several mathematical operations on the one or several identified attributes of the first acoustic event based on the one or several filter values. These operations and/or the identification or calculation of the attributes of the first acoustic event in the second theater 108-B can be performed by the processor 102.
After the identification of the one or several attributes of the first acoustic event in the second theater 108-B, the process 470 proceeds to block 478, wherein a cancellation attribute for the second acoustic event is identified. The identification of the cancellation attribute can be based on the identified attribute of the first acoustic event in the second theater 108-B determined in block 476. In some embodiments, the identification of the cancellation attribute can include identification a second acoustic event from the event database 106-C that has one or several attributes such that, when generated, the second acoustic event destructively interferes with the first acoustic event in the second theater 108-B. In some embodiments, this can include selecting a second acoustic event having the same or similar frequency to the first acoustic event or to the attribute identified in block 476, an opposite phase to the phase of the first acoustic event or to the attribute identified in block 476, and in some embodiments, a desired volume or amplitude. In some embodiments, this desired volume or amplitude can be equal to the volume or amplitude of the first acoustic event in the second theater 108-B, and in some embodiments, this volume or amplitude can be less than or greater than the volume or amplitude of the first acoustic event in the second theater 108-B. In some embodiments, identifying the cancellation attribute for the second acoustic event can include the modification of one or several attributes of a second acoustic event so that destructive interference between the second acoustic event and the first acoustic event in the second theater 108-B is maximized. This can include, for example, modifying the frequency of the second acoustic event, modifying the volume or amplitude of the second acoustic event, or modifying the phase of the second acoustic event. The identifying of the cancellation attribute for the second acoustic event in the second theater 108-B can be performed by the processor 102.
After the cancellation attribute for the second acoustic event has been identified, the process 470 can advance to block 310 of the process 300 shown in
V. Computer System
The one or more data processors or central processing units (CPUs) 1005 execute program code to implement the processes described herein. The one or more graphics processor or graphical processing units (GPUs) 1010 execute logic or program code associated with graphics or for providing graphics-specific functionality. Memory subsystem 1015 can store information, e.g., using machine-readable articles, information storage devices, or computer-readable storage media. Storage subsystem 1020 can also store information using machine-readable articles, information storage devices, or computer-readable storage media. Storage subsystem 1020 may store information using storage media 1045 that can be any desired storage media.
The one or more input/output (I/O) interfaces 1025 can perform I/O operations and the one or more output devices 1055 can output information to one or more destinations for computer system 1000. One or more input devices 1050 and/or one or more output devices 1055 may be communicatively coupled to the one or more I/O interfaces 1025. The one or more input devices 1050 can receive information from one or more sources for computer system 1000. The one or more output devices 1055 may allow a user of computer system 1000 to view objects, icons, text, user interface widgets, or other user interface elements.
Communications interface 1030 can perform communications operations, including sending and receiving data. Communications interface 1030 may be coupled to communications network/external bus 1060, such as a computer network, a USB hub, or the like. A computer system can include a plurality of the same components or subsystems, e.g., connected together by communications interface 1030 or by an internal interface.
Computer system 1000 may also include one or more applications (e.g., software components or functions) to be executed by a processor to execute, perform, or otherwise implement techniques disclosed herein. These applications may be embodied as data and program code 1040. Such applications may also be encoded and transmitted using carrier signals adapted for transmission via wired, optical, and/or wireless networks conforming to a variety of protocols, including the Internet.
The above description of exemplary embodiments of the invention has been presented for the purposes of illustration and description. It is not intended to be exhaustive or to limit the invention to the precise form described, and many modifications and variations are possible in light of the teaching above. The embodiments were chosen and described in order to best explain the principles of the invention and its practical applications to thereby enable others skilled in the art to best utilize the invention in various embodiments and with various modifications as are suited to the particular use contemplated.
This application claims the benefit of U.S. Provisional Application No. 62/610,862, filed on Dec. 27, 2017, and entitled “SYSTEMS AND METHODS FOR ACTIVE SOUND COMPENSATION”, the entirety of which is hereby incorporated by reference herein.
Number | Name | Date | Kind |
---|---|---|---|
20070266395 | Lee | Nov 2007 | A1 |
20080192945 | McConnell | Aug 2008 | A1 |
20130230175 | Bech | Sep 2013 | A1 |
20140380350 | Shankar | Dec 2014 | A1 |
20150264507 | Francombe | Sep 2015 | A1 |
20150319518 | Wilson | Nov 2015 | A1 |
20170236512 | Williams | Aug 2017 | A1 |
20180018984 | Dickins | Jan 2018 | A1 |
Number | Date | Country | |
---|---|---|---|
62610862 | Dec 2017 | US |