One or more implementations relate generally to surround sound audio rendering, and more specifically to impedance matching filters and equalization systems for headphone rendering.
Virtual rendering of spatial audio over a pair of speakers commonly involves the creation of a stereo binaural signal that represents the desired sound arriving at the listener's left and right ears and is synthesized to simulate a particular audio scene in three-dimensional (3D) space, containing possibly a multitude of sources at different locations. For playback through headphones rather than speakers, binaural processing or rendering can be defined as a set of signal processing operations aimed at reproducing the intended 3D location of a sound source over headphones by emulating the natural spatial listening cues of human subjects. Typical core components of a binaural renderer are head-related filtering to reproduce direction dependent cues as well as distance cues processing, which may involve modeling the influence of a real or virtual listening room or environment. One example of a present binaural renderer processes each of the 5 or 7 channels of a 5.1 or 7.1 surround in a channel-based audio presentation to 5/7 virtual sound sources in 2D space around the listener. Binaural rendering is also commonly found in games or gaming audio hardware, in which case the processing can be applied to individual audio objects in the game based on their individual 3D position. With the growing importance of headphone listening and the additional flexibility brought by object-based content (such as the Dolby® Atmos™ system), there is greater opportunity and need to have the mixers create and encode specific binaural rendering metadata at content creation time to maintain the spatial cues of the original content.
During headphone playback, matching the response at a person's ear drum to a free field response is important for recreating the perception of spatiality and obtaining the correct timbre. Unlike loudspeakers, headphones are generally not designed to have a flat frequency response but instead should compensate for the spectral coloration caused by the sound path to the ear. For correct headphone reproduction it is essential to control the sound pressure at the listener's ears, and there is no general consensus about the optimal transfer function and equalization of headphones. A great multitude of different headphone models can be derived to model playback through different types of headphones (e.g., open, closed, earbuds, in-ear monitors, hearing aids, and so on), and different directional placements. The creation and distribution of such models can be a challenge in environments that feature different audio playback scenarios, such as different client devices (e.g., mobile phones, portable or desktop computers, gaming consoles, and so on), as well as audio content (e.g., music, games, dialog, environmental noise, and so on).
What is needed, therefore, is an equalization system that enhances the perceptual quality and spatial representation of object-based audio content for playback through headphones. What is further needed is a system for efficiently defining and distributing headphone models for a variety of different headphone types and listening environments.
The subject matter discussed in the background section should not be assumed to be prior art merely as a result of its mention in the background section. Similarly, a problem mentioned in the background section or associated with the subject matter of the background section should not be assumed to have been previously recognized in the prior art. The subject matter in the background section merely represents different approaches, which in and of themselves may also be inventions.
Embodiments are described for systems and methods for designing a filter in a magnitude domain for filtering function over a frequency domain to compensate for directional cues for the left and right ears of the listening subject as a function of virtual source angles during headphone virtual sound reproduction by obtaining blocked ear canal and open ear canal transfer functions for loudspeakers placed in a room, obtaining an open ear canal transfer function for a headphone placed on a listening subject, and dividing the loudspeaker transfer functions by the headphone transfer function to invert a headphone response at the entrance of the ear canal and map the ear canal function from the headphone to free field. The method may further comprise constraining the frequency domain to a frequency range spanning a mid to high frequency range of the audible sound domain, wherein the frequency range is selected based on a degree of variation observed in the ratio due to transverse dimensions of the ear canal relative to the wavelength of sound transmitted to the listening subject. The filter may comprise a time-domain filter designed by modeling a magnitude response and phase using one of: a linear-phase design or minimum phase design. The smoothing of the magnitude response may by performed by a fractional octave smoothing function, such as either a ⅓ octave smoother or a ⅙ octave smoother.
In this method, the headphone is configured to playback audio content rendered through a digital audio processing system, and comprising channel-based audio and object-based audio including spatial cues for reproducing an intended location of a corresponding sound source in three-dimensional space relative to the listening subject. The method may comprise a measurement process in which the listening subject comprises a head and torso (HATS) manikin, the method further comprising: placing the manikin centrally in the room surrounded by the loudspeakers; placing the headphones on the manikin; transmitting acoustic signals through the loudspeakers and headphones for reception by microphones placed in or proximate the headphones; deriving measurements of the transfer functions by deconvolving the received acoustic signals with the transmitted signals to obtain binaural room impulse responses (BRIRs) for the loudspeaker blocked ear canal and open ear canal transfer functions; and converting the BRIRs to gated head related transfer function (HTRF) impulses. The method may also comprise placing subminiature microphones in cylindrical foam inserts placed in ear canal entrances of the manikin; measuring headphone sound response through the subminiature microphones; and correcting the headphone sound response to match a flat frequency response pressure microphone through a fractional octave smoothing and minimum-phase equalization component. The method may yet further comprise measuring a Headphone-Ear-Transfer-Function for each of a plurality of headphones by placing a selected headphone is on the manikin a plurality of times each; measuring a transfer function/impulse response for both ears for both ears of the manikin for each placement; and deriving an average response by RMS (root mean squared) averaging the magnitude frequency response of both ears and all placements for each respective headphone to generate a single headphone model for each headphone. The fractional (n) octave smoothing may be performed by one of: RMS averaging all the frequency components over a sliding-frequency, 1/n octave frequency interval or by a weighted RMS average, where the weighting is a sliding-frequency, prototypical 1/n octave frequency filter shape.
In an embodiment, the method comprises storing each headphone model in a networked storage device accessible to client computers and mobile devices over a network, and downloading a requested headphone model to a target client device upon request by the client device. The networked storage device may comprise a cloud-based server and storage system. The requested headphone model may be selected from a user of the client device through a selection application configured to allow the user to identify and download an appropriate headphone model; or it may be determined by automatically detecting a make and model of headphone attached to the client device, and downloading a respective headphone model as the requested headphone model based on the detected make and model of headphone, the headphone comprising one of an analog headphone and a digital headphone. The automatic detection may be performed by one of: measuring electrical characteristics of the analog headphone and comparing to known profiled electrical characteristics to identify a make and type of analog headphone, and using digital metadata definitions of the digital headphone to identify a make and type of digital headphone.
In the method, the client device comprises one of a client computing device, or a mobile communication device, and wherein the method further comprises applying the downloaded headphone model to a virtualizer that renders audio data through the headphones to the user.
Embodiments are further directed to a method comprising: deriving a base filter transfer curve for a headphone over a frequency domain to compensate for directional cues for the left and right ears of the listening subject as a function of virtual source angles during headphone virtual sound reproduction by obtaining blocked ear canal and open ear canal transfer functions for loudspeakers, obtaining an open ear canal transfer function for the headphone, and dividing the loudspeaker transfer functions by the headphone transfer function; deriving additional filter transfer curves for the headphone by changing placement of the headphone relative to a listening device; deriving an average response for the headphone by RMS (root mean squared) averaging the magnitude frequency response of the base filter transfer curve and additional filter transfer curves to generate a single headphone model for each headphone; and applying the average response to a virtualizer for rendering of audio content to a listener through the headphones.
Embodiments are yet further directed to a system comprising an audio renderer rendering audio for playback, a headphone coupled to the audio renderer receiving the rendered audio through a virtualizer function, and a memory storing a filter for use by the headphone, the filter configured to compensate for directional cues for the left and right ears of a listener as a function of virtual source angles during headphone virtual sound reproduction by obtaining blocked ear canal and open ear canal transfer functions for loudspeakers, obtaining an open ear canal transfer function for the headphone, and dividing the loudspeaker transfer functions by the headphone transfer function. The filter can be derived using an offline process and stored in a database accessible to a product or in memory in the product, and applied by a processor in a device connected to the headphones. Alternatively, the filters may be loaded into memory integrated in the headphone that includes resident processing and/or virtualizer componentry.
Embodiments are further directed to systems and articles of manufacture that perform or embody processing commands that perform or implement the above-described method acts.
In the following drawings like reference numbers are used to refer to like elements. Although the following figures depict various examples, the one or more implementations are not limited to the examples depicted in the figures.
Systems and methods are described for virtual rendering of object-based audio over headphones, and impedance matching and equalization system for headphone surround rendering, though applications are not so limited. Aspects of the one or more embodiments described herein may be implemented in an audio or audio-visual system that processes source audio information in a mixing, rendering and playback system that includes one or more computers or processing devices executing software instructions. Any of the described embodiments may be used alone or together with one another in any combination. Although various embodiments may have been motivated by various deficiencies with the prior art, which may be discussed or alluded to in one or more places in the specification, the embodiments do not necessarily address any of these deficiencies. In other words, different embodiments may address different deficiencies that may be discussed in the specification. Some embodiments may only partially address some deficiencies or just one deficiency that may be discussed in the specification, and some embodiments may not address any of these deficiencies.
Embodiments are directed to an audio rendering and processing system including impedance filter and equalizer components that optimize the playback of object and/or channel-based audio over headphones. Such a system may be used in conjunction with an audio source that includes authoring tools to create audio content, or an interface that receives pre-produced audio content.
In an embodiment, the audio processed by the system may comprise channel-based audio, object-based audio or object and channel-based audio (e.g., hybrid or adaptive audio). The audio comprises or is associated with metadata that dictates how the audio is rendered for playback on specific endpoint devices and listening environments. Channel-based audio generally refers to an audio signal plus metadata in which the position is coded as a channel identifier, where the audio is formatted for playback through a pre-defined set of speaker zones with associated nominal surround-sound locations, e.g., 5.1, 7.1, and so on; and object-based means one or more audio channels with a parametric source description, such as apparent source position (e.g., 3D coordinates), apparent source width, etc. The term “adaptive audio” may be used to mean channel-based and/or object-based audio signals plus metadata that renders the audio signals based on the playback environment using an audio stream plus metadata in which the position is coded as a 3D position in space. In general, the listening environment may be any open, partially enclosed, or fully enclosed area, such as a room, but embodiments described herein are generally directed to playback through headphones or other close proximity endpoint devices. Audio objects can be considered as groups of sound elements that may be perceived to emanate from a particular physical location or locations in the environment, and such objects can be static or dynamic. The audio objects are controlled by metadata, which among other things, details the position of the sound at a given point in time, and upon playback they are rendered according to the positional metadata. In a hybrid audio system, channel-based content (e.g., ‘beds’) may be processed in addition to audio objects, where beds are effectively channel-based sub-mixes or stems. These can be delivered for final playback (rendering) and can be created in different channel-based configurations such as 5.1, 7.1.
As shown in
In an embodiment, the audio content from authoring tool 102 includes stereo or channel based audio (e.g., 5.1 or 7.1 surround sound) in addition to object-based audio. For the embodiment of
It should be noted that the components of
In an embodiment, the rendering system of
In spatial audio reproduction, certain sound source cues are virtualized. For example, sounds intended to be heard from behind the listeners may be generated by speakers physically located behind them, and as such, all of the listeners perceive these sounds as coming from behind. With virtual spatial rendering over headphones, on the other hand, perception of audio from behind is controlled by head related transfer functions (HRTF) that are used to generate the binaural signal. In an embodiment, the metadata-based headphone processing system 100 may include certain HRTF modeling mechanisms. The foundation of such a system generally builds upon the structural model of the head and torso. This approach allows algorithms to be built upon the core model in a modular approach. In this algorithm, the modular algorithms are referred to as ‘tools.’ In addition to providing ITD and ILD cues, the model approach provides a point of reference with respect to the position of the ears on the head, and more broadly to the tools that are built upon the model. The system could be tuned or modified according to anthropometric features of the user. Other benefits of the modular approach allow for accentuating certain features in order to amplify specific spatial cues. For instance, certain cues could be exaggerated beyond what an acoustic binaural filter would impart to an individual.
As illustrated in
In general, the equalization function computes the Fast Fourier Transform (FFT) of each response and performs an RMS (root-mean squared) averaging of the derived response. The responses may be variable, octave smoothed, ERB smoothed, etc. The process then computes the inversion, |F(ω)|, of the RMS average with constraints on the limits (+/−x dB) of the inversion magnitude response at mid- and high-frequencies. The process then determines the time-domain filter.
The process then computes the FFT for each impulse response, block 404, and performs an RMS averaging of the derived magnitude response, block 406. The responses may be smoothed (⅓ octave, ERB etc.). In block 408, the computes the filter value, |F(ω)|, by inverting the RMS average with constraints on the limits+/−x dB of the inversion magnitude response. The process then determines the time-domain filter by modeling the magnitude and phase using either a linear-phase (frequency sampling) or minimum phase design.
The post-process may also include a closed-to-open transform function to provide an impedance matching filter function 304. This pressure-division-ratio (PDR) method involves designing a transform to match the acoustical impedance between eardrum and free-field for closed-back headphones with modifications in terms of how the measurements are obtained for free-field sound transmission as a function of direction of arrival first-arriving sound. This indirectly enables matching the ear-drum pressure signals between closed-back headphones and free-field equivalent conditions without requiring complicated eardrum measurements. In an embodiment, a Pressure-Division-Ratio (PDR) for synthesis of impedance matching filter is used. The method involves designing a transform to match the acoustical impedance between ear-drum and free-field for closed-back headphones in particular. The modifications described below are in terms of how the measurements are obtained for free-field sound transmission expressed as function of direction of arrival of first-arriving sound.
For this model, the ratio of P2(ω)/P1(ω) is calculated as follows:
In an embodiment, a headphone sound transmission (headphone acoustical impedance analog model) is used.
For this model, the ratio of P5(ω)/P4(ω) is calculated as follows:
The value P4(ω) is measured at the entrance of the blocked ear canal with a headphone (RMS averaged) steady-state measurement. The measurement of P5(ω) can be done at entrance to ear canal or at distance X mm inside ear canal (or at eardrum) from opening for same headphone placement used for measuring P4(ω). The PDR is computed for both the left and right ears using Eq. 1 below:
PDR(ω,θ)=P2,direct(ω,θ)/P1,direct(ω,θ)÷P5(ω)/P4(ω) (1)
The PDR is computed for both the left and right ears. The filter is then applied in cascade with the equalization filter designed for the corresponding channel/driver (left or right) of the headphone (where the left headphone driver signal delivers audio to the left-L ear, and the right headphone driver delivers audio to the right-R ear). Accordingly, with the knowledge that the two headphone drivers are matched, Eq. 1 can be recast as PDR values associated with the left or right ear:
PDRL(ω,θ)=P2,direct,L(ω,θ)/P1,direct,L(ω,θ)÷P5(ω)/P4(ω) (2a)
PDRR(ω,θ)=P2,direct,R(ω,θ)/P1,direct,R(ω,θ)÷P5(ω)/P4(ω) (2b)
Equations (2a) and (2b) can be combined using the logical-OR (V) expression as:
PDRLVR(ω,θ)=P2,direct,LVR(ω,θ)/P1,direct,LVR(ω,θ)÷P5(ω)/P4(ω) (3b)
As shown in
For open-back headphones, in theory, the acoustical impedance match between free-field and ear-drum and between headphone and ear-drum should be close to identical since the headphone impedance approximates the radiation impedance for “open” condition. This would result in a unity PDR.
As found through the investigation, there is a directional element to the PDR from measurements obtained from an ITU loudspeaker setup (with the ITU setup being an example). This directional aspect manifests as different PDRs for the ipsilateral and contralateral ears as well as differences in PDRs for different channels (resulting in coupling differences by the individual ear-drums to source at angle θ in the free-field, with the angle θ being measured at center of head). The center loudspeaker exhibits a smaller difference in PDR between the ipsilateral and contralateral ears. The angular dependence is captured in a modified nomenclature of PDR(ω,θ). Accordingly, each of the headphone virtualized signals corresponding to a given channel/loudspeaker to the ipsi/contra-ear would need to be transformed by the corresponding ipsilateral and contralateral PDRs through the impedance filter associated with the angle of the loudspeaker.
In an embodiment, the impedance filter can be normalized to a hold amplitude value at higher frequencies to reduce the effect of non-uniform transmission associated with variability in headphone placements. Specifically, the amplitude is held at the amplitude of the bin value corresponding to the boundary frequencies, x and y Hz or to a mean amplitude value in between x and y Hz (where the interval between x and y Hz is the frequency region where PDR variations are observed). The smoothing may be done using n-th octave or ERB or variable octave. In the examples shown, the smoothing is done by a ⅓rd octave smoother.
The closed-to-open transform |G(ω)| to give matched eardrum signals (matching between headphone and free-field) is expressed as:
G(ω,θ)=F|(ω)∥PDR(ω,θ)∥M(ω)|−1
where |M(ω)|−1 is the inverted microphone amplitude response. For
For purposes of comparison with the open-back headphone case,
In an embodiment, the synthesis of the impedance matching filter is performed using ear-canal mapping from the headphone to the free-field and headphone entrance to ear canal transfer function inversion. This is essentially a modification to the PDR method described above, and is a more realistic analogy for the synthesis process in most cases, since it does not involve a blocked canal measurement for the headphone. Measurements show that this approach using filters as obtained using the calculations of Eqs. 4a and 4b below are preferred over the above-described method for various content.
PressuretransformL(ω,θ)=P2,direct,L(ω,θ)/P1,direct,L(ω,θ)÷P5(ω) (4a)
PressuretransformR(ω,θ)=P2,direct,R(ω,θ)/P1,direct,R(ω,θ)÷P5(ω) (4b)
The denominator term (P5(ω)) of each of Eqs. 4a and 4b only have an open ear transfer function, and not the blocked ear transfer function. Directional dependence is maintained because the loudspeaker term is maintained. The denominator term equalizes the ear-drum measurement of the headphone. Specifically, the eardrum measurement of the headphone is represented as:
P
5(ω)=(Pd(ω)+Pr(ω))hp-ecPec-ed(ω) (5)
Note that the numerator in each of Eqs. 4a and 4b involves the pressure transform from entrance of ear-canal to ear-drum in a free-field condition, and the denominator includes the pressure transform from entrance of ear-canal to ear-drum, Pec-ed(ω) in headphone condition of Eq. 3 (in addition to the headphone transfer function measure at the entrance to ear canal, the direct and reflected response, (Pd(ω)+Pr(ω))hp-ec). The ratio in Eqs. 4a and 4b inverts the headphone response at the entrance of the ear canal and maps the ear-canal function from the headphone to free field. It should be noted that the correction is constrained to only the mid-frequency to high-frequency region since this region is where the largest variation is observed in the ratio due to the transverse dimensions of the ear canal relative to the wavelength of the sound. This region was defined by determining the location of the first two resonances in a tube (closed at one end) using the empirical formula for a quarter-wave resonator (a tube closed at one end). For an average ear-canal the diameter is d=2r˜8 mm, the length L is ˜25 mm, which translates to frequencies of:
f
n
=nc/4(L+8r/3π) (n=1,3) f1≈3 kHz, f2≈10 kHz
Note there are other equations such as the simplified quarter-wavelength equations and giving similar frequencies since L>>(8r/3π), such as:
f
n
=nc/4(L) (n=1,3) f1≈3 kHz, f2≈10 kHz,
The binaural room impulse response (BRIR) transfer functions for the blocked canal and ear drum conditions were obtained by placing a HATS manikin in the center of a room of a certain size (e.g., 14.2′ wide by 17.6′ long by 10.6′ high) surrounded by the source loudspeakers. Similarly, the headphone measurements were made by placing the headphones on the manikin. The manikin ears were set at a specific height (e.g., 3.5′) from the floor and the acoustic centers of the loudspeakers were set at approximately that same height and a set distance (e.g., 5′) from the center of the manikin head. In a specific example configuration, seven horizontal loudspeakers were placed a 0°, ±30°, ±90°, and ±135° azimuth, at 0° elevation, while two height loudspeakers were placed at ±90° azimuth and 63° elevation. Other speaker configurations and orientations are also possible.
The measurements of the transfer functions were made by deconvolution of the received acoustic signals with the source four-second long exponential sweep in a 5.46 second long file. The BRIRs were trimmed to 32768 samples long and then further converted to head-related transfer function (HRTF) impulses by time gating the BRIRs to only include the first two milliseconds from the direct arrival sound, followed by 2.5 milliseconds of fade down interval.
Two measurements were made for each source loudspeaker location and headphone fitting. First the internal “ear drum” microphones of the manikin were used for the ear drum measurements. Next, the blocked measurements were made by the use of subminiature microphones (e.g., Sonion 8002MP) placed in small cylindrical foam inserts so that both microphone diaphragms were flush with the manikin conchae and completely sealing the manikin ear canal entrances. The responses of these microphones were also corrected to match a flat frequency response pressure microphone (e.g., B&K ⅛th 4138) via ⅓-octave smoothed, minimum-phase equalization covering the 50-15,000 Hz frequency range.
With regard to the test data measurements and filter design, the divisions between loudspeaker and headphone measurements, leads to a filter in the magnitude domain. The filter is designed over frequency domain [x1, x2] Hz. The filter is constrained in the range (y-axis) to be set at a value of 20*log 10(abs(H(x1))) for all frequencies x<x1 through DC, and is constrained to a value of 20*log 10(abs(H(x2))) for all frequencies x>x2 through Nyquist. Other options are also possible, and not precluded by the specific example values provided herein, such as constraining to 0 dB, constraining to the mean value between x1 and x2 or between 500 Hz and 2 kHz. One example case keeps the values x1 and x2 as 500 Hz and 9 kHz respectively. As can be appreciated by those of ordinary skill in the art, there can be multiple ways to design the filter in the time domain.
After constraining, proper bins are set to values above the Nyquist rate before the inverse FFT process. A frequency sampling approach (e.g., fir2 in matlab) could be used to approximate the frequency response from DC to Nyquist.
In an example embodiment, the basic measurement process comprises measuring the transfer function embodied by a 48 kHz sample rate impulse response. This impulse response is measured by the use of a four-second exponential chirp in a 5.46-second file, where the measured signal is deconvolved with the source signal to result in the impulse response. This impulse response is trimmed to result in a 32768-sample impulse response where the direct arrival impulse is located a few hundred samples from the beginning of the source file. The source file is used to either drive each channel of the headphone or the appropriate loudspeaker, while the measured signal is taken from the internal “ear drum” or blocked-canal microphone in a HATS manikin (e.g., B&K 4128 HATS manikin). The magnitude frequency response is measured by taking the Fast Fourier Transform (FFT) of the impulse response and finding the magnitude component of the FFT frequency bins.
For the measurement of the Headphone-Ear-Transfer-Function P5(ω), a selected headphone is placed on the HATS manikin multiple times or fittings and the transfer function/impulse response measured for both ears. An average response is obtained by RMS averaging the magnitude frequency response of both ears and all fittings for that particular headphone. Fractional-octave smoothing (e.g., ⅓ octave smoothing) is performed by RMS averaging all the frequency components over a sliding-frequency, ⅓ octave frequency interval or by a weighted RMS average, where the weighting can be a sliding-frequency, prototypical ⅓ octave frequency filter shape.
For the measurement of the Head-Related-Transfer-Functions (HRTFs) to the Ear Drum P2(ω) or Blocked Ear Canal P1(ω), the HATS manikin is placed in the center of a room, away from the walls, ceiling, and floor surfaces. Loudspeakers are individually driven by the source signal and then signals at the HATS “ear drum” microphones are used to derive the “Ear Drum” impulse responses for both ears. Alternately, the transfer functions for the blocked canal condition are obtained by placing a foam plug at the ear canal entrance and a small microphone in the center, where both the microphone diaphragm and the foam plug surface are flush with the manikin conchae. These microphones are equalized to be flat over the audible frequency range and the signals from these microphones are combined with the source signals to create the blocked canal impulse responses. These impulse responses are converted to HRTFs by removing all room reflections by only including the first two millisecond time interval after the first arrival sounds, followed by a 2.5 millisecond fade down to zero.
In an embodiment, an automated process is implemented that allows for detection and identification of headphone model/make and which would enable download of appropriate headphone filter coefficients. The device connected to a host could be identified based on manufacturer, make. Such a detection and identification protocol may be provided by the communication system coupling the headphones to the system, such as through USB bus, Apple Lightning connector, and so on. For this embodiment, a device descriptor table using class codes for various interfaces and devices may be used to specify product IDs, vendors, manufacturers, versions, serial numbers, and other relevant product information.
For the embodiment of
In one embodiment the filter models can be derived using an offline process and stored in a database accessible to a product or in memory in the product, and applied by a processor in a device connected to the headphones 1210 (e.g., virtualizer 1208). Alternatively, the filters may be applied to a headphone set that includes resident processing and/or virtualizer componentry, such as headphone set 1220, which is a headphone that includes certain on-board circuitry and memory 1221 sufficient to support and execute downloaded filters and virtualization, rendering or post-processing operations.
Aspects of the methods and systems described herein may be implemented in an appropriate computer-based sound processing network environment for processing digital or digitized audio files. Portions of the adaptive audio system may include one or more networks that comprise any desired number of individual machines, including one or more routers (not shown) that serve to buffer and route the data transmitted among the computers. Such a network may be built on various different network protocols, and may be the Internet, a Wide Area Network (WAN), a Local Area Network (LAN), or any combination thereof. In an embodiment in which the network comprises the Internet, one or more machines may be configured to access the Internet through web browser programs.
One or more of the components, blocks, processes or other functional components may be implemented through a computer program that controls execution of a processor-based computing device of the system. It should also be noted that the various functions disclosed herein may be described using any number of combinations of hardware, firmware, and/or as data and/or instructions embodied in various machine-readable or computer-readable media, in terms of their behavioral, register transfer, logic component, and/or other characteristics. Computer-readable media in which such formatted data and/or instructions may be embodied include, but are not limited to, physical (non-transitory), non-volatile storage media in various forms, such as optical, magnetic or semiconductor storage media.
Unless the context clearly requires otherwise, throughout the description and the claims, the words “comprise,” “comprising,” and the like are to be construed in an inclusive sense as opposed to an exclusive or exhaustive sense; that is to say, in a sense of “including, but not limited to.” Words using the singular or plural number also include the plural or singular number respectively. Additionally, the words “herein,” “hereunder,” “above,” “below,” and words of similar import refer to this application as a whole and not to any particular portions of this application. When the word “or” is used in reference to a list of two or more items, that word covers all of the following interpretations of the word: any of the items in the list, all of the items in the list and any combination of the items in the list.
While one or more implementations have been described by way of example and in terms of the specific embodiments, it is to be understood that one or more implementations are not limited to the disclosed embodiments. To the contrary, it is intended to cover various modifications and similar arrangements as would be apparent to those skilled in the art. Therefore, the scope of the appended claims should be accorded the broadest interpretation so as to encompass all such modifications and similar arrangements.
This application claims priority to U.S. Provisional Patent Application No. 62/072,953, filed on Oct. 30, 2014, which is hereby incorporated by reference in its entirety.
Filing Document | Filing Date | Country | Kind |
---|---|---|---|
PCT/US15/57906 | 10/28/2015 | WO | 00 |
Number | Date | Country | |
---|---|---|---|
62072953 | Oct 2014 | US |