The invention relates to configuring the playback of audio via a home audio playback system, where the audio comprises one or more audio objects.
A typical home audio playback system is arranged to receive and play back audio in a home listening environment. Such a home audio playback system may comprise an Audio Video Receiver (AVR) connected to multiple speakers in a surround sound configuration so as to play back audio via the speakers, e.g. in a home living room or home cinema room. For instance, the AVR may be connected to six speakers in a 5.1 surround sound configuration or eight speakers in a 7.1 surround sound configuration. That is, such an AVR may be configured to playback audio via 6, 8, or in the future even more speaker channels.
The played back audio may be based on a received object-based audio program. The object-based audio program may comprise many different audio objects, e.g. up to 128 different audio objects in some modern formats.
Clearly, if a home audio playback system having 6, 8 or even more speaker channels is to play back audio from an object-based audio program comprising up to 128 audio objects, then there will be many possible permutations of which of the audio objects are to be played back and via which of the speaker channels.
The present disclosure provides a method for configuring playback of audio via a home audio playback system, the audio comprising one or more audio objects, the method comprising:
Thus, in the method it is suggested to make it possible for the user to choose an apparent playback position within the listening environment where the user wants the sound to appear.
In an embodiment the positioning area comprises a first visual element with a visual indication that the at least one of the one or more control objects has been selected by the user. When a control object is selected in an object selection list area e.g. an icon representing that control object is shown in the positioning area.
The positioning area may also comprise a second visual element with a visual indication that the movable control object is movable, or wherein the first visual element comprises the visual indication that the movable control object is movable. For example the control object can have a specific shape when it is movable or just being movable along a circumference of a circle.
The first visual element may also comprise a visual indication when the control object is active. A control object is active when the audio object that the control object represents carries sound. For example when a commentator is speaking the control object is active. An icon that represents the control object on the positioning area may have a special color or highlighting when it is active. This makes it easy for the user to identify the control objects and move selected control objects to certain places in the listening environment.
In an embodiment, the positioning area comprises a further visual element arranged to indicate at least one valid location with respect to the visual representation of where the one or more speakers are located, the valid location being a location which the movable control object can occupy. Depending on the speaker set-up, the user can see on the positioning area of the GUI where it would be possible to place the audio objects. E.g it may be along a circumference of a circle.
Further it is possible to determine a respective content of control object of the, or each of the, audio objects. In one embodiment the respective content is any one of:
In a further embodiment, the graphical user interface further comprises an object selection list area, wherein the object selection list area comprises at least one selectable control object, the at least one selectable control object being configured to allow the user to enable playback of one or more of the control objects. If the user enables control objects in the object selection list area they will show up in the positioning area.
The object selection list area also may comprise at least one further selectable control object, the at least one further selectable control object being configured to allow the user to select a playback volume of the one or more enabled control objects.
The object selection list area may comprise plural control objects, and the control objects are arranged into plural groups in the object selection list area. The number of control objects and the allocation of control objects to the groups are based on the respective content of the control objects. E.g. the user may find all audio objects with a certain audio content under a common group.
In one embodiment the home audio playback system receives information related to the audio, the audio-related information comprising resources specific for a current program of the audio, wherein the resources comprise images for object icons to be displayed in association with the control object(s) in the object selection list area and/or in the positioning area. The resources for a current program are downloaded through a file transfer protocol (FTP) link.
This embodiment makes it possible to have pre-determined programs for each sporting event. In for example a motor sports event, a specific icon can be team radio.
The home audio playback system may comprise two or more speakers which are arranged according to one of:
Further embodiments comprise determining, from the input signal, a user command to store the selection of the one of the control objects as a preset but can also comprise a user command to store the user-desired playback position as a preset.
Further embodiment comprises communicating the configuration signal to the home audio playback system. This can be done by using an Internet-based protocol.
A further embodiment relates to a method of configuring playback of audio via a home audio playback system, the audio comprising two or more audio objects, the method comprising:
This method makes it possible to use only one control object for effecting two actions. When moving the control object it will result in an increase of the volume level of certain pre-selected audio objects and at the same time decrease the volume level of other pre-selected audio objects. If the user wants to increase intelligibility of foreground objects and decrease volume of background objects, the user increases the volume level for these foreground objects with the balance adjuster by moving the balance adjuster to the first position. This will at the same time result in a decrease in volume level for the background objects.
In one embodiment the movable control object may be a slider which is movable in a straight line between the first and second position. A slider would make it easy and flexible to only need one control object for effecting two actions.
In one embodiment it is possible to determine from the input signal, a user command to store:
This embodiment would make it comfortable to know where the audio objects are positioned in the listening environment at a specific preset and what volume levels the different audio objects have for a specific preset.
This embodiment would work well for example with two or more speakers which are arranged according to one of:
The embodiments as recited above for the methods are also applicable to the respective apparatus.
In a further embodiment, the corresponding apparatus would be suitable to receive the input signal via wireless communications channel such as IR-channel and transmit the output signal via a wired channel such as for example HDMI, VGA, DVI or SCART cable.
In a further embodiment of the apparatus, the display and the input device are integrated in a touch screen.
The objective of the present disclosure is also achieved by a home audio playback system as claimed in claim 57 and a broadcasting system as claimed in claim 58.
Throughout this disclosure including in the claims, the expression “system” is used in a broad sense to denote a device, system, or subsystem. For example, a subsystem that implements a decoder may be referred to as a decoder system, and a system including such a subsystem (e.g., a system that generates X output signals in response to multiple inputs, in which the subsystem generates M of the inputs and the other X-M inputs are received from an external source) may also be referred to as a decoder system.
Throughout this disclosure including in the claims, the term “processor” is used in a broad sense to denote a system or device programmable or otherwise configurable (e.g., with software or firmware) to perform operations on data (e.g., audio, or video or other image data). Examples of processors include a field-programmable gate array (or other configurable integrated circuit or chip set), a digital signal processor programmed and/or otherwise configured to perform pipelined processing on audio or other sound data, a programmable general purpose processor or computer, and a programmable microprocessor chip or chip set.
Throughout this disclosure including in the claims, the expression “audio video receiver” (or “AVR”) denotes a receiver in a class of consumer electronics equipment used to control playback of audio and video content, for example in a home theater.
Throughout this disclosure including in the claims, the expression “soundbar” denotes a device which is a type of consumer electronics equipment (typically installed in a home theater system), and which includes at least one speaker (typically, at least two speakers) and a subsystem for rendering audio for playback by each included speaker (or for playback by each included speaker and at least one additional speaker external to the soundbar).
Throughout this disclosure including in the claims, the expression “metadata” (e.g., as in the expression “processing state metadata”) refers to separate and different data from corresponding audio data (audio content of a bitstream which also includes metadata). Metadata is associated with audio data, and indicates at least one feature or characteristic of the audio data (e.g., what type(s) of processing have already been performed, or should be performed, on the audio data, or the trajectory of an object indicated by the audio data). The association of the metadata with the audio data is time-synchronous. Thus, present (most recently received or updated) metadata may indicate that the corresponding audio data contemporaneously has an indicated feature and/or comprises the results of an indicated type of audio data processing.
Throughout this disclosure including in the claims, the following expressions have the following definitions:
With reference to
In
In this embodiment, a handheld computing device 204 interacts with the home playback system 200. The handheld computing device 204 is preferably a tablet computer, a mobile phone or the like.
The TV 201 typically communicates with the STB 202 and the AVR 203 through a wired connection or a wireless connection. The wired connection is preferably via a cable like an HDMI (High Definition Multimedia Interface), VGA (Video Graphics Array), SCART (Syndicat des Constructeurs d'Appareils Radiorécepteurs et Téléviseurs), or DVI (Digital Visual Interface) cable, or similar. The speakers 205 may have a wired or wireless connection to the AVR 203. The handheld computing device 204 may have a wired or wireless connection to the STB 202.
As shown in
A downstream renderer 309 may be part of the AVR 203 (or a soundbar), and is configured to render audio for playback to the speakers 205 in the home playback system.
The user control unit 303 is preferably the handheld computing device 204.
The decoder 301 receives audio related data in a bit stream, e.g. an AC-3 encoded bit-stream.
The audio comprises audio objects. The bit-stream comprises data informing of available audio objects in the bit-stream.
With the user control unit 303 a user can select which audio objects to render. The user control unit 303 may be the handheld computing device 204 which is programmed to implement a graphical user interface (GUI). The GUI may provide to the user a menu of selectable “preset” mixes of objects and speaker channel content. The decoder decodes the channels of the selected audio objects, and outputs to the object processing unit 302 these selected audio object channels and object related metadata corresponding to the selected object channels.
The object processing unit 302 is controlled by control data from the user control unit 303 and object related metadata from decoder 301 and is configured to determine inter alia a spatial position and audio level for each of the selected objects.
The spatial rendering system 304 is configured to render the audio objects from object processing unit 302 for playback by speakers 312 of the home playback system. The spatial rendering system map to available speaker channels, the audio channels which has been selected by object processing unit 302 using the rendering parameters output from object processing unit 302.
From the audio-related data 402 the processor 401 determines 501 presence of audio objects. The processor 401 then determines 502 an audio content for each of the audio objects.
The respective content of the audio objects may be captured audio of any of: commentary, fan crowd, team radio, extras or social media chat. The commentary can be captured audio of a commentator for home fans, for away fans, for radio, or alternative commentaries e.g in different languages. The fan crowd may comprise home, away or neutral crowd noise. The team radio may comprise radio communication between driver and engineer when watching a motor sports event. Extras may comprise stadium announcements (e.g. substitutions of players, emergency information), or a goal flash from other events. Social media chat may comprise text messages exchanged between friends during a game or a race. The text may be converted to speech by using Text-To-Speech (TTS) synthesis.
The processor 401 receives information related to the audio. The processor 401 may use a file transfer protocol (FTP) link to download resources specific for a current program. The current program can be a sporting event e.g. a rugby game, a soccer game or another sporting event. For each of the programs there are predefined resources. The resources are mainly images with icons or bars that are displayed on the GUI.
The processor 401 also obtains system information 403, e.g. by retrieving the system information from a memory. The system information may have been saved to the memory during a discovery phase.
From the system information, playback capability of the user's audio system is received. Depending on the user's audio playback system a speaker configuration of one or more speakers can be determined 503 for the home audio playback system 200.
The speaker configuration can e.g. be any one of: a 2.0 speaker set-up, a 5.1 set-up, a 7.1 set up, a 3D set-up or a soundbar set-up.
The processor 401 is then generates 504 the output signal 404 for an output device. For example the output device can in various embodiments comprise a display. The display may be integrated in a touch screen of the handheld computing device 204.
Depending on the output device the output signal can be transmitted via a wireless communications channel, or wired channel via a HDMI, VGA, SCART or DVI cable.
The output signal 404 can comprise data which is suitable for causing the output device to present to a user an indication of which audio objects are present in the audio. At least part of the data is suitable for causing the output device to generate a graphical user interface (GUI) 600.
In some embodiments the object selection list area 602 may comprise at least one further control object configured to allow the user to select a playback volume of at least one of the audio objects. This further control object need not be in the object selection area 602.
The control objects may be arranged in plural groups in a list in the object selection list area. The number of control objects in the groups, and allocation of control objects to the groups may be based on the respective content of the control objects that is predetermined by a Content creator. The object selection list area may be a scrollable region if there are many control objects such as 16 control objects.
When the control objects are selected in the object selection list area the respective control object will appear in the positioning area. In the positioning area these control objects may be visualized as icons.
As shown in
A 2.0 speaker set-up area is limited to an angle of −/+45 degrees from the center of the listening area. A 5.1 speaker set-up area is limited to a circle, which has an angle of 360 degree from the center of the listening area. A 7.1 speaker set-up area is limited to a circle, which has an angle of 360 degree from the center of the listening area.
A 3D set-up area is limited to half of a sphere of the listening area. A soundbar set-up area is also limited to half of a sphere of the listening area.
The positioning area 601, 700 comprises of at least one movable control object 702 which represents one of the enabled control objects in the object selection list area. This movable control object 702 is movable with respect to the visual representation. As shown in
The current location of the movable control object 702 is selected by the user, as will be discussed below with reference to
The positioning area 601, 700 may comprise a first visual element which is a visual indication that the at least one of the one or more control objects has been selected by the user. E.g. an icon is shown in the positioning area. The first visual element can further comprise a visual indication of whether a control object is movable, or the positioning area can comprise a second visual element with a visual indication that the movable control object is movable. As shown in
The first visual element may also comprise a visual indication when the control object is active. For example, an icon that represents the control object may be highlighted blue, when the audio object carries sound, e.g. wherein a commentator is speaking.
The positioning area 700 may also comprise a further visual element arranged to indicate at least one valid location with respect to the visual representation in the listening environment, the valid location 701 being a location which the movable control object 702 can occupy. It also comprises a visual indication of at least one invalid location, the invalid location being a location which the movable control object cannot occupy. The control objects 702 in the positioning area 700 are movable along the circumference of the circle 701, which may be e.g. displayed in red in order to visually indicate to the user that the movable control object 702 can occupy any point on the circumference. When the user is moving the icon in allowable positions on the circumference of the circle in the positioning area, the icon is typically highlighted with a green/red shadow around the icon.
As shown in
Further the balance adjustment area comprises a second icon at a second position. This second position can be in a left-hand or lower part of a balance adjustment area. The second icon represents one or more of the audio objects which are background objects 802. A movable control object is movable between the first position and the second position whereby a current location of the movable control object can be selected by the user. The current position of the movable control object relative to the first position is representative of a user-selected volume level for the one or more foreground objects 801. The current position of the movable control object relative to the second position is representative of a user-selected volume level for the one or more background objects 802.
The audio objects that are possible to be background 802 and foreground objects 801 are pre-selected by a Content creator through metadata. The metadata is specific dependent on the different type of applications and can e.g. be sent in the bit-stream or be sent as external metadata.
In one embodiment the movable control object may be a slider which is movable along a straight line. If the slider is moved to the right (or upwards) the slider increases audio volume for foreground objects 801 and at the same time decreases audio volume for the background objects 802. The user might for example want to increase intelligibility of foreground objects that he would like to hear better and decrease the volume of background ambience that he still wants to hear but at a lower volume.
If the slider is moved to the left (or downwards) the audio volume increases for the background objects 802 and decreases for foreground objects 801. In this way only one control object is needed for effecting two actions (increasing and decreasing audio volume for pre-selected audio object(s) at the same time).
The volume level of the background 802 and foreground objects 801 can also be stored as presets.
In
In one example of a use scenario, there is a 5.1 speaker configuration for a home playback system as in
Three control objects are selected out of the possible 5 control objects. The respective three control objects have captured audio content of a home team commentator, a social media chat, and home team crowd noise. In this example the 5.1 speaker configuration makes it possible to position the control objects along the circumference of a circle as in
In the positioning area of the GUI the user will see a visual representation of the speaker set-up in the home playback environment. In the object selection list area the user will see the control objects. The control objects may on the object selection list area 602 appear as element bars and on the positioning area as icons. The user can select some of the control objects and these bars may then get blue colored. The selected control objects will then appear as icons on the positioning area 601.
The user may for example position the home team commentator icon along the circumference of the circle to the left of the TV, the social media chat icon along the circumference of the circle to the right of the TV and the team crowd noise icon along the circumference of the circle behind the sofa. The user will then in his living room hear the audio of the home team commentator as it appears from left of the TV, the audio of the social media chat as it appears from the right of the TV and the audio of the home team crowd noise as it appears from behind the sofa. The user can then on the GUI 600 of the handheld device move a slider which controls the audio volume level of the foreground objects (the home team commentator and the social media chat in this example) and the background audio objects (home team crowd noise). If it is desirable to decrease the audio volume of the home crowd noise and at the same time increase the audio volume of the home team commentator and the social media chat, the slider is moved towards the first icon 801.
The processor 401 receives an input signal 405 from the user input device. The input signal comprises data representative of:
The input signal also comprises data indicative for determining a user command 1002 to store the selection of the one of the audio objects as a preset, to store the perceived spatial position relative to the speaker configuration as a preset and/or to store the playback volume level as a preset.
The processor 401 then generates a configuration signal 406, 1003 for configuring the home audio playback system. The configuration signal data is suitable for causing the home audio playback system to selectively play back one or more audio objects.
Said data is suitable for causing the home audio playback system to perform at least one of: playback one of the audio objects according to a user-desired playback position; playback two or more audio objects according to respective user-selected volume levels for one or more foreground objects and one or more background objects.
The configuration signal data is also using presets to store predefined configurations of enabled audio objects, positions and volume for the enabled objects. The different types of presets for the audio object are preferably:
Said data is also suitable to recall presets any time to restore object configuration, and is suitable to communicate with the audio playback system over a protocol as Transmission Control Protocol/Internet Protocol (TCP/IP.)
The present disclosure describes various example embodiments below, with reference to the drawings. It is to be noted that, for purpose of clarity, representations and descriptions of components and processes which are known to those skilled in the art, but which are not necessary to understand the various example embodiments, are omitted in the drawings and the description.
As will be appreciated by one skilled in the art, aspects of the present application may be embodied as a system, a device (e.g., a cellular telephone, a portable media player, a personal computer, a server, a television set-top box, or a digital video recorder, or any other media player), a method or a computer program product. Accordingly, aspects of the present application may take the form of an hardware embodiment, an software embodiment (including firmware, resident software, microcodes, etc.) or an embodiment combining both software and hardware aspects that may all generally be referred to herein as a “circuit,” “module” or “system.” Furthermore, aspects of the present application may take the form of a computer program product embodied in one or more computer readable mediums having computer readable program code embodied thereon.
Any combination of one or more computer readable mediums may be utilized. The computer readable medium may be a computer readable signal medium or a computer readable storage medium. A computer readable storage medium may be, for example, but not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or any suitable combination of the foregoing. More specific examples (a non-exhaustive list) of the computer readable storage medium would include the following: an electrical connection having one or more wires, a portable computer diskette, a hard disk, a random access memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or Flash memory), an optical fiber, a portable compact disc read-only memory (CD-ROM), an optical storage device, a magnetic storage device, or any suitable combination of the foregoing. In the context of this document, a computer readable storage medium may be any tangible medium that can contain, or store a program for use by or in connection with an instruction execution system, apparatus, or device.
A computer readable signal medium may include a propagated data signal with computer readable program code embodied therein, for example, in baseband or as part of a carrier wave. Such a propagated signal may take any of a variety of forms, including, but not limited to, electro-magnetic or optical signal, or any suitable combination thereof.
A computer readable signal medium may be any computer readable medium that is not a computer readable storage medium and that can communicate, propagate, or transport a program for use by or in connection with an instruction execution system, apparatus, or device.
Program code embodied on a computer readable medium may be transmitted using any appropriate medium, including but not limited to wireless, wired line, optical fiber cable, RF, etc., or any suitable combination of the foregoing.
Computer program code for carrying out operations for aspects of the present application may be written in any combination of one or more programming languages, including an object oriented programming language such as Java, Smalltalk, C++ or the like and conventional procedural programming languages, such as the “C” programming language or similar programming languages. The program code may execute entirely on the user's computer as a stand-alone software package, or partly on the user's computer and partly on a remote computer or entirely on the remote computer or server. In the latter scenario, the remote computer may be connected to the user's computer through any type of network, including a local area network (LAN) or a wide area network (WAN), or the connection may be made to an external computer (for example, through the Internet using an Internet Service Provider).
Aspects of the present application are described below with reference to flowchart illustrations and/or block diagrams of methods, apparatus (systems) and computer program products according to embodiments of the application. It will be understood that each block of the flowchart illustrations and/or block diagrams, and combinations of blocks in the flowchart illustrations and/or block diagrams, can be implemented by computer program instructions. These computer program instructions may be provided to a processor of a general purpose computer, special purpose computer, or other programmable data processing apparatus to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable data processing apparatus, create means for implementing the functions/acts specified in the flowchart and/or block diagram block or blocks.
These computer program instructions may also be stored in a computer readable medium that can direct a computer, other programmable data processing apparatus, or other devices to function in a particular manner, such that the instructions stored in the computer readable medium produce an article of manufacture including instructions which implement the function/act specified in the flowchart and/or block diagram block or blocks.
The computer program instructions may also be loaded onto a computer, other programmable data processing apparatus, or other devices to cause a series of operational operations to be performed on the computer, other programmable apparatus or other devices to produce a computer implemented process such that the instructions which execute on the computer or other programmable apparatus provide processes for implementing the functions/acts specified in the flowchart and/or block diagram block or blocks.
While implementations have been described by way of example and in terms of exemplary specific embodiments, it is to be understood that implementations of the invention are not limited to the disclosed embodiments. On the contrary, it is intended to cover various modifications and similar arrangements as would be apparent to those skilled in the art. Therefore, the scope of the appended claims should be accorded the broadest interpretation so as to encompass all such modifications and similar arrangements.
This application claims the benefit of priority to U.S. Provisional Patent Application No. 62/001,193 filed 21 May 2014, which is hereby incorporated by reference in its entirety.
Filing Document | Filing Date | Country | Kind |
---|---|---|---|
PCT/EP2015/061138 | 5/20/2015 | WO | 00 |
Number | Date | Country | |
---|---|---|---|
62001193 | May 2014 | US |