This patent application also makes reference to:
Each of the above stated applications is hereby incorporated herein by reference in its entirety.
Certain embodiments of the invention relate to audio processing. More specifically, certain embodiments of the invention relate to a method and system for adjusting audio based on captured depth information.
Support and demand for video systems that support three-dimensional (3-D) video has increased rapidly in recent years. Both literally and physically, 3-D video provides a whole new way to watch video, in home and in theaters. However, 3-D video systems are still in their infancy in many ways and there is much room for improvement in terms of both cost and performance.
Further limitations and disadvantages of conventional and traditional approaches will become apparent to one of skill in the art, through comparison of such systems with some aspects of the present invention as set forth in the remainder of the present application with reference to the drawings.
A system and/or method is provided for adjusting audio based on captured depth information, substantially as illustrated by and/or described in connection with at least one of the figures, as set forth more completely in the claims.
These and other advantages, aspects and novel features of the present invention, as well as details of an illustrated embodiment thereof, will be more fully understood from the following description and drawings.
Certain embodiments of the invention may be found in a method and system for adjusting based on captured depth information. In various embodiments of the invention, a two-dimensional image may be captured via one or more image sensors of a monoscopic camera and depth information may be captured via one or more depth sensors of the monoscopic camera. The depth sensor may utilize infrared waves transmitted by an infrared emitter of the camera. A video stream may be generated from the captured two-dimensional image data based on the captured depth information. The generated video stream may be a 3-D video stream. Corresponding audio to accompany the generated video stream may be generated based on the captured depth information. The generated corresponding audio may be adjusted based on the captured depth information. A location and/or movement of an object appearing in the captured two-dimensional image data may be determined based on the depth information and based on pixel data of the captured two-dimensional image data. Characteristics of the generated audio, such as volume, frequency, delay, and balance, may be adjusted based on the determined location and/or movement. As utilized herein a “3-D image” refers a stereoscopic image, and “3-D video” refers to stereoscopic video.
The monoscopic camera 102 may capture images via a single viewpoint corresponding to the lens 101c. In this regard, EM waves in the visible spectrum may be focused on one or more image sensors by the lens 101c. The image sensor(s) may capture brightness and/or color information. The captured brightness and/or color information may be represented in any suitable color space such as YCrCb color space or RGB color space. The monoscopic camera 102 may also capture depth information via the lens 101c (and associated optics). For example, the monoscopic cameral 102 may comprise an infrared emitter, an infrared sensor, and associated circuitry operable to determine the distance to objects based on reflected infrared waves. Additional details of the monoscopic camera 102 are described below.
The monoscopic camera 102 may comprise a processor 124, a memory 126, and one or more sensors 128. The processor 124 may comprise suitable logic, circuitry, interfaces, and/or code that may be operable to manage operation of various components of the monoscopic camera 102 and perform various computing and processing tasks. A single processor 124 is utilized only for illustration but the invention is not so limited. In an exemplary embodiment of the invention, various portions of the monoscopic camera 102 depicted in
The processor 104 may comprise suitable logic, circuitry, interfaces, and/or code. The processor 104 may be operable to coordinate operation of the various components of the monoscopic camera 102. The processor 104 may, for example, run an operating system of the monoscopic camera 102 and control communication of information and signals between components of the monoscopic camera 102. The processor 104 may execute instructions stored in the memory 106.
The memory 106 may comprise, for example, DRAM, SRAM, flash memory, a hard drive or other magnetic storage, or any other suitable memory devices. For example, SRAM may be utilized to store data utilized and/or generated by the processor 104 and a hard-drive and/or flash memory may be utilized to store recorded image data and depth data.
The video encoder/decoder 107 may comprise suitable logic, circuitry, interfaces, and/or code. The video encoder/decoder 107 may be operable to process captured color, brightness, and/or depth data to make the data suitable for conveyance to, for example, the display 120 and/or to one or more external devices via the I/O block 114. For example, the video encoder/decoder 107 may convert between, for example, raw RGB or YcrCb pixel values and an MPEG encoding. Although depicted as a separate block 107, the video encoder/decoder 107 may be implemented in the DSP 110.
The depth sensor 108 may comprise suitable logic, circuitry, interfaces, and/or code. The depth sensor 108 may be operable to detect EM waves in the infrared spectrum and determine distance to objects based on reflected infrared waves. In an embodiment of the invention, distance may be determined based on time-of-flight of infrared waves transmitted by the emitter 109 and reflected back to the sensor 108. In an embodiment of the invention, depth may be determined based on distortion of a captured grid.
The audio encoder/decoder 109 may comprise suitable logic, circuitry, interfaces, and/or code. The audio encoder/decoder 109 may be operable to process captured audio data to make the data suitable for conveyance to, for example, the speaker 111 and/or to one or more external devices via the I/O block 114. For example, the video encoder/decoder 107 may convert between, for example, raw pulse-code-modulated audio and an MP3 or AAC encoding. Although depicted as a separate block 109, the audio encoder/decoder 109 may be implemented in the DSP 110.
The digital signal processor (DSP) 110 may comprise suitable logic, circuitry, interfaces, and/or code. The DSP 110 may be operable to perform complex processing of captured image data, captured depth data, and captured audio data. The DSP 110 may be operable to, for example, compress and/or decompress the data, encode and/or decode the data, and/or filter the data to remove noise and/or otherwise improve perceived audio and/or video quality for a listener and/or viewer.
The input/output module 112 may comprise suitable logic, circuitry, interfaces, and/or code that may enable the monoscopic camera 102 to interface with other devices in accordance with one or more standards such as USB, PCI-X, IEEE 1394, HDMI, DisplayPort, and/or analog audio and/or analog video standards. For example, the I/O module 112 may be operable to send and receive signals from the controls 122, output video to the display 120, output audio to a speaker 111, handle audio input from the microphone 113, read from and write to cassettes, flash cards, or other external memory attached to the monoscopic camera 102, and/or output audio and/or video via one or more ports such as a IEEE 1394 or USB port.
The microphone 113 may comprise a transducer and associated logic, circuitry, interfaces, and/or code operable to convert acoustic waves into electrical signals. The microphone 113 may be operable to amplify, equalize, and/or otherwise process captured audio signals. The directionality of the microphone 113 may be controlled electronically and/or mechanically. The monoscopic camera 102 may also be operable to receive audio signals from one or more remotely located microphones.
The image sensor(s) 114 may each comprise suitable logic, circuitry, interfaces, and/or code that may be operable to convert optical signals to electrical signals. Each image sensor 114 may comprise, for example, a charge coupled device (CCD) images sensor or a complimentary metal oxide semiconductor (CMOS) image sensor. Each image sensor 114 may capture 2D brightness and/or color information.
The error protection module 315 may comprise suitable logic, circuitry, interfaces and/or code that may be operable to perform error protection functions for the video monoscopic camera 300. For example, the error protection module 315 may provide error protection to encoded 2D video images and corresponding depth information and/or encoded audio data for transmission to a 3-D video rendering device such as the 3-D video rendering device 204. The error protection module 315 may apply one or more levels of error protections to an encoded 2D video image frame and/or corresponding depth information or data based on one or more of interest within the encoded 2D video image frame.
The optics 116 may comprise various optical devices for conditioning and directing EM waves received via the lens 101c. The optics 116 may direct EM waves in the visible spectrum to the image sensor 114 and direct EM waves in the infrared spectrum to the depth sensor 108. The optics 116 may comprise, for example, one or more lenses, prisms, color filters, and/or mirrors.
The lens 118 may be operable to collect and sufficiently focus electromagnetic waves in the visible and infrared spectra.
The digital display 120 may comprise an LCD, LED, OLED, or other digital display technology on which images recorded via the monoscopic camera 102 may be displayed. In an embodiment of the invention, the digital display 120 may be operable to display 3-D images.
The controls 122 may comprise suitable logic, circuitry, interfaces, and/or code. The controls 122 may enable a user to interact with the monoscopic camera 102. For example, controls for controlling recording and playback. In an embodiment of the invention, the controls 122 may enable a user to select whether the monoscopic camera 102 records and/or outputs video in 2D or 3-D modes.
The optical viewfinder 124 may enable a user to see what the lens 101c “sees,” that is, what is “in frame.”
In operation, the image sensor(s) 114 may capture frames of 2D video. The depth sensor 108 may capture depth information associated with the objects appearing in the video frames. An audio track to accompany the video may be processed and/or generated based on the pixel data of the video and/or based on the captured depth information. In this regard, characteristics of the audio, such as volume, frequency, delay, left-right balance, and front-back balance, may be adjusted based on the captured pixel data and/or depth information. The audio track may comprise, for example, audio captured concurrently with the capturing of the video and depth information and/or may comprise sound effects and/or other audio captured and/or generated separately from the video.
In one exemplary scenario, a object which didn't make any sound while being recorded may appear in the video and audio to be perceived by a viewer of the video as emanating from that object may be added during editing of the video. Characteristics of the added audio may be adjusted based on a location and/or movement of the object. The location and/or movement of the object may be determined based on the pixel data and/or captured depth information.
In another exemplary scenario, the video may comprise images of a sound source and the audio from the sound source may be captured concurrently by the microphone 113. Subsequently, during editing of the video, the camera angle or location of the camera may be artificially manipulated such that the perceived location of the sound source relative to the viewer of the video is different than was the location of the sound source relative to the monoscopic camera 102. Accordingly, the audio may be manipulated such that the origin of the audio, as perceived by a viewer of the video, corresponds to the location of the sound source, as perceived by the viewer of the video. In other words, as the video is manipulated, the audio may be manipulated to ensure that the audio tracks its source as the source moves around relative to the viewer.
In the frame 130, the line weight is used to indicate depth—heavier lines being closer to the viewer. Thus, the object 138 is farthest from the monoscopic camera 102, the object 142 is closest to the monoscopic camera 102 and the object 104 is at an intermediate distance. In various embodiments of the invention, depth information may be mapped to a grayscale, or pseudo-grayscale, image for display to a viewer. Such mapping may be performed, for example, by the DSP 110.
The image in the frame 134 is a conventional 2D image. A viewer of the frame 134, for example, on the display 120 or on a device connected to the monoscopic camera 102 via the I/O module 112, perceives the same distance between himself and each of the objects 138, 140, and 142. That is, each of the objects 138, 140, and 142 each appear to reside on the plane 132.
The image in the frame 136 is a 3-D image. A viewer of the frame 136, for example, on the display 120 or on a device connected to the monoscopic camera 102 via the I/O module 112, perceives the object 138 being furthest from him the object 142 being closest to him, and the object 140 being at an intermediate distance. In this regard, the object 138 appears to be behind the reference plane, the object 140 appears to be on the reference plane, and the object 142 appears to be in front of the reference plane.
Location of the object may be described utilizing the three-dimensional coordinate system depicted in
The graph 406 illustrates control of the overall volume. As the object 402 moves closer to the monoscopic camera 102, and thus the viewer, the overall volume, that is the combined volume of the four audio channels, may be increased.
The graph 408 illustrates control of audio frequency. Audio frequency may be adjusted to simulate a Doppler effect. Thus, as the object 402 is moving toward the monoscopic camera 102, and thus the viewer, the frequency of the signal 404 may be increased and as the object 402 is moving away from the monoscopic camera, and thus the viewer, the frequency may be decreased.
The graph 410 illustrates control of left-right balance. As the object 402 moves from the viewer' left to the viewer's right, the volume of the FL and BL channels may be reduced and the volume of the FR and BR channels may be increased.
The graph 412 illustrates control of the front-back fading. Since the object appears in front of the viewer, the volume of the FL and FR channels may be higher than the volume of the BL and BR channels. However, simulated echo or other ancillary sounds may be added onto the BL and BR channels to provide a more realistic audio experience.
A location of the object may be described utilizing the three-dimensional coordinate system depicted in
The graph 506 illustrates control of the overall volume. As the object 502 moves closer to the monoscopic camera 102, and thus the viewer, the overall volume, that is the combined volume of the four audio channels, may be increased. As the object 502 moves away from the viewer the overall volume may be decreased.
The graph 508 illustrates control of audio frequency. Audio frequency may be adjusted to simulate a Doppler effect. Thus, as the object 502 is moving toward the monoscopic camera 102, and thus the viewer, the frequency of the signal 504 may be increased and as the object 502 is moving away from the monoscopic camera 102, and thus the viewer, the frequency may be decreased.
The graph 510 illustrates control of left-right balance. Because the object appears to the right of the viewer, the volume of the FR and BR channels may be higher than the volume of the FL and BL channels. However, simulated echo or other ancillary sounds may be added onto the FL and FR channels to provide a more realistic audio experience.
The graph 512 illustrates control of the front-back fading. As the object 402 moves from in front of the viewer to behind the viewer, the volume of the FL and FR channels may be reduced and the volume of the BL and BR channels may be increased.
Various aspects of a method and system for adjusting audio based on captured depth information are provided. In an exemplary embodiment of the invention, characteristics of one or more audio signals associated with a video, such as signals 404 and 504, may be adjusted based on pixel data of the video and based on depth information captured while recording the video. The video may be captured via one or more image sensors 114 of a monoscopic camera 102 and the depth information may be captured via one or more depth sensors 108 of the monoscopic camera 102. The depth sensor(s) 108 may utilize infrared waves transmitted by an emitter 109 integrated into the monoscopic camera 102. The captured depth information may be stored in memory separately from the pixel data of the captured video. The captured video may comprise two-dimensional video. The monoscopic camera 102 may be operable to process frames of two-dimensional video, such as the frame 130, utilizing frames of captured depth information, such as the frame 134, to generate frames of a three-dimensional video, such as frame 136. The depth information may be utilized to determine a location and/or movement of an object appearing in the video, such as objects 402 and 502. The location of the object may be stored as one or more three-dimensional coordinates. A volume, frequency, delay, and/or balance of the one or more audio signals may be adjusted based on the determined location and/or movement.
In various embodiments of the invention, two-dimensional image may be captured via the image sensor(s) 114 of the monoscopic camera 102, depth information may be captured via the depth sensor(s) 108 of the monoscopic camera 102, a video stream may generated from the captured two-dimensional image data based on the captured depth information, and corresponding audio may be generated based on the captured depth information. The generated video stream may be a 3-D video stream. The generated audio and video may be combined to generate a multimedia stream. The generated corresponding audio may be adjusted based on the captured depth information.
Other embodiments of the invention may provide a non-transitory computer readable medium and/or storage medium, and/or a non-transitory machine readable medium and/or storage medium, having stored thereon, a machine code and/or a computer program having at least one code section executable by a machine and/or a computer, thereby causing the machine and/or computer to perform the steps as described herein for generating three-dimensional video utilizing a monoscopic camera.
Accordingly, the present invention may be realized in hardware, software, or a combination of hardware and software. The present invention may be realized in a centralized fashion in at least one computer system, or in a distributed fashion where different elements are spread across several interconnected computer systems. Any kind of computer system or other apparatus adapted for carrying out the methods described herein is suited. A typical combination of hardware and software may be a general-purpose computer system with a computer program that, when being loaded and executed, controls the computer system such that it carries out the methods described herein.
The present invention may also be embedded in a computer program product, which comprises all the features enabling the implementation of the methods described herein, and which when loaded in a computer system is able to carry out these methods. Computer program in the present context means any expression, in any language, code or notation, of a set of instructions intended to cause a system having an information processing capability to perform a particular function either directly or after either or both of the following: a) conversion to another language, code or notation; b) reproduction in a different material form.
While the present invention has been described with reference to certain embodiments, it will be understood by those skilled in the art that various changes may be made and equivalents may be substituted without departing from the scope of the present invention. In addition, many modifications may be made to adapt a particular situation or material to the teachings of the present invention without departing from its scope. Therefore, it is intended that the present invention not be limited to the particular embodiment disclosed, but that the present invention will include all embodiments falling within the scope of the appended claims.
This patent application makes reference to, claims priority to and claims benefit from U.S. Provisional Patent Application Ser. No. 61/439,201 filed on Feb. 3, 2011 and U.S. Provisional Patent Application Ser. No. 61/377,867 filed on Aug. 27, 2010. Each of the above stated applications is hereby incorporated herein by reference in its entirety.
Number | Date | Country | |
---|---|---|---|
61439201 | Feb 2011 | US | |
61377867 | Aug 2010 | US |