1. The Field of the Invention
This invention relates to systems, methods, and computer program products related to conversion and presentation of three-dimensional video content.
2. Background and Relevant Art
Three-dimensional (3D) display technology involves presenting two-dimensional images in such a manner that the images appear to the human brain to be 3D. The process typically involves presenting “left” image data to the left eye, and “right” image data to the right eye. When received, the brain perceives this data as a 3D image. 3D display technology generally incorporates the use of a filtering device or blanking device, such as glasses, which filter displayed image data to the correct eye. Filtering devices can be passive, meaning that image data is filtered passively (e.g., by color code or by polarization), or active, meaning that the image data is filtered actively (e.g., by shuttering).
Recently, 3D display devices designed specifically for displaying 3D content have become increasingly popular. These 3D display devices are generally used in connection with active filtering devices (e.g., shuttering glasses) to produce 3D image quality not previously available from traditional display devices. These 3D display devices, however, are often based on proprietary signals and shuttering patterns. For example, one manufacturer of 3D display devices may display right eye information for a frame before left eye information for the frame while another manufacturer may display left eye information for a frame before right eye information for the frame. Other timing and signal protocol differences may exist as well. However, hardware for 3D display devices, including the active shuttering glasses, may be expensive.
The subject matter claimed herein is not limited to embodiments that solve any disadvantages or that operate only in environments such as those described above. Rather, this background is only provided to illustrate one exemplary technology area where some embodiments described herein may be practiced.
One embodiment illustrated herein is directed to a method of controlling a shuttering device. The method includes detecting from a use environment a set of encoded signals for use in controlling a shuttering device. The set of encoded signals are analyzed to determine an encoding scheme for the set of encoded signals. The encoded signals are prospectively decoded in the environment according to the determined encoding scheme to control the shuttering device.
This Summary is provided to introduce a selection of concepts in a simplified form that are further described below in the Detailed Description. This Summary is not intended to identify key features or essential features of the claimed subject matter, nor is it intended to be used as an aid in determining the scope of the claimed subject matter.
Additional features and advantages of exemplary implementations of the invention will be set forth in the description which follows, and in part will be obvious from the description, or may be learned by the practice of such exemplary implementations. The features and advantages of such implementations may be realized and obtained by means of the instruments and combinations particularly pointed out in the appended claims. These and other features will become more fully apparent from the following description and appended claims, or may be learned by the practice of such exemplary implementations as set forth hereinafter.
In order to describe the manner in which the above-recited and other advantages and features of the invention can be obtained, a more particular description of the invention briefly described above will be rendered by reference to specific embodiments thereof which are illustrated in the appended drawings. It should be noted that the figures are not drawn to scale, and that elements of similar structure or function are generally represented by like reference numerals for illustrative purposes throughout the figures. Understanding that these drawings depict only typical embodiments of the invention and are not therefore to be considered to be limiting of its scope, the invention will be described and explained with additional specificity and detail through the use of the accompanying drawings in which:
Some implementations described herein are directed to shuttering devices, such as 3D shuttering glasses, that are configured to work with various different manufacturers and protocols, such that a single shuttering device can be adaptively used for different display devices or content devices. Some embodiments allow the shuttering devices to store information about different encoding schemes for 3D shuttering signals. The shuttering devices can detect encoded signals in an environment. The detected encoded signals can be compared to stored encoding schemes to determine the encoding scheme for the detected encoded signals. The encoding scheme can then be used by the shuttering devices to control shuttering of the shuttering devices according to encoded signals in the environment.
Some embodiments may further include functionality for obtaining new encoding schemes at the shuttering devices. For example, the shuttering devices may detect encoded signals in an environment. The shuttering devices may determine that they do not have a stored encoding scheme corresponding to the detected encoded signals. The shuttering devices can then obtain an appropriate encoding scheme corresponding to the detected encoded signals. The obtained encoding scheme can then be used to control shuttering of the shuttering devices according to encoded signals in the environment.
Encoding schemes can be obtained by the shuttering devices in a number of different ways. For example, the shuttering devices may include network connectivity capabilities. For example, the shuttering devices may include a wireless network transceiver, such as an 802.11a, b, n or g transceiver or a Bluetooth transceiver. The shuttering devices can connect to a network using the transceiver. Through the network, the shuttering devices may be able to provide information about the detected signals to a service. The service may be able to determine the encoding scheme for the detected signals. The service can then provide the encoding scheme to the shuttering devices. In an alternative embodiment, the service can periodically provide additional encoding schemes to the shuttering devices as they become available without a request from the shuttering devices. In yet another alternative embodiment, the shuttering devices may be updated over the same channel as encoded signals are received, such as infrared or Bluetooth. These embodiments will be discussed in more detail below.
Similarly, in state 106, the shuttering device 116 can process a shuttering instruction 128 (blank left) received from the video processing device 122. In response, the shuttering device 116 can use a shuttering component 118 to shutter all or part of the user's left eye view of the display device 108. Additionally, the shuttering device 116 can place shuttering component 120 in an un-shuttered or open state, allowing the user's right eye to view the video frame 114 displayed by the display device 108. State 106 may correspond to at least a portion of a period of time in which the display device 108 displays a video frame 114 comprising 3D video content intended for viewing by the user's right eye.
One will appreciate that states 102 and 106 are not limited to shuttering left and right eyes in the manner illustrated. For instance, in state 102, the shuttering device 116 can use the shuttering component 118 to shutter the viewer's left eye during display of right eye content. Furthermore, in state 106, the shuttering device 116 can use the shuttering component 120 to shutter the viewer's right eye during display of left eye content. It will be appreciated, then, that states 102, 104, 106 should not be interpreted as necessarily occurring in the illustrated order.
In state 104, the shuttering device 116 can process and respond to an inter-frame shuttering instruction 126 (blank both) by using both shuttering components 118, 120 to concurrently shutter the user's left and right eye views of the display device 108. State 104 may correspond to at least a portion a frame transition period in which the display device 108 transitions between displaying two video frames. For instance, the display device 108 might be transitioning from display of the “left eye” video frame 110 to the “right eye” video frame 114, or vice versa. Thus, in one or more implementations, the display device 108 displays a frame transition 112, in which the display device 108 concurrently displays at least a portion of two or more different video frames (e.g., video frame 110 and video frame 114). By shuttering both of the user's eyes during display of the frame transition 112, the shuttering device 116 can prevent the user from viewing all or part of the frame transition 112 during all or part of the frame transition period.
One will appreciate that the appropriate shuttering of a single eye, as in states 102 and 106, synchronous with the display of appropriate 3D video content, can provide the illusion that two-dimensional images are actually 3D. Furthermore, inter-frame shuttering, or the synchronous shuttering of both eyes during frame transition periods, can enhance the clarity of the perceived 3D image. For example, inter-frame shuttering can reduce or eliminate undesirable effects such as motion blurring and ghosting. Thus, the disclosed inter-frame shuttering techniques can allow for viewing 3D content on display devices that may have lower frame-rates and/or longer frame overlap or transition periods.
The shuttering device 116 can also receive an inter-frame shuttering instruction 126 (“BB,” or blank both) in connection with a frame transition 112 of the displayed 3D content 202. The frame transition 112 may occur as a result of the display device 108 transitioning between display of the “left eye” video frame 110 and a “right eye” video frame 114. The inter-frame shuttering instruction 126, can instruct the shuttering device 116 to concurrently shutter both of the user's eyes during all or part of the time period 208 in which the display device 108 displays the frame transition 112. In response, the shuttering device 116 can shutter both of shuttering components 118, 120 during all, or part of, the time period 208, preventing the viewer from seeing all or part of the frame transition 112 with either eye.
In addition, the shuttering device can receive a shuttering instruction 128 (“BL,” or blank left) in connection with the “right eye” video frame 114 of the displayed 3D content 202. This instruction 128 can instruct the shuttering device 116 to shutter the user's left eye during all or part of the time period 210 in which the display device 108 displays the “right eye” video frame 114. This may occur, for example, after the display device 108 fully transitions to display of the “right eye” video frame 114. In response, the shuttering device 116 can shutter the shuttering component 118 during all or part of the time period 210. The shuttering device 116 can open, or un-obscure, the other shuttering component 120 during all or part of the time period 210, allowing the viewer to view the right eye video frame 114 with the right eye.
Additionally, in time periods 212 and 214, the display device 108 can subsequently pass through another frame transition 220 to display another left video frame 222. In connection therewith, the shuttering signal 204 can include appropriate shuttering instructions 126, 124. In response to which, the shuttering device can shutter the appropriate shuttering components 118, 120.
Thus, one will appreciate in light of the disclosure herein that the shuttering signal 204 can include any appropriate sequence of shuttering instructions that correspond to the displayed 3D content 202. For instance, if displayed 3D content 202 includes a different sequence of left and right video frames, the shuttering signal 204 can include an appropriate different sequence of shuttering instructions. Furthermore, the shuttering signal 204 can depart from the illustrated shuttering instructions. For example, shuttering signal 204 can refrain from shuttering during one or more frame transition time periods. Furthermore, shuttering signal 204 can include any number of additional shuttering instructions, such a shuttering instruction that does no shuttering (e.g., when the display device 108 displays a video frame intended for viewing with both eyes).
In one or more implementations, the shuttering device 116 receives the shuttering signal 204 prior to the display of the 3D content 202 by the display device 108. Thus, the shuttering device 116 may store (e.g., cache) at least a portion of the shuttering signal 204 prior to performing any instructed shuttering. In additional implementations, the shuttering device 116 receives the shuttering signal 204 concurrently (or substantially concurrently) with the display of the 3D content 202 by the display device 108. Furthermore, in some instances, the shuttering instructions of the shuttering signal 204 can instruct the shuttering device 116 to shutter entire time periods. One will appreciate, however, that the shuttering instructions can also instruct the shuttering device 116 to shutter only a portion of a corresponding time period. Furthermore, the shuttering signal 204 can also instruct the shuttering device 116 to shutter more than a corresponding time period.
In any event, the receiver 302 can communicate the shuttering signal 302 as a whole, or the shuttering instructions individually, to a processing component 308.
Illustratively, the processing component 308 can use the shuttering signal processor 310 to identify shuttering instructions in the shuttering signal 302. Once identified, the shuttering signal processor 310 can instruct one or more of the shuttering components 118, 120 to alter a shuttering state (e.g., shuttered or not shuttered), as appropriate for the shuttering instruction. The shuttering signal processor 310 can use any number of other processing components or modules to perform the processing and to instruct the shuttering components. For instance, when the shuttering signal processor 310 identifies an inter-frame shuttering instruction 126, the shuttering signal processor 310 can instruct both the shuttering components 118, 120 to enter a shuttered state. If one or more of the shuttering components 118, 120 is already in the shuttered state, the shuttering signal processor 310 can instruct the shuttering component to remain in the shuttered state.
The user input component 314 can comprise any means for receiving user input, including any combination of one or more buttons, switches, touch devices, microphones, cameras, light sensing devices, pressure sensing devices, etc. The user input processor 312 can process user input received via the user input component 314.
The user input can comprise any appropriate type of user input, and the shuttering device 116 can use the user input in any appropriate manner. For example, user input may comprise volume control, selection of a shuttering signal type, user feedback to be sent to another device (e.g., the video processing device 122), selection of a mode of the shuttering device (e.g., on, off, standby, configuration, update), etc.
As illustrated by the vertical ellipses between the user input component 314 and an “other” component 322, the shuttering device 116 can include any number of additional components. The additional components can communicate with the processing component 308 or any other component where appropriate. Additional components may include, for example, one or more speakers, a power source, lights, one or more microphones, etc. When additional components are present, components already discussed may be modified accordingly. For instance, when the other components include speakers, the receiver 302 can also receive an analog or digital audio signal, either as part of the shuttering signal 204 or as a separate signal. Then, the processing component 308 may include an audio decoder which sends a decoded audio signal to the speakers.
In one or more implementations, the other component(s) 322 can include a shuttering signal continuity component, which ensures that shuttering occurs, even for intermittent shuttering signals 204 or when the shuttering signal 204 is lost. For instance, the shuttering signal continuity component can implement a phase lock loop (or any similar logic) that analyzes the shuttering signal 204 and generates a replacement signal at the shuttering device 116, when appropriate. The analysis can include determining average sequence and timing information about the shuttering instructions contained in the shuttering signal 204, and developing a model of this information. Then, the shuttering signal continuity component can use phase lock loop (or other logic) to generate a series of substitute shuttering instructions that are estimated to duplicate any shuttering instructions that would have been received if the shuttering signal 204 had not been lost. These generated shuttering instructions can be used to control the shuttering components 118, 120 in the absence of the shuttering signal 204. When the shuttering signal 204 is regained, the shuttering signal continuity component can synchronize the generated substitute shuttering instructions with the shuttering signal 204. In some embodiments, when the shuttering signal 204 is again detected in the environment, re-synching can be done at the shuttering device 116 to correct for any time, frequency, or phase drift that may have occurred when the shuttering signal 204 is not present in the environment.
The shuttering signal 204 may be lost for a variety of reasons, such as physical or electromagnetic interference, loss of signal strength, insufficient transmission bandwidth, etc. For example, if the shuttering signal 204 is transmitted via an infrared transmission, the shuttering signal 204 may be lost if a person or animal walks between the wearer of shuttering glasses and the transmission source, if the wearer turns his or her head, etc. Similarly, if the shuttering signal 204 is transmitted via a BLUETOOTH or Wi-Fi transmission, the shuttering signal 204 may be lost due to interference on the same frequency as the transmission (e.g., from a microwave oven or other wireless devices), if the wearer moves out of range, if data packets in the transmission arrive out of order or delayed, etc.
In some embodiments, the shuttering signal may be intentionally intermittent as a consequence of how the shuttering signal is transmitted. For example, in one embodiment, the shuttering signal 204 may be transmitted using a Bluetooth interface. In one illustrative example, the shuttering signal 204 may be transmitted on an asynchronous Bluetooth audio channel, such as an a2dp channel or other audio channel. The Bluetooth channel is a multiplexed channel, meaning that the channel is shared amongst a number of different entities resulting in periods where the shuttering signal 204 is not transmitted on the channel. Thus, by reconstructing shuttering instructions at the shuttering device 116, based on previously received shuttering instructions 124, 126, and/or 128, control of the shuttering components 118 and 120 can be maintained even when the shuttering signal 204 is not being transmitted on the channel. When the shuttering signal 204 is transmitted on the channel, re-synching can be done at the shuttering device 116 to correct for any time, frequency, or phase drift that may have occurred when the shuttering signal 204 is not present in the environment.
The shuttering components 118, 120 can comprise any component that can selectively obfuscate/shutter all or a portion of a user's view of the display device 108. For example, in one or more implementations the shuttering components 118, 120 can comprise one or more liquid crystal layers that respond to applied voltage. The liquid crystal layers can have the property of becoming opaque (or substantially opaque) when voltage is applied (or, alternatively, when voltage is removed). Otherwise, the liquid crystal layers can have the property of being transparent (or substantially transparent) when voltage is removed (or, alternatively, when voltage is applied). One will recognize in view of the disclosure herein that liquid crystal layers are not the only available shuttering technology. For example, alternate electronic shuttering technologies (e.g., polarized lead lanthanum zirconate titanate (PLZT)) and mechanical shuttering technologies are also available and within the scope of the disclosure herein.
In one or more implementations, the shuttering device 116 can comprise a universal shuttering device. When the shuttering device 116 is a universal device, the processing component 308 can include any number of other processing components for identifying a signal type of the shuttering signal 204. Thus, the receiver 302 can receive any of a number of types of shuttering signals 204, both analog and digital, and the processing component 308 can determine the type of signal and process it accordingly.
Illustratively,
The shuttering device 116 can detect encoded signals, such as the shuttering signal 204, in an environment. The detected encoded signals can be compared using the processing component 308 to encoding schemes stored in the storage device 318 to determine the encoding scheme for the detected encoded signals. The determined encoding scheme can then be used by the shuttering device 116 to control shuttering of the shuttering components 118 and 120 according to encoded signals in the environment.
Some embodiments may further include functionality for obtaining new encoding schemes at the shuttering device 116 for storage on the storage device 318. For example, the shuttering device 116 may detect encoded signals 204 in an environment. The shuttering device 116 may determine that it does not have an encoding scheme stored at the storage device 318 corresponding to the detected encoded signals 204. The shuttering device 116 can then obtain an appropriate encoding scheme corresponding to the detected encoded signals 204. The obtained encoding scheme can then be used to control shuttering of the shuttering components 118 and 120 according to encoded signals 204 in the environment.
Encoding schemes can be obtained by the shuttering devices in a number of different ways. Reference is now made to
In an alternative embodiment, as illustrated in
In yet another alternative embodiment, illustrated in
In one or more implementations, the shuttering device 116 can comprise a configurable shuttering device. When the shuttering device 116 is a configurable shuttering device, the processing component 308 can include any number of other processing components for receiving updates. Thus, the shuttering device 116 can receive any of a number of updates, such as updates to processing components, updates to signal types, new signal types, etc.
The shuttering device 116 may be a battery operated device in that one or more battery units may provide power for operating the various electronic components within the shuttering device 116. The battery units may be single use user replaceable batteries or rechargeable batteries. In some embodiments, the rechargeable batteries may be recharged using mechanical electrical contacts. Alternatively, embodiments may use an inductive charging system. In particular, an inductive charging system may be used to couple inductive coils from a charger to inductive coils in the shuttering device. The inductive coils in the shuttering device 116 may be coupled to charging circuitry, which in turn is coupled to one or more battery units to charge the battery units.
In one or more implementations, the video processing device 122 receives 3D content from a media device. The media device can comprise any number of devices capable of transmitting 3D content to the video processing device 122. For example,
The video processing device 122 can communicate with the display device 108 and the shuttering device(s) 116 in any appropriate manner. For instance, an appropriate wired mechanism, such as HDMI, component, composite, coaxial, network, optical, and the like can couple the video processing device 122 and the display device 108 together. Additionally, or alternatively, an appropriate wireless mechanism, such as BLUETOOTH, Wi-Fi, etc., can couple the video processing device 122 and the display device 108 together. Furthermore, as discussed herein above, any appropriate wired or wireless mechanism (e.g., BLUETOOTH, infrared, etc.) can couple the video processing device 122 and the shuttering device(s) 116 together.
The video processing device 122 can generate an appropriate output signal comprising 3D content received from a media device. For example, when the video processing device 122 and the display device 108 are coupled via a digital mechanism (e.g., HDMI), the video processing device 122 can generate a digital output signal. On the other hand, when the video processing device 122 and the display device 108 are coupled via an analog mechanism (e.g., component, composite or coaxial), the video processing device 122 can generate an analog output signal. The video processing device 122 can process 3D content received from the media device to convert the received 3D content to a format more suited for the particular display device 108. Additionally, the video processing device 122 can generate and send a shuttering signal to the shuttering device(s) 116 that is synchronized to the output signal.
One will appreciate in view of the disclosure herein that the video processing device 122 can take any of a variety of forms. For example, the video processing device 122 may be a set-top box or other customized computing system. The video processing device 122 may also be a general purpose computing system (e.g., a laptop computer, a desktop computer, a tablet computer, etc.). Alternatively, the video processing device 122 may be a special purpose computing system (e.g., a gaming console, a set-top box, etc.) that has been adapted to implement one or more disclosed features.
The display device 108 can be any one of a broad range of display devices that incorporate a variety of display technologies, both current and future (e.g., Cathode Ray, Plasma, LCD, LED, OLED). Furthermore, the display device 108 can take any of a number of forms, such as a television set, a computer display (e.g., desktop computer monitor, laptop computer display, tablet computer display), a handheld display (e.g., cellular telephone, PDA, handheld gaming device, handheld multimedia device), or any other appropriate form. While the display device 108 can be a display device designed specifically to displaying 3D content, the display device 108 can also be a more traditional display device, such as a lower frame-rate device. One will appreciate in light of the disclosure herein that the display device 108 can include both digital and analog display devices.
Accordingly,
Additionally, implementations of the present invention can also be described in terms of flowcharts comprising one or more acts in a method for accomplishing a particular result. Along these lines,
Referring now to
The method 600 further includes analyzing the set of encoded signals to determine an encoding scheme for the set of encoded signals (act 604). For example, the processing component 308 may include functionality for analyzing the shuttering signal 204 to determine an encoding scheme for the shuttering signal. In some embodiments, the determination can be made by comparing the encoded signals 204 to encoding schemes stored in the storage device 318.
The method 600 further includes prospectively decoding encoded signals in the environment according to the determined encoding scheme to control the shuttering device (act 606). For example, shuttering signals 204 can be decoded, such as for example to determine shuttering instructions, to control the shuttering components 118 and 120.
The method 600 may be practiced where analyzing the set of encoded signals to determine an encoding scheme for the set of encoded signals comprises identifying a four state encoding scheme comprising a left eye open, right eye closed state; a left eye closed, right eye open state; a left eye open, right eye open state; and a left eye closed, right eye closed state.
The method 600 may be practiced where detecting a set of encoded signals includes detecting infrared signals. Alternatively, the method 600 may be practiced where detecting a set of encoded signals includes detecting RF signals. The infrared signals may be analog or digital signals. When the signals are digital signals, shuttering instructions may be transmitted in data packets. When the signals are analog signals, shuttering instructions may be transmitted such that detecting signals comprises detecting frequency shift keyed (FSK) signals. Alternatively, when the signals are analog signals, shuttering instructions may be transmitted such that detecting signals comprises detecting phase shift keyed (PSK) signals.
The method 600 may be practiced where detecting a set of encoded signals comprises detecting Bluetooth signals. In these embodiments, analyzing the set of encoded signals to determine an encoding scheme for the set of encoded signals may comprises analyzing one or more packets of data.
As noted above, the method 600 may be practiced where analyzing the set of encoded signals to determine an encoding scheme for the set of encoded signals comprises comparing the detected set of signals to information defining encoding schemes stored in a computer readable storage medium at the shuttering device. For example, comparisons may be made to encoding schemes store on a storage device 318.
As illustrated above, when the computer readable storage medium does not include information defining an encoding scheme that matches the detected set of signals, some embodiments of the method 600 may be practiced to further include obtaining the encoding scheme for the detected set of signals from another entity using a same channel that the detected set of signals were transmitted over. For example,
As illustrated above, when the computer readable storage medium does not include information defining an encoding scheme that matches the detected set of signals, some embodiments of the method 600 may be practiced to further include obtaining the encoding scheme for the detected set of signals from another entity using a different channel than the channel that the detected set of signals were transmitted over. The different channel may include a wired connection such as a USB connection or wired network connection. For example, the shuttering device 116 may be connected to a general purpose computing system capable of receiving encoding schemes 410 from a service and transmitting them to the shuttering device through the wired connection. Alternatively or additionally, the shuttering device may be able to connect to a network 404 through a wired connection to obtain encoding schemes 410 from a service 406.
Alternatively or additionally, the different channel may include a wireless channel, such as a Bluetooth connection or wireless network connection. Examples of such a system are illustrated in
The implementations of the present invention can comprise a special purpose or general-purpose computing systems. Computing systems may, for example, be handheld devices, appliances, laptop computers, desktop computers, mainframes, distributed computing systems, or even devices that have not conventionally considered a computing system, such as DVD players, Blu-Ray Players, gaming systems, and video converters. In this description and in the claims, the term “computing system” is defined broadly as including any device or system (or combination thereof) that includes at least one physical and tangible processor, and a physical and tangible memory capable of having thereon computer-executable instructions that may be executed by the processor.
The memory may take any form and may depend on the nature and form of the computing system. A computing system may be distributed over a network environment and may include multiple constituent computing systems. In its most basic configuration, a computing system typically includes at least one processing unit and memory. The memory may be physical system memory, which may be volatile, non-volatile, or some combination of the two. The term “memory” may also be used herein to refer to non-volatile mass storage such as physical storage media. If the computing system is distributed, the processing, memory and/or storage capability may be distributed as well. As used herein, the term “module” or “component” can refer to software objects or routines that execute on the computing system. The different components, modules, engines, and services described herein may be implemented as objects or processes that execute on the computing system (e.g., as separate threads).
Implementations of the present invention may comprise or utilize a special purpose or general-purpose computer including computer hardware, such as, for example, one or more processors and system memory, as discussed in greater detail below. Embodiments within the scope of the present invention also include physical and other computer-readable media for carrying or storing computer-executable instructions and/or data structures. Such computer-readable media can be any available media that can be accessed by a general purpose or special purpose computer system. Computer-readable media that store computer-executable instructions are physical storage media. Computer-readable media that carry computer-executable instructions are transmission media. Thus, by way of example, and not limitation, embodiments of the invention can comprise at least two distinctly different kinds of computer-readable media: computer storage media and transmission media.
Computer storage media includes RAM, ROM, EEPROM, CD-ROM or other optical disk storage, magnetic disk storage or other magnetic storage devices, or any other medium which can be used to store desired program code means in the form of computer-executable instructions or data structures and which can be accessed by a general purpose or special purpose computer.
A “network” is defined as one or more data links that enable the transport of electronic data between computer systems and/or modules and/or other electronic devices. When information is transferred or provided over a network or another communications connection (either hardwired, wireless, or a combination of hardwired or wireless) to a computer, the computer properly views the connection as a transmission medium. Transmissions media can include a network and/or data links which can be used to carry or desired program code means in the form of computer-executable instructions or data structures and which can be accessed by a general purpose or special purpose computer. Combinations of the above should also be included within the scope of computer-readable media.
Further, upon reaching various computer system components, program code means in the form of computer-executable instructions or data structures can be transferred automatically from transmission media to computer storage media (or vice versa). For example, computer-executable instructions or data structures received over a network or data link can be buffered in RAM within a network interface module (e.g., a “NIC”), and then eventually transferred to computer system RAM and/or to less volatile computer storage media at a computer system. Thus, it should be understood that computer storage media can be included in computer system components that also (or even primarily) utilize transmission media.
Computer-executable instructions comprise, for example, instructions and data which, when executed at a processor, cause a general purpose computer, special purpose computer, or special purpose processing device to perform a certain function or group of functions. The computer executable instructions may be, for example, binaries, intermediate format instructions such as assembly language, or even source code. Although the subject matter has been described in language specific to structural features and/or methodological acts, it is to be understood that the subject matter defined in the appended claims is not necessarily limited to the described features or acts described above. Rather, the described features and acts are disclosed as example forms of implementing the claims.
Those skilled in the art will appreciate that the invention may be practiced in network computing environments with many types of computer system configurations, including, personal computers, desktop computers, laptop computers, message processors, hand-held devices, multi-processor systems, microprocessor-based or programmable consumer electronics, network PCs, minicomputers, mainframe computers, mobile telephones, PDAs, pagers, routers, switches, and the like. The invention may also be practiced in distributed system environments where local and remote computer systems, which are linked (either by hardwired data links, wireless data links, or by a combination of hardwired and wireless data links) through a network, both perform tasks. In a distributed system environment, program modules may be located in both local and remote memory storage devices.
The present invention may be embodied in other specific forms without departing from its spirit or essential characteristics. The described embodiments are to be considered in all respects only as illustrative and not restrictive. The scope of the invention is, therefore, indicated by the appended claims rather than by the foregoing description. All changes which come within the meaning and range of equivalency of the claims are to be embraced within their scope.
The present application is a U.S. National Stage Application corresponding to PCT Patent Application No. PCT/US2011/032549, filed Apr. 14, 2011, which claims priority to U.S. Provisional Application No. 61/416,708, filed Nov. 23, 2010, entitled “3D VIDEO CONVERTER.” The present application is also a continuation-in-part of: PCT Patent Application No. PCT/US2011/025262, filed Feb. 17, 2011, entitled “BLANKING INTER-FRAME TRANSITIONS OF A 3D SIGNAL;” PCT Patent Application No. PCT/US2011/027175, filed Mar. 4, 2011, entitled “FORMATTING 3D CONTENT FOR LOW FRAME-RATE DISPLAYS;” PCT Patent Application No. PCT/US2011/027933, filed Mar. 10, 2011, entitled “DISPLAYING 3D CONTENT ON LOW FRAME-RATE DISPLAYS;” PCT Patent Application No. PCT/US2011/027981, filed Mar. 10, 2011, entitled “SHUTTERING THE DISPLAY OF INTER-FRAME TRANSITIONS;” and PCT Patent Application No. PCT/US2011/031115, filed Apr. 4, 2011, entitled “DEVICE FOR DISPLAYING 3D CONTENT ON LOW FRAME-RATE DISPLAYS.” The entire content of each of the foregoing applications is incorporated by reference herein.
Filing Document | Filing Date | Country | Kind | 371c Date |
---|---|---|---|---|
PCT/US11/32549 | 4/14/2011 | WO | 00 | 12/19/2011 |
Number | Date | Country | |
---|---|---|---|
61416708 | Nov 2010 | US |
Number | Date | Country | |
---|---|---|---|
Parent | PCT/US11/25262 | Feb 2011 | US |
Child | 13379317 | US | |
Parent | PCT/US11/27175 | Mar 2011 | US |
Child | PCT/US11/32549 | US | |
Parent | PCT/US11/27933 | Mar 2011 | US |
Child | PCT/US11/27175 | US | |
Parent | PCT/US11/27981 | Mar 2011 | US |
Child | PCT/US11/27933 | US | |
Parent | PCT/US11/31115 | Apr 2011 | US |
Child | PCT/US11/27981 | US |