This application claims priority of Chinese Application No. CN201510278144.8, filed on May 27, 2015, the entire contents of which are hereby incorporated by reference.
The present disclosure generally relates to the field of three-dimensional (3D) display technologies and, more particularly, relates to a video processing method and a video processing device.
In three-dimensional (3D) display technologies, human-computer interaction is not limited to two-dimensional (2D) plane and has been extended to three-dimensional (3D) space. In pursuit of realism, interaction in three-dimensional space must be closely integrated with visual effects. As smart phones, tablet computers and other portable electronic devices are getting widely used, users may capture photos and videos using the built-in cameras and then may use application software to process the captured photos and videos, for example, for image enhancement, image modification, etc. However, such processing of the photos and videos may be still limited to 2D plane effects rather than dynamic 3D spatial effects. For more realistic 3D viewing experience, users are demanding dynamic 3D alteration of video sequence data captured by the portable electronic devices or stored in the portable electronic devices.
The disclosed video processing method and image processing device are directed to solve one or more problems set forth above and other problems in the art.
One aspect of the present disclosure provides a video processing method. The video processing method includes acquiring a to-be-processed 3D video sequence consisting of a plurality of to-be-processed images, wherein each to-be-processed image is a 3D image including a first view image and a second view image, and the first view image and the second view image have a horizontal parallax between them. The video processing method also includes receiving a user instruction and determining special-effect data to be inserted into each of the plurality of to-be-processed images, wherein the special-effect data corresponds to a special effect to be inserted into each of the plurality of to-be-processed images. The video processing method also includes, based on special-effect attribute information of the special-effect data, respectively combining the special-effect data with the first view image and the second view image of each of the plurality of to-be-processed images to obtain a plurality of 3D special-effect images one-to-one corresponding to the plurality of to-be-processed images included in the to-be-processed 3D video sequence. The video processing method also includes storing the plurality of 3D special-effect images one-to-one corresponding to the plurality of to-be-processed images included in the to-be-processed 3D video sequence, and displaying the plurality of 3D special-effect images.
Another aspect of the present disclosure provides a video processing device. The image processing device includes a video acquisition unit configured to acquire a to-be-processed 3D video sequence, wherein the to-be-processed 3D video sequence includes a plurality of to-be-processed images, and each to-be-processed image is a 3D image including a first view image and a second view image having a parallax between them. The image processing device also includes an image processing unit configured to sequentially capture each of the plurality of to-be-processed images from the to-be-processed 3D video sequence; a special effect selection unit configured to receive a user instruction and determine special-effect data to be inserted into each to-be-processed image, wherein the special-effect data corresponds to a special effect to be inserted into each to-be-processed image; a combining unit, based on special-effect attribute information of the special-effect data, configured to respectively combine the special-effect data with the first view image and the second view image of each of the plurality of to-be-processed images to obtain a plurality of 3D special-effect images one-to-one corresponding to the plurality of to-be-processed images included in the to-be-processed 3D video sequence; and a storage unit configure to store the plurality of 3D special-effect images corresponding to the plurality of to-be-processed images included in the to-be-processed 3D video sequence.
Other aspects of the present disclosure can be understood by those skilled in the art in light of the description, the claims, and the drawings of the present disclosure.
The following drawings are merely examples for illustrative purposes according to various disclosed embodiments and are not intended to limit the scope of the present disclosure.
Reference will now be made in detail to exemplary embodiments of the invention, which are illustrated in the accompanying drawings. Hereinafter, embodiments consistent with the disclosure will be described with reference to drawings. It is apparent that the described embodiments are some but not all of the embodiments of the present invention. Based on the disclosed embodiments, persons of ordinary skill in the art may derive other embodiments consistent with the present disclosure, all of which are within the scope of the present invention.
When no conflict exists, the exemplary features illustrated in various embodiments may be combined and/or rearranged. The specific details provided in the descriptions of various embodiments are intended to help understanding the present disclosure. However, the present disclosure may be implemented in other manners that are not described herein. Thus, the scope of the present disclosure is not limited to the disclosed embodiments. In various embodiments, the terms “first” and “second”, etc., are used to describe technical differentiations, and such terms may be replaced without departing from the scope of the present disclosure.
The present disclosure provides a video processing method for 3D image capturing, 3D video capturing, or other image and video processing applications by, for example, portable devices equipped with built-in camera, such as mobile terminals, tablet computers, etc. The video processing method according to the present disclosure may also apply to, but is not limited to, 3D mobile phones, tablet computers, and other portable electronic devices that are integrated with light splitting devices such as gratings. The video processing method according to the present disclosure may also apply to 2D mobile phones. The video processing method according to the present disclosure may be implemented into a device.
The processor 802 may include any appropriate type of central processing unit (CPU), graphic processing unit (GPU), general purpose microprocessor, digital signal processor (DSP) or microcontroller, and application specific integrated circuit (ASIC). The processor 802 may execute sequences of computer program instructions to perform various processes associated with the device 800.
The display 804 may be any appropriate type of display, such as plasma display panel (PDP) display, field emission display (FED), cathode ray tube (CRT) display, liquid crystal display (LCD), organic light emitting diode (OLED) display, light emitting diode (LED) display, or other types of displays.
The camera 806 may be an internal camera in the device 800 or may be an external camera connected to the device 800 over a network. The camera 806 may take images and/or videos to be processed by the processor 802.
The system memory 808 is just a general term that may include read-only memory (ROM), random access memory (RAM) and etc. The ROM may store necessary software for a system, such as system software. The RAM may store real-time data, such as images for displaying.
The system bus 810 may provide communication connections, such that the display device may be accessed remotely and/or communicate with other systems via various communication protocols, such as transmission control protocol/internet protocol (TCP/IP), hypertext transfer protocol (HTTP), etc.
The input/output unit 812 may be provided for users to input information into the display device or for the users to receive information from the display device. For example, the input/output unit 812 may include any appropriate input device, such as a remote control, a keyboard, a mouse, an electronic tablet, voice communication devices, or any other optical or wireless input devices.
Further, the mass storage device 814 may include any appropriate type of mass storage medium, such as a CD-ROM, a hard disk, an optical storage, a DVD drive, or other type of storage devices.
During operation, the processor 802 may implement various software units, which may perform certain functions/steps to process and/or display images or videos to one or more users.
A user may take a 3D video by his/her smartphone and want to beautify the 3D video through adding 3D dynamic effects, for example, floating feathers, snowflakes, bubbles or raindrops. Such a desire may be realized through a video processing method consistent with disclosed embodiments.
Step S201: acquiring a to-be-processed 3D video sequence. The to-be-processed 3D video sequence may include a plurality of to-be-processed images. Each to-be-processed image may be a 3D image.
The to-be-processed 3D video sequence may be acquired by the user using the camera, retrieved from system memory of an electronic device, downloaded or transmitted through a network, etc., which are only for illustrative purposes and are not intended to limit the scope of the present invention. The to-be-processed 3D video sequence may include a plurality of 3D images (i.e., to-be-processed images) sequentially arranged. When the plurality of 3D images are sequentially played, a dynamic video may be observed. The dynamic video may correspond to the to-be-processed 3D video sequence.
Step S202: sequentially capturing each of the plurality of to-be-processed images included in the to-be-processed 3D video sequence. Each to-be-processed image may include a first view image and a second view image, and the first view image and the second view image may have a parallax between them.
In the disclosed embodiments, to realize a 3D effect, each to-be-processed image may include the first view image and the second view image. The first view image may be a left view image of the 3D image (i.e., the to-be-processed image) and the second view image may be a right view image of the 3D image. Meanwhile, the first view image and the second view image of the 3D image may have a horizontal parallax between them. The to-be-processed image is also called as a 3D background image in the following descriptions, and the to-be-processed 3D video sequence is also called as a 3D background video in the following descriptions.
In another embodiment, the first view image may be a right view image of the 3D image (i.e., the to-be-processed image) and the second view image may be a left view image of the 3D image. The first view image and the second view image in the 3D image may have a horizontal parallax in the horizontal direction. In another embodiment, the first view image may be an upper view image of the 3D image and the second view image may be a lower view image of the 3D image. In another embodiment, the first view image may be a lower view image of the 3D image and the second view image may be an upper view image of the 3D image.
Step S203: receiving a user instruction and determining special-effect data to be inserted to each of the plurality of to-be-processed images included in the to-be-processed 3D video sequence.
In this step, the user may select the desired special-effect data and may instruct certain image processing operations. Special-effect data, as used herein, may refer to data used for creating certain display or viewing effect in one or more images. An electronic device (for example, a smartphone) may obtain the special-effect data required for processing the to-be-processed images based on the received user instruction.
In one embodiment, the disclosed video processing method may be implemented into an application program. The user may select the requested special-effect data from a dropdown list provided by the application program, from a search engine interface provided by the application program, or from a preconfigured template or model. For example, the user may enter text command in the search engine interface. According to the text command, the search engine may search the special-effect database to obtain the special-effect data searched by the user, which is then presented to the user to be selected by the user.
Step S204: based on special-effect attribute information, respectively combining the special-effect data with the first view image and the second view image of each of the plurality of to-be-processed images to obtain a plurality of 3D special-effect images or a plurality of images with the 3D special effect, which may be one-to-one corresponding to the plurality of to-be-processed images included in the to-be-processed 3D video sequence.
In particular, the electronic device may combine the special-effect data and the to-be-processed image based on the special-effect attribute information to obtain the 3D special-effect image or the image with the 3D special effect. After the special-effect data is respectively combined with the first view image and the second view image of the to-be-processed image, the 3D special-effect image may be a single view image including the special-effect data, content of the first view image and the second view image of the to-be-processed image.
After combining the special-effect data with the first view image and the second view image of the to-be-processed image, the special effect corresponding to the special-effect data may have a horizontal parallax. The special-effect attribute information may include position information of the special effect located in the to-be-processed image, variation information of the special effect, and a number of frames of the 3D special-effect images required to be generated.
When the number of frames of the 3D special-effect images reaches a certain value, the desired dynamic effect may be observed when the plurality of 3D special-effect images are continuously displayed on a device capable of playing 3D images and/or videos, such as a 3D display device, etc. The value may be determined by a refresh rate on the 3D display device, i.e., a frame rate of the 3D display device.
The position information of the special effect in the to-be-processed image may include positions of the special effect in the first view image and the second view image of the to-be-processed image, such as insertion positions (or initial positions) of the special effect. The variation information of the special effect may include size variation information, trajectory variation information, color variation information, etc.
Further, before combining the special-effect data with the first view image and the second view image of the to-be-processed image, parallax information between the first view image and the second view image of the to-be-processed image may be calculated. The parallax information between the first view image and the second view image of the to-be-processed image may be retrieved from the smartphone. For example, the smartphone may capture the 3D image (i.e., the to-be-processed image), automatically compare the left view image (i.e., the first view image) and right view image (i.e., the second view image) of the 3D image to calculate the parallax information between the left view image and the right view image, and then store the parallax information in system memory.
In another embodiment, the smartphone may capture the 3D image and output the 3D image to a third party. The third party may compare the left view image (i.e., the first view image) and right view image (i.e., the second view image) of the 3D image to calculate the parallax information between the left view image and the right view image, and then return the parallax information to the smartphone.
The position information of the special effect in the 3D image (i.e., the to-be-processed image) may correspond to the parallax information between the first view image and the second view image of the 3D image. Based on the position information of the special effect in the 3D image (i.e., the positions of the special effect in the first view image and the second view image) and the parallax information between the first view image and the second view image of the 3D image, the special effect may be combined with the first view image and the second view image of the 3D image at corresponding positions, respectively. Combining the special effect with the first view image and the second view image of the 3D image may also change pixel data in the first view image and the second view image of the 3D image, respectively.
In certain other embodiments, the parallax information between the first view image and the second view image might not be calculated, the special effector the special-effect data corresponding to the special effect may be directly combined with the 3D background image.
The special effect may be combined with the to-be-processed image through various approaches. For example, in one embodiment, when butterflies are used as the special effect, pixels in a certain region of the first view image and the second view image of the 3D image may be replaced by pixel data corresponding to the butterfly, respectively. Thus, an image with the butterfly may be obtained. When being displayed on the 3D display device, the 3D special-effect of the butterfly may be observed.
It should be noted that, each 3D special-effect image may also include two view images: the first view image after being inserted with the special-effect data and the second view image after being inserted with the special-effect data. After the special-effect data is respectively combined with the first view image and the second view image of the to-be-processed image, a same special effect in the first view image and the second view image of the 3D special-effect image may have a parallax larger than the parallax between the first view image and the second view image of the to-be-processed image at corresponding positions.
The parallax of the same special effect between the first view image and the second view image of the 3D special-effect image may be determined based on the position information of the same special effect in the 3D special-effect image, i.e., the positions of the same special effect in the first view image and the second view image of the 3D special-effect image. For example, the position coordinate of the special-effect data in the first view image may be (100, 100). Both the row and column coordinates of the first view image are 100. The corresponding position coordinate in the second view image may be (102, 100).
The insertion position of the special effect in each view image of different 3D special-effect images may be selected randomly, may be determined based on the user input, or may be determined by certain rules, such as a trajectory algorithm.
The special-effect attribute information may also include the variation information of the special effect. The special effect may change from various aspects, for example, size, shape, etc.
In the disclosed embodiments, the special effect may be equivalent to the special-effect data. In certain other embodiment, the special effect be different from the special-effect data.
Step S205: storing the plurality of 3D special-effect images corresponding to the plurality of to-be-processed images included in the to-be-processed 3D video sequence.
The plurality of 3D special-effect images may be one-to-one corresponding to the plurality of to-be-processed images included in the to-be-processed 3D video sequence. After the 3D special-effect data is inserted into the first view and the second view of each of the plurality of the to-be-processed image included in the to-be-processed 3D video sequence in steps S201-S204, the 3D special-effect images may be stored in, for example, an MP4 format or other displayable format. Then, the stored 3D special-effect images may be displayed or played back by the 3D display device, and the 3D dynamic effect may be observed on the 3D background video. For example, referring to
Based on the user instruction, the electronic device (e.g., a smartphone) may determine the to-be-processed image and desired dynamic special effect to be inserted into the to-be-processed image. For example, the user may select snowflakes as the special effect through a dropdown list.
In one embodiment, the position information included in the special effect attribute information may be generated randomly or selected by the user through clicking a display screen of the electronic device. The number of frames of the 3D special-effect images may be determined by a model algorithm or may be input by the user. The special effect attribute information may also include variation information to control size variation, trajectory variation, and/or color variation.
It should be noted that, the first view image and the second view image in the to-be-processed image may have the horizontal parallax in the horizontal direction, and parallax information between the first view image and the second view image may be acquired before combining the special-effect data with the to-be-processed image. In particular, the position information of the special effecting the first view image and the second view image of the to-be-processed image may correspond to the parallax information between the first view image and the second view image of the to-be-processed image, which may enable a fast and accurate combination of the special effect and the to-be-processed image.
Based on the position information of the special effecting the first view image and the second view image of the to-be-processed image and the parallax information between the first view image and the second view image of the to-be-processed image, the electronic device may combine the special effector the special-effect data with the first view image and the second view image at the corresponding positions, respectively, such that a 3D special-effect image may be obtained.
Combining the special effect or the special-effect data with the to-be-processed image may also change pixel data in the first view image and the second view image of the to-be-processed image, respectively. For example, the pixel data at the corresponding position of each view image may be replaced by the special effect pixel data.
In particular, the 3D special-effect image may also include two view images: the first view image after being inserted with the special-effect data and the second view image after being inserted with the special-effect data. A same special effect in the first view image and the second view image of the 3D special-effect image may have a parallax larger than the parallax between the first view image and the second view image of the to-be-processed image at corresponding positions.
In certain embodiments, the parallax information between the first view image and the second view image of the to-be-processed image may not be calculated, and the special effector the special-effect data may be directly combined with the 3D background image.
After inserting the 3D special effect into the first view and the second view of the to-be-processed image, a 3D special-effect image may be obtained. The 3D special-effect image may also include two view images: the first view image after being inserted with the special-effect data and the second view image after being inserted with the special-effect data.
In particular, a same special effect in the first view image and the second view image of the 3D special-effect image may have a parallax larger than the parallax between the first view image and the second view image of the to-be-processed image at corresponding positions. For example, a snowflake 401 may be combined with the first view image (the left view image) shown in
The parallax between the snowflake 401 in the first view image and the second view image of the 3D special-effect image may be larger than the parallax between the first view image and the second view image of the to-be-processed image at the corresponding positions. That is, the parallax between the position Bin
Returning to
Further, the position of the special effect snowflake in one 3D special-effect image may vary from the position of the corresponding snowflake in another 3D special-effect image. That is, the positions of the special effect snowflake in a plurality of 3D special-effect images may vary.
Thus, when the plurality of 3D special-effect images are continuously displayed, a dynamic effect may be observed. Especially, when the plurality of 3D special-effect images, each containing two view images inserted with the special effect, are continuously displayed on the 3D display device (e.g., a smartphone), a 3D dynamic effect may be observed due to the horizontal parallax of the special effect, as
For example, referring to
To interactively generate snowflakes, the user may first click an insertion position of the special effect, and the electronic device may insert snowflakes at the position clicked by the user and then spread out circularly or radially.
To randomly generate snowflakes, the user may not need to indicate any insertion position of the special effect, and the electronic device may randomly determine an initial insertion position and may insert snowflakes at the determined position and then spread out circularly or radially.
The interactive generation method and the random generation method may be similar except that the interactive method may generate the first special effect at the position clicked by the user. The special effect motion model may be a circular motion model. The position clicked by the user may be the center of a circle. The special effect at the radius r may be moved to the position at the radius r′. r′=kr. k is a scalar value and may not alter the direction of r.
The interactive generation method and the random generation method to combine the special effector the special-effect data with the to-be-processed image are for illustrative purpose, which are not intended to limit the scope of the present inventions. The details of the interactive generation method and the random generation method are not going to be explained here.
Through the disclosed video processing method, the 3D effect as well as the dynamic 3D effect may be observed on the 3D display device. However, the dynamic 3D effect may not be observed on a device which is only capable of displaying 2D images and/or videos, such as a 2D display device. On the 2D display device, only a plurality of continuously animated snowflake view images may be observed, like animated GIF images.
To observe 3D like dynamic effect of the 3D video sequence on the 2D display device, before the step S204, i.e., before respectively combining the special-effect data with the first view image and the second view image of each of the plurality of to-be-processed images to obtain the plurality of 3D special-effect images corresponding to the plurality of to-be-processed images included in the to-be-processed 3D video sequence, the electronic device may insert a border to the first view image or the second view image of the to-be-processed image.
After the 3D dynamic special effect has been inserted to the to-be-processed view image, the electronic device may store the 3D special-effect images. In particular, storing the 3D special-effect images may include the following approaches.
Table 1 illustrates an exemplary method for storing 3D special-effect images consistent with disclosed embodiments.
As shown in Table 1, a plurality of 3D special-effect images are going to be stored, and each 3D special-effect image may include a first view image and a second view image. The plurality of 3D special-effect images may be arranged in a sequence corresponding to the to-be-processed 3D video sequence, i.e., forming a sequence of the 3D special-effect images.
The first view image of each 3D special-effect image may be sequentially stored according to the sequence of the 3D special-effect images, forming a first view image sequence. The second view image of each 3D special-effect image may also be sequentially stored according to the sequence of the 3D special-effect images, forming a second view image sequence.
The first view image sequence and the second view image sequence may be stored in separated files and then packed. When the stored files (i.e., the first view image sequence and the second view image sequence) are played back, the first view images and the second view images may be sequentially arranged or combined and then played.
Table 2 illustrates another exemplary method for storing 3D special-effect images consistent with disclosed embodiments.
As shown in Table 2, a plurality of 3D special-effect images may be stored, and each 3D special-effect image may include a first view image and a second view image. The first view image and the second view image in each 3D special-effect image may be combined or paired together to form a 3D special-effect image sequence corresponding to the frame number of the to-be-processed 3D video sequence. That is, the to-be-processed 3D video sequence may include a plurality of to-be-processed images forming a to-be-processed image sequence, and the 3D special-effect image sequence may be corresponding to the to-be-processed image sequence.
The 3D special-effect images included in the 3D special-effect image sequence may be sequentially stored in separated files and then packed. When the stored files (i.e., the 3D special-effect image sequence or the combine first view image and the second view image sequence) are played back, the 3D special-effect images may be played sequentially.
In certain embodiments, the border may be inserted. Storing the 3D special-effect images with the border may include the following methods.
Table 3 illustrates another exemplary method for storing 3D special-effect images consistent with disclosed embodiments.
As shown in Table 3, a plurality of 3D special-effect images may be stored, and each 3D special-effect image may include a first view image and a second view image. The first view image of each 3D special-effect image may be sequentially stored according to a sequence of the 3D special-effect images, forming a first view image sequence. The second view image of each 3D special-effect image may be sequentially stored according to the sequence of the 3D special-effect images, forming a second view image sequence. The border corresponding to each 3D special-effect image may be stored sequentially, forming a border sequence.
The first view image sequence, the second view image sequence and the border sequence may be stored in separated files. When the user plays the stored files on the 3D display device, the user may simply open the stored files and play sequentially.
When the user wants to transmit the stored files to the 2D display device that does not support 3D display, the user may only transmit the border sequence and one of the first view image sequence and the second view image sequence. Thus, not all the three sequences may be transmitted to the 2D display device, and data (i.e., files) required to be transmitted may be reduced.
Table 4 illustrates another exemplary method for storing 3D special-effect images consistent with disclosed embodiments.
As shown in Table 4, each 3D special-effect image may include a first view image and a second view image. The first view image and the second view image in each 3D special-effect image may be combined or paired together to form a 3D special-effect image sequence corresponding to the frame number of the to-be-processed 3D video sequence. The border corresponding to each 3D special-effect image may be stored sequentially, forming a border sequence.
When the user wants to play the stored data on the 3D display device or to transmit the stored data to the 3D display device, the user may only need to play or transmit the 3D special-effect image sequence and may not transmit the border sequence. On the other hand, when the user wants to transmit the stored data to the 2D display device, the user may need to transmit the 3D special-effect image sequence and the border sequence, which may be sequentially combined before being played.
Further, the border may be combined with the first view image and the second view image of each 3D special-effect image to form a 3D special-effect image sequence with the border corresponding to the frame number of the special effect attribute information.
Table 5 illustrates another exemplary method for storing 3D special-effect images consistent with disclosed embodiments.
As shown in Table 5, each 3D special-effect image may include a first view image and a second view image. The first view image and the second view image in each 3D special-effect image as well as the border corresponding to each 3D special-effect image may be combined or paired sequentially, forming a 3D special-effect image plus border sequence corresponding to the frame number of the to-be-processed 3D video sequence. Then the 3D special-effect image sequence with the border may be stored.
When transmitting the stored data, the user may transmit the entire 3D special-effect image sequence with the border. When received by the 2D display device, the 3D special-effect image sequence with the border may be played directly and the 3D like dynamic effect may be observed.
Further, to meet requirements of data transmission, the to-be-processed image, dynamic special-effect attribute information, and the special effect may be stored in separate files. A sending electronic device may not process the images, but transmit the to-be-processed image, the special-effect attribute information, and the special-effect to a receiving electronic device.
Based on the dynamic special-effect attribute information, the receiving electronic device may process the to-be-processed image to obtain the images with the 3D dynamic effect, which may be identical as if the images with the 3D dynamic effect are generated by the sending electronic device. Because only the to-be-processed image, the dynamic special-effect attribute information, and the special-effect data are transmitted, the data required for transmission may be significantly reduced.
Table 6 illustrates an exemplary method for storing 3D special-effect images consistent with disclosed embodiments.
As shown in Table 6, the to-be-processed 3D video sequence and the dynamic special-effect attribute information may be stored in separated files. The dynamic special-effect attribute information may include special-effect attribute information and special-effect data.
Table 7 illustrates another exemplary method for storing 3D special-effect images consistent with disclosed embodiments. The to-be-processed 3D video sequence may include a plurality of to-be-processed images. A border may be inserted into a first view image or a second view image of each to-be-processed image.
As shown in Table 7, the to-be-processed 3D video sequence, dynamic special-effect attribute information and border information may be stored in separated files. The dynamic special-effect attribute information may include special-effect attribute information and special-effect data.
Table 8 illustrates another exemplary method for storing 3D special-effect images consistent with disclosed embodiments.
As shown in Table 8, the to-be-processed 3D video sequence may include a plurality of to-be-processed images, and each to-be-processed image may include a first view image and a second view image. The first view image of each to-be-processed image may be sequentially stored, forming a first view image sequence. The second view image of each to-be-processed image may also be sequentially stored, forming a second view image sequence. The first view image sequence, the second view image sequence and the dynamic special-effect attribute information may be stored in separated files. The dynamic special-effect attribute information may include special-effect attribute information and special-effect data.
Table 9 illustrates another exemplary method for storing 3D special-effect images consistent with disclosed embodiments.
As shown in Table 9, the to-be-processed 3D video sequence may include a plurality of to-be-processed images, and each to-be-processed image may include a first view image and a second view image. The first view images of each to-be-processed image may be sequentially stored, forming a first view image sequence. The second view images of each to-be-processed image may also be sequentially stored, forming a second view image sequence. A border may be inserted into the first view image or the second view image.
The first view image sequence, the second view image sequence, the dynamic special-effect attribute information, and the border information may be stored in separated files may be stored in separated files. The dynamic special-effect attribute information may include special-effect attribute information and special-effect data.
The present disclosure provides the video processing method in various embodiments. Each to-be-processed image included in the 3D video sequence may be acquired. Based on the user instructions, the special-effect data to be inserted into the to-be-processed image and the special-effect attribute information may be determined. Then the special-effect data may be combined with each to-be-processed image respectively, such that a plurality of 3D special-effect images may be obtained. Accordingly, a 3D special-effect video corresponding to the plurality of 3D special-effect images may be obtained. When the 3D special-effect video is played, desired 3D viewing experience may be observed to meet the user's requirements.
The present disclosure further provides a video processing device.
The video acquisition unit 601 may be configured to acquire a to-be-processed 3D video sequence. The to-be-processed 3D video sequence may include a plurality of to-be-processed images, and each to-be-processed image may be a 3D image.
The image processing unit 602 may be configured to capture each of the plurality of to-be-processed images from the to-be-processed 3D video sequence. In particular, each to-be-processed image may include a first view image and a second view image. The first view image and the second view image of the to-be-processed image may have a parallax between them.
The special effect selection unit 603 may be configured to receive user instructions and determine special-effect data for a 3D special effect to be inserted into the to-be-processed images.
Based on special-effect attribute information, the combining unit 604 may be configured to respectively combine the special-effect data with the first view image and the second view image of each of the plurality of to-be-processed images, such that a plurality of 3D special-effect images corresponding to the plurality of the to-be-processed images included in the to-be-processed 3D video sequence may be obtained. Accordingly, a corresponding 3D special-effect video may be obtained. The 3D special-effect image may also include a first view image and a second view image.
In particular, after a same special effect corresponding to the special-effect data is respectively combined with the first view image and the second view image of the to-be-processed image, the same special effecting the first view image and the second view image of the 3D special-effect image may have a horizontal parallax between them. The special-effect attribute information may include position information of the special effect in the to-be-processed image.
The storage unit 605 may be configured to store the plurality of 3D special-effect images corresponding to the plurality of the to-be-processed images included in the to-be-processed 3D video sequence.
The device 600 may be an electronic device implemented with the disclosed video processing methods. The details of the disclosed video processing device are not repeated here. It should be noted that, names of the software units are only for illustrative purposes, which are not intended to limit the scope of the present invention.
The video processing device consistent with disclosed embodiments may be able to acquire the to-be-processed 3D video sequence, and process each of the plurality of to-be-processed images included in the to-be-processed 3D video sequence. That is, based on the user instructions, the disclosed video processing device may be able to determine the special-effect data to be inserted into the to-be-processed image and the special-effect attribute information, combine the special-effect data with each to-be-processed image, and generate the 3D special-effect video. When the 3D special-effect video is played, desired 3D viewing experience may be observed to meet the user's requirements.
Those of skill would further appreciate that the various illustrative units and algorithm steps disclosed in the embodiments may be implemented as electronic hardware, computer software, or combinations of both. To clearly illustrate this interchangeability of hardware and software, various illustrative units and steps have been described above generally in terms of their functionality. Whether such functionality is implemented as hardware or software depends upon the particular application and design constraints imposed on the overall system. Skilled artisans may implement the described functionality in varying ways for each particular application, but such implementation decisions should not be interpreted as causing a departure from the scope of the present invention.
The steps of a method or algorithm disclosed in the embodiments may be embodied directly in hardware, in a software unit executed by a processor, or in a combination of the two. A software unit may reside in RAM, flash memory, ROM, EPROM (erasable programmable read-only memory), EEPROM (electrically erasable programmable read-only memory), registers, hard disk, a removable disk, a CD-ROM, or any other form of storage medium known in the art.
The description of the disclosed embodiments is provided to illustrate the present invention to those skilled in the art. Various modifications to these embodiments will be readily apparent to those skilled in the art, and the generic principles defined herein may be applied to other embodiments without departing from the spirit or scope of the invention. Thus, the present invention is not intended to be limited to the embodiments shown herein but is to be accorded the widest scope consistent with the principles and novel features disclosed herein.
Number | Date | Country | Kind |
---|---|---|---|
2015-10278144.8 | May 2015 | CN | national |