1. Field of the Invention
The invention is in the field of processing image data, for example with the aim of generating visual effects, and/or with the aim of treating (temporarily) relevant data different from (temporarily) irrelevant data.
2. Description of Related Art
Several approaches exist to change perspectives and generate virtual viewpoints based on imagery recorded by TV-cameras. To minimize visual artefacts and create the best rendering result, various geometric representations and rendering algorithms are known.
For example, virtual viewpoints for sport scenes for example of television have been done according to a first option by using deformed 3D template meshes with pre-stored textures. The final rendering, in accordance with this option, is a completely virtual scene, where actors/players as well as the surroundings (e.g. stadium) are rendered this way.
According to a second option, this is done by using deformed 3D template meshes with pre-stored textures on a background, which is based on an approximate geometry of the surroundings using textures from the real cameras.
In accordance with a third option, the virtual viewpoints have been implemented by using approximate geometry and textures taken from the real camera images. This representation is then used for all the objects of the scene.
A further issue in image processing is generating visual effects, for example for camera transitions. A camera transition is a transition between camera images from two cameras, wherein the camera images can be the same during the entire transition or can change during the transition. In this, the cameras may be real or virtual cameras. Camera transitions with synthetic images and video sequences, generated for virtual cameras, are described, for example in US 2009/0315978.
Image effects are also used to convey motion or alter the perception of an image by transforming a source image based on this effect. Possible effects are motion blur, depth of field, color filters and more. These kinds of image effects are used in the movie industry for computer generated movies, or post processing effects.
The image effects for camera transitions can be divided in different categories:
Applying a spatially independent image effect to a camera transition does not allow controlling the effect spatially. Therefore, regions or objects, which should remain in the viewer's focus or should be handled differently, cannot be handled in a particular way. E.g. applying a motion blur on a camera transition might also blur the object, which should always remain recognizable for the viewer.
A known method applies the image effect to the background object but not to the foreground object (e.g. defined by a color based cutout). This is comparable to combining two separate images from the camera transitions, whereas one will not be modified by the image effect and the other one will be modified by the spatially independent image effect. In both cases the image effect is not spatially dependent, which results in a camera transition where the foreground object seems to be stitched on top of the background object.
Accordingly, it is an object of the present invention to provide methods overcoming drawbacks of prior art methods and especially of allowing seamless integration of image or scene parts to be handled specially into the entire image or scene.
In accordance with a first aspect of the invention, a method comprises the steps of
In the step of applying at least one image effect to the transition region, the image effect may be linearly or non-linearly blended between ‘close to the ROI’ and ‘close to the default region’. For example the parameter or at least one of the parameter may be a continuous, monotonous function of a position along a trajectory from the ROI to the default region through the transition region.
In this, it is understood that if an image effect is not applied to a certain region, this can be described by a particular parameter value (or particular parameter values) of this particular image effect, for example a parameter value of 0—depending on the representation.
In accordance with a second aspect of the invention, a method comprises the steps of
In this, a transition region according to the first aspect may be present between the region-of-interest and the default region.
For example, the 3D data may comprise stereo disparity data (i.e. a relative displacement of a certain object in the image in images taken from different cameras, this depending on the distance of the object from the cameras), or may comprise a 3D position, a depth value, etc.; any data representing information on the 3D properties of the scene and being obtainable from camera images and/or from other data may be used.
The region-of-interest in the image data may be a 2D region-of-interest or a 3D region-of-interest.
A 2D region-of-interest is a region defined in the 2D image sequence, which region, however, in accordance with the second aspect is identified based on 3D data. For example, in a sports event, in the step of identifying a region-of-interest in at least one image of the image sequence, a certain player is chosen to be central to the region-of-interest. Then, in the image sequence, 3D data is used to identify objects (other players, a ball, etc.) that are close to the chosen player. Everything that is, in the 2D images of the image sequence, close to the object is then defined to belong to the region-of-interest.
Alternatively, the region-of-interest in the image data may be a 3D region-of-interest.
An example of a 3D ROI is as follows. To determine a 3D ROI, after choosing an object (for example a player) to belong to the ROI, a projection on a plane different from the normal to the line that connects this object with the (real or virtual) camera is made. An environment of the projection in this plane (for example a 5 m radius circle around the player's location) defines the ROI, which by definition comprises every object that is projected onto a region (the said environment) on said plane. A specific example is as follows. In an image sequence of a sports event, the location of a particular player is projected to the ground plane. A region on the ground plane around the player then defines the ROI, and every object that is in this region on the ground plane belongs to the ROI—independent on how far away this object is from the player in the actual camera image.
In accordance with the second aspect of the invention, it is proposed to use depth information of the scene for the image transition. This is in contrast with the prior art that does not consider any depth information of the scene. Nevertheless, merely an image effect is applied to the ROI and/or the default region (and if applicable to the transition region). An image effect is generally an effect applied to the 2D image(s). Thus it is not necessary to do a computationally extremely complex computation of a 3D model of the object to which the effect is applied (which laborious 3D model computation is sometimes applied to moving objects in animated pictures).
In accordance with a third aspect, a method for generating camera transitions or other image sequences based on a spatially adaptive image effect comprises of the following steps.
The sequence of the steps may be from A. through F., or it may optionally be interchanged, however with step F. generally being the last of the mentioned steps.
In step F., the parameter(s) may be chosen to gradually vary through the transition region (as a function of position or speed, for example along a trajectory) from a value corresponding to the value in the ROI at the interface to the ROI to a value corresponding to the default region value at the interface to the default region. The gradual variation may for example be in function of a 3D distance from the ROI, of a distance projected onto a plane (such as the ground) or in function of a 2D distance in the image.
Alternatively, the parameter(s) may vary discontinuously or be constant in the transition region, for example somewhere between the ROI and default region values.
The third aspect of the invention may be implemented together with the first and/or second aspect of the invention.
In accordance with a fourth aspect, the method concerns a virtual camera transition from one perspective to another where the rendering of the virtual intermediate perspectives are altered based on one or more region(s) of interest (ROI), an ROI being a spatial function, with
“Altering” based on that spatial function means that
A difference to prior art methods is that in accordance with the prior art, the whole virtual camera image would be “inside” or “outside” but not a combination, possibly including a transition.
The fourth aspect of the invention can be combined with the first, second, and/or third aspect of the invention.
In all aspects, there can be several ROI's and (if applicable) transition regions at the same time, with specific parameters each. The transition region (if any) starts with the border of the ROI and spreads to the default region or to the next ROI. If different ROIs are present, same or different image effects can be applied to the different ROIs and the transition regions (if any) around them. If same effects are applied to different ROIs, same or different parameter values can be used for the different ROIs.
Aspects of the invention may be viewed as a method of spatially adapting image effects, which is applied and combines a single or multiple images of a video or camera transition to one output image, to a 3D scene rendering or a stereoscopic image pair. A camera transition may be the generation of images as seen by a virtual camera when moving from one camera/position to another. The start and end camera/position can be static or moving to evaluate the transition. A moving start or end position can for example be given if the input is a video of a moving camera. The camera represents an image or video of a physical camera or viewpoint or view trajectory. The image or video can be a 2D image, a 2D video, a stereo image pair a stereo video stream or a 3D scene or animated 3D scene.
The spatial dependence of the image effect may mean that the image effect will be evaluated in a different way depending on the 3D location of the corresponding 3D position (this includes the possibility that to as 3D location information, or the corresponding depth of a stereo image point described by the disparity of an image point is used), or the 2D location of an image point.
Generally, aspects and embodiments may comprise the following concepts:
In aspects and embodiments of the invention, the following characteristics are valid for the ROI:
The following possibilities can for example be used to determine the ROI:
This enumeration is not exhaustive. Combinations (as long as they are not contradictory) are possible, for example, any one of 1-5 can be combined with 7, etc.
A further, degenerate possibility of determining the ROI may be choosing the empty ROI, which results in a default image effect over the entire image, without transition regions. In most aspects and embodiments of the invention, a non-empty ROI is determined.
The image effect can be based on single or multiple 2D image effects, on single or multiple 3D based image effects or single or multiple stereo based 2D or 3D image effects. E.g. the image effect could also consist in a change of stereo disparity, which can be changed for a certain ROI resulting in one object of the scene appearing closer or further away during a camera transition.
A first example of an image effect is motion blur. For example, a plausible motion blur for the given camera transition may be applied. In a sports scene, if the camera transition is chosen in such a way, that the player, who should remain in focus of attention during the entire transition, features a relative motion on the transition images would result in a blurred appearance of this player. However, by defining a 3D ROI around this player, he can be kept in focus, although this would physically not have been the case.
A second example of an image effect is a motion blur with transition adjustment. The image effect can consist of a motion blur, which represents a plausible motion blur for the given camera transition. Again, a ROI can be defined around an object or a region, which should remain in focus. The ROI could, however, also be used to adjust the camera trajectory in order to minimize the travelled distance of pixels of the ROI during the camera transition. This results in a region/object appearing in focus, while correctly calculating motion blur for the entire image.
Depth of field: Instead of setting a fix focal length with a fix depth of field, the distance of the center of the ROI to the virtual camera can determine the focal length, while the radius of the ROI (for example projected on the ground plane) can be reflected as the depth of field. During the camera transition the image effect is being adjusted.
Enhance stereo images: A known image effect can transform a 2D video into a 3D stereo video. The perceived depth can then be increased by increasing disparity variance over the entire image. With the ROI, one could change the disparity locally, which allows making it appear to be closer or further away. This way the viewers' attention can be drawn to a specific object or region.
Embodiments of all aspects of the invention comprise building on approximate scene geometry and on a camera path used to generate virtual viewpoints being part of the images of the transition.
Embodiments of the invention may comprise the following features, alone or in combination. The features relate to all aspects of the invention unless otherwise stated:
Turning to embodiments of the image effect, the image effect can be an arbitrary effect, which is based on 2D images, 2D videos or stereo videos.
The implementation of the image effect can be any combination of the items above.
Possible image effects are:
Again, these examples of image effects may, with the partial exception of effects 2 and 3, that are only partially compatible, arbitrarily combined.
In an embodiment, a computer program product for the processing image data according to the aspects described in the above, alone or in combination, is loadable into an internal memory of a digital computer or a computer system, and comprises computer-executable instructions to cause one or more processors of the computer or computer system execute the respective method. In another embodiment, the computer program product comprises a computer readable medium having the computer-executable instructions recorded thereon. The computer readable medium preferably is non-transitory; that is, tangible. In still another embodiment, the computer program is embodied as a reproducible computer-readable signal, and thus can be transmitted in the form of such a signal.
The method can be performed by computer system comprising, typically, a processor, short- and long-term memory storage units and at least one input/output device such as a mouse, trackball, joystick, pen, touchscreen, display unit etc.
The subject matter of the invention will be explained in more detail in the following text with reference to exemplary embodiments which are illustrated in the attached drawings, in which:
A schematic example of an image is depicted in
An example relating to a sports event on a playing field defining a ground plane is described hereinafter. The ROI is chosen by drawing, with a drawing pen, an approximate ellipse on an image by hand. Into this, a mathematical ellipse approximating the drawing ellipse as well as possible is fitted. The ellipse is than projected onto the ground plane. The ellipse on the floor is then projected into the (real or virtual) camera. From a previous calibration, the relationship between the pixel number and the actual distance (in metres) is known. The transition region may be defined to be a certain region around the ellipse in on the ground (for example 3 m around the ellipse) or may be a certain pixel number (such as 100 pixels) around the ROI projected into the virtual camera. In the former case, the transition region automatically adapts if the camera for example zooms in, in the latter case it does not.
The image effect in this example is a motion blur applied to the background (to the default region, and, in part, to the transition region.). The motion blur is a combination of a velocity blur (see for example Gilberto Rosado. Motion Blur as a Post-Processing Effect, chapter 27. Addison-Wesley Professional, 2007.) in which the velocity is the parameter, and an accumulation motion blur (the averaging of the last n images, with n being a parameter). In the transition region, the respective parameter (the velocity v, the number n of images) is continuously varied from the value of the default region to 0 and 1, respectively, at the interface to the ROI.
Filing Document | Filing Date | Country | Kind | 371c Date |
---|---|---|---|---|
PCT/CH12/00077 | 4/2/2012 | WO | 00 | 10/18/2013 |
Number | Date | Country | |
---|---|---|---|
61473870 | Apr 2011 | US |