This disclosure relates to head-mountable display systems.
The “background” description provided herein is for the purpose of generally presenting the context of the disclosure. Work of the presently named inventors, to the extent it is described in this background section, as well as aspects of the description which may not otherwise qualify as prior art at the time of filing, are neither expressly or impliedly admitted as prior art against the present disclosure.
A head-mountable display (HMD) is one example of a head-mountable apparatus. In an HMD, an image or video display device is provided which may be worn on the head or as part of a helmet. Either one eye or both eyes are provided with small electronic display devices.
Some HMDs allow a displayed image to be superimposed on a real-world view. This type of HMD can be referred to as an optical see-through HMD and generally requires the display devices to be positioned somewhere other than directly in front of the user's eyes. Some way of deflecting the displayed image so that the user may see it is then required. This might be through the use of a partially reflective mirror placed in front of the user's eyes so as to allow the user to see through the mirror but also to see a reflection of the output of the display devices. In another arrangement, disclosed in EP-A-1 731 943 and US-A-2010/0157433, a waveguide arrangement employing total internal reflection is used to convey a displayed image from a display device disposed to the side of the user's head so that the user may see the displayed image but still see a view of the real world through the waveguide. Once again, in either of these types of arrangement, a virtual image of the display is created (using known techniques) so that the user sees the virtual image at an appropriate size and distance to allow relaxed viewing. For example, even though the physical display device may be tiny (for example, 10 mm×10 mm) and may be just a few millimetres from the user's eye, the virtual image may be arranged so as to be perceived by the user at a distance of (for example) 20 m from the user, having a perceived size of 5 m×5 m.
Other HMDs, however, allow the user only to see the displayed images, which is to say that they obscure the real world environment surrounding the user. This type of HMD can position the actual display devices in front of the user's eyes, in association with appropriate lenses or other optical components which place a virtual displayed image at a suitable distance for the user to focus in a relaxed manner—for example, at a similar virtual distance and perceived size as the optical see-through HMD described above. This type of device might be used for viewing movies or similar recorded content, or for viewing so-called virtual reality content representing a virtual space surrounding the user. It is of course however possible to display a real-world view on this type of HMD, for example by using a forward-facing camera to generate images for display on the display devices.
Although the original development of HMDs was perhaps driven by the military and professional applications of these devices, HMDs are becoming more popular for use by casual users in, for example, computer game or domestic computing applications.
The foregoing paragraphs have been provided by way of general introduction, and are not intended to limit the scope of the following claims. The described embodiments, together with further advantages, will be best understood by reference to the following detailed description taken in conjunction with the accompanying drawings.
Various aspects and features of the present disclosure are defined in the appended claims and within the text of the accompanying description and include at least a head mountable apparatus such as a display and a method of operating a head-mountable apparatus as well as a computer program.
A more complete appreciation of the disclosure and many of the attendant advantages thereof will be readily obtained as the same becomes better understood by reference to the following detailed description when considered in connection with the accompanying drawings, wherein:
Referring now to
Note that the HMD of
The HMD of
The HMD has associated headphone audio transducers or earpieces 60 which fit into the user's left and right ears 70. The earpieces 60 replay an audio signal provided from an external source, which may be the same as the video signal source which provides the video signal for display to the user's eyes.
The combination of the fact that the user can see only what is displayed by the HMD and, subject to the limitations of the noise blocking or active cancellation properties of the earpieces and associated electronics, can hear only what is provided via the earpieces, mean that this HMD may be considered as a so-called “full immersion” HMD. Note however that in some embodiments the HMD is not a full immersion HMD, and may provide at least some facility for the user to see and/or hear the user's surroundings. This could be by providing some degree of transparency or partial transparency in the display arrangements, and/or by projecting a view of the outside (captured using a camera, for example a camera mounted on the HMD) via the HMD's displays, and/or by allowing the transmission of ambient sound past the earpieces and/or by providing a microphone to generate an input sound signal (for transmission to the earpieces) dependent upon the ambient sound.
A front-facing camera 122 may capture images to the front of the HMD, in use. A Bluetooth® antenna 124 may provide communication facilities or may simply be arranged as a directional antenna to allow a detection of the direction of a nearby Bluetooth transmitter.
In operation, a video signal is provided for display by the HMD. This could be provided by an external video signal source 80 such as a video games machine or data processing apparatus (such as a personal computer), in which case the signals could be transmitted to the HMD by a wired or a wireless connection 82. Examples of suitable wireless connections include Bluetooth® connections. Audio signals for the earpieces 60 can be carried by the same connection. Similarly, any control signals passed from the HMD to the video (audio) signal source may be carried by the same connection. Furthermore, a power supply 83 (including one or more batteries and/or being connectable to a mains power outlet) may be linked by a cable 84 to the HMD. Note that the power supply 83 and the video signal source 80 may be separate units or may be embodied as the same physical unit. There may be separate cables for power and video (and indeed for audio) signal supply, or these may be combined for carriage on a single cable (for example, using separate conductors, as in a USB cable, or in a similar way to a “power over Ethernet” arrangement in which data is carried as a balanced signal and power as direct current, over the same collection of physical wires). The video and/or audio signal may be carried by, for example, an optical fibre cable. In other embodiments, at least part of the functionality associated with generating image and/or audio signals for presentation to the user may be carried out by circuitry and/or processing forming part of the HMD itself. A power supply may be provided as part of the HMD itself.
Some embodiments of the disclosure are applicable to an HMD having at least one electrical and/or optical cable linking the HMD to another device, such as a power supply and/or a video (and/or audio) signal source. So, embodiments of the disclosure can include, for example:
(a) an HMD having its own power supply (as part of the HMD arrangement) but a cabled connection to a video and/or audio signal source;
(b) an HMD having a cabled connection to a power supply and to a video and/or audio signal source, embodied as a single physical cable or more than one physical cable;
(c) an HMD having its own video and/or audio signal source (as part of the HMD arrangement) and a cabled connection to a power supply or its own power supply; or
(d) an HMD having a wireless connection to a video and/or audio signal source and a cabled connection to a power supply or its own power supply.
If one or more cables are used, the physical position at which the cable 82 and/or 84 enters or joins the HMD is not particularly important from a technical point of view. Aesthetically, and to avoid the cable(s) brushing the user's face in operation, it would normally be the case that the cable(s) would enter or join the HMD at the side or back of the HMD (relative to the orientation of the user's head when worn in normal operation). Accordingly, the position of the cables 82, 84 relative to the HMD in
Accordingly, the arrangement of
In the example of
Referring to
An alternative arrangement is shown in
In the case of an HMD in which the user's view of the external surroundings is entirely obscured, the mirror 210 can be a substantially 100% reflective mirror. The arrangement of
In the case where separate respective displays are provided for each of the user's eyes, it is possible to display stereoscopic images. An example of a pair of stereoscopic images for display to the left and right eyes is shown in
Note that the lateral displacements in
In some situations, an HMD may be used simply to view movies and the like. In this case, there is no change required to the apparent viewpoint of the displayed images as the user turns the user's head, for example from side to side. In other uses, however, such as those associated with virtual reality (VR) or augmented reality (AR) systems, the user's viewpoint needs to track movements with respect to a real or virtual space in which the user is located.
The video displays in the HMD 20 are arranged to display images generated by the games console 300, and the earpieces 60 in the HMD 20 are arranged to reproduce audio signals generated by the games console 300. Note that if a USB type cable is used, these signals will be in digital form when they reach the HMD 20, such that the HMD 20 comprises a digital to analogue converter (DAC) to convert at least the audio signals back into an analogue form for reproduction.
Images from the camera 122 mounted on the HMD 20 are passed back to the games console 300 via the cable 82, 84. Similarly, if motion or other sensors are provided at the HMD 20, signals from those sensors may be at least partially processed at the HMD 20 and/or may be at least partially processed at the games console 300. The use and processing of such signals will be described further below.
The USB connection from the games console 300 also provides power to the HMD 20, according to the USB standard.
It will be appreciated that the localisation of processing in the various techniques described in this application can be varied without changing the overall effect, given that an HMD may form part of a set or cohort of interconnected devices (that is to say, interconnected for the purposes of data or signal transfer, but not necessarily connected by a physical cable). So, processing which is described as taking place “at” one device, such as at the HMD, could be devolved to another device such as the games console (base device) or the break-out box. Processing tasks can be shared amongst devices. Source signals, on which the processing is to take place, could be distributed to another device, or the processing results from the processing of those source signals could be sent to another device, as required. So any references to processing taking place at a particular device should be understood in this context. Similarly, where an interaction between two devices is basically symmetrical, for example where a camera or sensor on one device detects a signal or feature of the other device, it will be understood that unless the context prohibits this, the two devices could be interchanged without any loss of functionality.
As mentioned above, in some uses of the HMD, such as those associated with virtual reality (VR) or augmented reality (AR) systems, the user's viewpoint needs to track movements with respect to a real or virtual space in which the user is located.
This tracking is carried out by detecting motion of the HMD and varying the apparent viewpoint of the displayed images so that the apparent viewpoint tracks the motion.
Referring to
Consider the situation in which the user then moves his head to a new position and/or orientation 280. In order to maintain the correct sense of the virtual reality or augmented reality display, the displayed portion of the virtual environment also moves so that, at the end of the movement, a new portion 290 is displayed by the HMD.
So, in this arrangement, the apparent viewpoint within the virtual environment moves with the head movement. If the head rotates to the right side, for example, as shown in
In
The camera 322 is a video camera, capturing images at an image capture rate of, for example, 25 images per second. As each image is captured, it is passed to an image store 400 for storage and is also compared, by an image comparator 410, with a preceding image retrieved from the image store. The comparison uses known block matching techniques (so-called “optical flow” detection) to establish whether substantially the whole image has moved since the time at which the preceding image was captured. Localised motion might indicate moving objects within the field of view of the camera 322, but global motion of substantially the whole image would tend to indicate motion of the camera rather than of individual features in the captured scene, and in the present case because the camera is mounted on the HMD, motion of the camera corresponds to motion of the HMD and in turn to motion of the user's head.
The displacement between one image and the next, as detected by the image comparator 410, is converted to a signal indicative of motion by a motion detector 420. If required, the motion signal is converted by to a position signal by an integrator 430.
As mentioned above, as an alternative to, or in addition to, the detection of motion by detecting inter-image motion between images captured by a video camera associated with the HMD, the HMD can detect head motion using a mechanical or solid state detector 332 such as an accelerometer. This can in fact give a faster response in respect of the indication of motion, given that the response time of the video-based system is at best the reciprocal of the image capture rate. In some instances, therefore, the detector 332 can be better suited for use with higher frequency motion detection. However, in other instances, for example if a high image rate camera is used (such as a 200 Hz capture rate camera), a camera-based system may be more appropriate. In terms of
Other position or motion detecting techniques are of course possible. For example, a mechanical arrangement by which the HMD is linked by a moveable pantograph arm to a fixed point (for example, on a data processing device or on a piece of furniture) may be used, with position and orientation sensors detecting changes in the deflection of the pantograph arm. In other embodiments, a system of one or more transmitters and receivers, mounted on the HMD and on a fixed point, can be used to allow detection of the position and orientation of the HMD by triangulation techniques. For example, the HMD could carry one or more directional transmitters, and an array of receivers associated with known or fixed points could detect the relative signals from the one or more transmitters. Or the transmitters could be fixed and the receivers could be on the HMD. Examples of transmitters and receivers include infra-red transducers, ultrasonic transducers and radio frequency transducers. The radio frequency transducers could have a dual purpose, in that they could also form part of a radio frequency data link to and/or from the HMD, such as a Bluetooth® link.
As mentioned above in connection with
With reference to
In each of
The ideal separation of the display screen from the viewer depends upon properties of the camera system by which a 3D image is captured, and in particular relates to the distance from the camera of a so-called zero disparity point (ZDP), which is a position with respect to the camera at which the optical axes of the left and right image cameras converge. For an authentic representation of a 3D scene, the display screen should be placed at the same distance from the viewer as the ZDP of the camera system. So the two aspects are inter-related in that the convergence of the cameras can be set to provide a ZDP at a distance corresponding to the expected viewer-to-screen separation. For example, in the case of 3D material prepared for display on a television set, a typical viewer-to-television set distance may be assumed.
Referring to
Overall, it can be seen from
In the example mentioned above, the images were captured using a camera system in which the separation from the viewer (or from the camera) of the ZDP was defined by the convergence of the optical axes of the left and right cameras. In the examples to be discussed below, however, the separation of the ZDP from the viewer can be changed. Some examples relate to computer-generated images in which the position of the ZDP can be altered during generation of the images, but other examples to post-processing of captured or previously generated 3D images in order to vary the location of the ZDP.
Changing the position of a 3D display screen has particular significance in the context of head mountable displays (HMDs). In the context of a conventional (non-HMD) display screen such as a television set or a cinema screen, it is not particularly practical to vary the separation of the user and the display screen, particularly during presentation of a piece of content. But an HMD can make use of a “virtual display screen” position such that the separation from the viewer of the virtual display screen can be altered dynamically during presentation of a content item.
For the sake of the example of
As discussed above with reference to
Again, as discussed above with reference to
A particular problem, relevant primarily to HMDs, is illustrated in
A solution provided by the present disclosure is to adjust the parallax in the displayed images using (for example) a horizontal image translation so that the main image subject or region of interest (in this example, the character 630) has zero parallax. This will also lead to a corresponding but opposite change to parts of the image having a depth position different to that of the main subject. In order to provide a correct representation of the 3D scene, the virtual display screen 620 is also moved so that its separation from the viewer 600 is the same as that of the subject 630. Accordingly, the display of the character 630 is changed from the situation of
Comparing
Techniques for changing the parallax of a 3D image for display will be discussed below. First, however,
Referring back to
In the example of
So, a (potentially very slight) lateral adjustment of the position of the displayed images inwards (towards the other eye's image) brings the virtual display screen position towards the user. Lateral adjustment of the displayed image positions outwards (in a direction away from the other image) move the virtual display screen position away from the user. Note that these changes in the separation, from the user, of the virtual display screen for 3D presentation are independent of the focus distance of the eye on the optical virtual image provided by the display elements and the display optics. The focus distance can remain the same but the position, relative to the user, of the virtual display screen 720, 730 can be changed in this way.
It is possible to vary the lateral position of the displayed left and right images without any mechanical changes in the position of either the display devices 150 or the display optics 160, simply by providing display devices 150 which are slightly larger than the required size 740 of the displayed left and right images. This gives the scope to make subtle electronic changes to the position of the virtual display screen by simply laterally shifting the position, within the available display area of the display devices 150, of the displayed portion 740.
In particular,
But in general, given a ZDP location and depth parameters associated with image objects or pixels, the resulting L and R images are rendered using known techniques to provide an appropriate representation of depth when the L image is viewed by the left eye and the R image is viewed by the right eye.
While the selection of a particular VDS position in dependence upon the location of a plurality of virtual objects within a scene may provide a viewing experience that is acceptable to a viewer in many cases, there are a number of problems that may arise in such an arrangement. For example, if a viewer were to focus on object A in
By using such a VDS arrangement, the relative motion of objects as a user moves their head may appear to more closely mimic that which the user would expect from real-world experiences. While there may be some errors in the display of an object upon a change in viewpoint resulting from the shape of a VDS not exactly matching the shape of a corresponding object, the errors are generally not noticeable or significant when there is dynamic motion within the scene. Nevertheless, such an arrangement may be considered important for use in an arrangement that makes use of positional tracking.
While the VDS is illustrated as a single element in
One way in which this may be achieved is by defining a plurality of VDSs each possessing a virtual depth, with a function describing a deviation from a flat VDS—effectively, the function describes the shape of the screen with reference to a flat VDS. In such an arrangement, the depth of the VDS may be referred to as a macro-scale depth while the deviation function defines a micro-scale depth. This function may be derived using information about objects in the scene that is provided, for example in metadata or in a three dimensional model. Alternatively, or in addition, information enabling the definition of an appropriate function could be derived from disparity data either provided with the content to be displayed or derived from the content itself by image processing.
In
Of course, in embodiments in which functions are not used to define the VDSs it may be possible to simply define a VDS of a particular shape using line segments or a set of discrete points or the like, or to simply derive a VDS shape using the depth information associated with each object (for example, from a disparity map or the like).
In some embodiments the use of a function is not sufficient; for example, when defining VDSs that are curved in all three dimensions. In such embodiments it may be advantageous to define a VDS as a three-dimensional object; a mesh may be particularly appropriate. The mesh could be defined in the same manner as the VDSs as described above, for example by using a plurality of functions in conjunction so as to define a surface.
It is considered that the user often may not be able to focus on each object within a scene simultaneously; in such cases, it may be possible to only provide a non-planar VDS for objects that are focussed upon (or in the immediate vicinity of the point of viewer focus) in the virtual scene. This may be implemented using a gaze tracking arrangement (such as cameras mounted within the HMD to perform eye tracking), or by using contextual information about the displayed content to infer an object of interest, for example. The provision of gaze-dependent VDSs may be advantageous in reducing processing overheads, without significantly impacting the viewing experience. For example, if the viewer were focussed on object A in
In this arrangement, the VDS may match the shape of the tree 900 exactly (or provide a good approximation of the shape of the tree). In such a scene, it is advantageous that the contact region between the tree and the ground has zero parallax. This reduces the appearance of ‘floating’ that may be experienced from some angles, particularly during head rotations, as a result of different parallaxes causing even touching objects to move different in response to a change of viewpoint of the viewer. This may be applied at the contact point between any two objects in a virtual scene that is presented to the user.
In view of this consideration, it is possible that the only zero-disparity region that actually maps to the surface of the object is that of a contact point between the object and the ground (in the example of
The ground, being the backdrop for the tree in this scene, may be projected onto a flat VDS behind or below the VDS associated with the tree. While the ground could be grassy, it would not be considered essential that each blade of grass is associated with an individual VDS; this would generate an unnecessary processing overhead that would provide little benefit. This is because the user is unlikely to notice any errors in displaying the grass with regards to depth or the like—the blades of grass are small and not the focus of the image.
This is an example of selectively assigning VDSs to different objects within a scene; while each object could have a corresponding VDS defined only for that object, this may not be desirable. For example, static objects within a scene are not likely to be of a great interest to a viewer and so may share a flat VDS with one or more other objects with little impact on the viewing experience that is provided.
In
In
In
In yet additional or alternative examples, the mesh may be rotated based on the detected change in viewpoint of the viewer. For example, in
It will be appreciated here that the term ‘twisting’ refers to rotation of a portion of the mesh in three dimensions about one or more axes, relative to other portions of the mesh; whereas rotation refers to rotation of all of the mesh about an axis (i.e. a bulk rotation).
For example, twisting may involve rotating a portion of the mesh about one or more axes whilst other portions of the mesh are not rotated at all. This may be the case where e.g. only on the upper part of the mesh is rotated, and the lower parts of the mesh remain at their original position and orientation within the virtual scene. Alternatively or in addition, it may be that a first portion of the mesh is rotated in one direction (e.g. anti-clockwise, for example when viewed from above 1030) whilst another portion is rotated in a different direction (e.g. clockwise 1032). Generally, twisting a portion of the mesh results in a normal vector 1026 located at the surface of that portion being pointed in a different direction relative to normal vectors located at other points on the surface of the mesh. In some examples, the twisting may result in the normal vector 1026 at the twisted portion being rotated to face, or be closer to facing, the vector defining the viewer's overall gaze direction. An example of this is shown in
Generally, the shape of the VDS is changed based on the detected change in viewpoint of the stereoscopic image, by the viewer. In some examples, a virtual scene may include a plurality of VDSs with each of these having a stereoscopic image of an object projected thereon. In these examples, the object that a viewer is viewing may be detected, and the VDS associated with that object may be rotated and/or twisted, based on a detected change in viewpoint of the same object, by the viewer. The object that a user is looking at may be determined based e.g. on the pose of a HMD that viewer is wearing and/or on the gaze direction of the viewer.
In an embodiment in which objects are presented using a flat VDS, it is usually considered that the display of the objects may be compromised as a result of mapping a curved (or otherwise non-planar) surface onto a flat surface. However, this may be mitigated in the present arrangement using a depth map that describes the depth of an object relative to the VDS. Objects may be defined to be in front of or behind the VDS relative to the viewer, with a value indicating how far away from the VDS the object is, whilst still being projected onto the VDS.
When the viewpoint within a virtual scene moves, objects may have a relative motion applied to them that is dependent upon the value defining the depth relative to the VDS. For example, an object that is in front of the VDS may move in the opposite direction to the viewpoint whilst objects behind the VDS may move in the same direction as the viewpoint. The magnitude of this motion may be proportional to the depth difference between the VDS and the object that is defined in the depth map.
In some embodiments, the depth map uses a 0-255 register for each object (or other element within the image, such as a particular area of the image) to describe the depth relative to the VDS. A value of 127 indicates that an object is at the same depth as the VDS, while values of 0 and 255 indicate a maximum distance in front of/behind the VDS respectively. This maximum distance may be defined for each scene individually; for example, the maximum distance in front could indicate being at the position of the viewpoint in the scene with an equal maximum distance behind, with the intermediate depths being scaled accordingly. Alternatively, a maximum distance from a VDS may be defined as a number of arbitrary units within the scene or as a percentage of the depth range of the whole scene. In such an embodiment, the depth map may be encoded in the alpha channel of the transmitted content or may be provided as separate metadata for a scene.
It may be advantageous, particularly in more complex scenes comprising a larger number of objects, that the virtual scene is provided in a layered manner. For example, a plurality of images could be captured of different viewpoints within a scene so as to record sufficient information about the depth of objects within a scene.
The centre of the room is denoted by a filled circle, which corresponds to the position of a viewpoint (for rendering or capturing an image) with objects A, B, C and D arranged within the room 940. Each of the circles represents a depth layer boundary, so as to allow a distinction between different depth regions within a virtual environment.
Object A appears in the first depth layer, which is the same depth layer as the central viewpoint. Object B appears in the second depth layer (arranged so as to surround the first depth layer), object C appears in the third depth layer and object D appears in the fourth depth layer. The layers shown in
It is apparent that the outermost depth layer extends beyond the perimeter of the room; the information in this layer would therefore not include any information for the regions which are obscured by the walls of the room (as these regions are not visible from any angle) and instead would comprise a series of discrete areas (corresponding to the corners of the room).
Each of these depth layers may be rendered separately, in any order. For example, a painter's algorithm could be used to as to render the scene starting from the outermost depth layer and moving towards the centre of the scene.
While shown to be completely distinct, in some embodiments the layers may overlap in the depth direction. This means that multiple images are captured of each object in the environment, and this may be utilised to generate more accurate depth data than in an arrangement in which only a single image is captured of each object.
While the above discussion has been provided with reference to VDSs that are stationary, in some embodiments it is advantageous for content to be provided using dynamic VDSs. For example, a mesh that represents a person in a video may be controlled to move so as to follow the motion of the person within the video. This ensures that the correct display of an object may be maintained over several frames, rather than the motion of the person causing a separation between the person in the image and the VDS to which the image of the person was initially mapped.
The VDSs as described above may be suitable for use with any projection type for the content as a whole. For example, a cubic projection may be used, with VDSs defined relative to this. An equirectangular projection mapping may be a particularly advantageous projection to use in conjunction with the VDSs, as this is a projection which most closely matches the geometry generated when capturing content as a 360 degree panoramic. As a result of this, the number of display errors may be reduced.
A step 3300 comprises acquiring virtual scene information. This comprises obtaining mesh and texture information required to generate an image of a virtual scene for display, for example, or any other form of image data that may be used to generate an image for display representing at least a portion of a virtual environment. This information may be obtained from any suitable source; for example, a removable storage medium or the internet.
A step 3310 comprises setting the position of the VDS in the virtual scene, in dependence upon the position of objects to be displayed within the virtual scene. The setting of a VDS position for a stereoscopic image for display is performed such that the depth profile of one or more VDSs approximates the depth profile of one or more objects in a virtual scene to be displayed by that stereoscopic image.
A step 3320 comprises generating an image for display. This image is generated so as to have an image parallax appropriate to the VDS position or positions set by the setting step 3330.
A step 3400 comprises the identification of significant objects within the virtual scene. A significant object is one that is to be considered when generating the VDS; for example, objects that are important to a story may be considered in preference to background objects such as scenery.
In some embodiments, only one of two identical objects may be significant if it is only determined that the VDS should be generated in dependence upon one of the two identical objects. Significant objects may be recognised in any of a number of ways.
In some embodiments, a significant object is identified in metadata associated with the virtual scene or video/image content. Alternatively, or in addition, processing may be performed so as to identify the objects of interest; this could take the form of motion detection (as objects that move more in scenes often are more important) or colour detection (as important objects tend to be of a different colour to the background, to ensure visibility). A further alternative or additional method is to allow a user to identify objects of significance manually.
In some embodiments, the significant objects may be determined or defined with a degree of significance or some other manner of distinguishing between the significant objects in dependence upon their significance. This may be advantageous in that in systems with limited processing power or the like, only the most significant objects may have a corresponding VDS that is defined. In the context of the above examples, a significance may be defined in the associated metadata or objects may be assessed to be of different levels of significance based upon amount of motion, speed of motion, object size, the severity of difference in colour (such that a red object on a green background is seen as more significant than an object that is a different shade of green would be, for example), or using any other suitable indicator of potential significance.
Of course, this step may be omitted entirely; in some arrangements, there may be no distinction between significant and insignificant objects such that a VDS is generated for each object within a scene. In this case, the steps below are performed for each object rather than only significant objects.
A step 3410 comprises locating the significant objects; this may be in the virtual environment in which they are present, their location in the image frame that is to be displayed, or simply relative to one another for example. This may be performed by using a depth map, for example, or by reading out such information from metadata associated with the content.
A step 3420 comprises defining one or more VDSs using the location of the significant objects. As described above, this may comprise the definition of a VDS shape and location in the virtual scene. In some embodiments, information about the object (such as a shape and size) is used in order to define a VDS that is approximately at the surface of the object. In some other embodiments, a VDS is defined using the position of the object only; this may mean that the VDS intersects the object.
A step 3500 comprises identifying significant objects, in a manner similar to the step 3400 of
A step 3510 comprises locating object surfaces for the identified significant objects. As in step 3410 of
A step 3520 comprises defining a VDS in dependence upon the surface information that is acquired in step 3510. This may be performed by simply defining the VDS to be the same as the surface information, or a simplified VDS may be defined (for example, by approximating the surface information with a simplified function) so as to reduce computational complexity. The VDS may be defined as a mathematical function, or by defining a depth for the VDS at each point or each of a plurality of small areas (such that a number of tiles may make up the VDS) that each approximate a corresponding area of the surface.
At step 3600 a VDS position for displaying a stereoscopic image is set. The setting of the position of the VDS may be set according to any of techniques described previously.
At step 3610 a stereoscopic image for display at the VDS is generated. As described previously in relation to
At step 3620 a change in viewpoint of the viewer of the stereoscopic image is detected. If the viewer is wearing an HMD, this change in viewpoint may be detected based on a change in pose of the HMD. For example, it may generally be assumed that the user will be looking at a central region of the display element of the HMD, and so the change in viewpoint of the corresponding virtual scene can be inferred from a detected change in pose of the HMD. In other examples, this change in viewpoint may be detected without an HMD, e.g. by detecting a pose and/or gaze direction of a viewer, e.g. using one or more cameras to capture images of the viewer's eyes.
At step 3630 a configuration of the VDS is controlled, based on the detected change in viewpoint of the viewer of the stereoscopic image. The shape of the VDS may be defined by a three-dimensional mesh. The configuration of the VDS may be controlled by controlling a rotation of the mesh, for example. Alternatively or in addition, the configuration of the VDS may be controlled by twisting a portion of the mesh. The twisting may correspond to rotating a portion of the mesh towards the viewer. The twisting may be such that a surface normal at the twisted portion of the mesh is rotated towards the viewer.
The rotation and/or twisting of the mesh may result in a corresponding rotation and/or twisting of the stereoscopic image that is projected onto the mesh In one example, the upper part of the mesh may be twisted such that at least one the head and shoulders of a human in the stereoscopic image is turned towards the viewer. In this way, the human in the stereoscopic image may appear to remain facing the viewer, despite the viewing having changed their viewpoint of the stereoscopic image. As mentioned previously, the stereoscopic image may be a live-action video of a performer (not live as in being captured in real-time, but live as in the video is of a real person performing).
In some examples, the stereoscopic image may comprise a first object and a second object. The first object may be e.g. a human and the second object may be e.g. an instrument or a physical surface such as the floor that the human is contact with. In such examples, the VDS may be shaped so as to conform to the shape of the first object at the point of contact between the first object and the second object. As mentioned previously, this shaping of the VDS ensures that the first object does not appear to float relative to the second object, which may appear unnatural. The shape of the VDS may be substantially flat (for example, perpendicular to a direction from the viewer) at the regions of the stereoscopic image in which the first object is not in contact with the second object.
It will be appreciated that while the method described in relation
It will be further be appreciated that, while the change in configuration has been described in relation to a mesh, the change may correspond to a change in shape of any surface onto which the stereoscopic image(s) is (or are) projected. Moreover, the change in configuration need not be limited to a rotation or twisting operation, but may more generally involve controlling a three-dimensional shape of the mesh.
The virtual scene information acquiring unit 3700 is operable to acquire information about a virtual scene. This may be acquired in the form of mesh and texture data for a virtual scene, information from metadata associated with a virtual scene or image content, or information derived by performing some form of image analysis on an image (or images) for display, for example.
The VDS setting unit 3710 is operable to set the position of one or more VDSs for a virtual scene. This may include defining a particular shape, so as to approximate (or replicate) the surface of an object in the scene, for example.
The image generating unit 3720 is operable to generate a image for display in dependence upon the one or more VDSs that are set by the VDS setting unit 3710. For example, this may comprise using mesh and texture information about a virtual scene in conjunction with the information about the one or more VDSs to generate an image for display.
The image display unit 3730 may be an HMD or any other hardware capable of displaying 3D content to a viewer.
The object identification unit 3800 is operable to identify which objects are to have an associated VDS defined. As discussed above, this may comprise identifying significant objects in a scene based upon metadata or image/scene analysis methods.
The object locating unit 3810 is operable to locate one or more of the objects that are identified by the object identification unit 3800. In some embodiments, each of the identified objects are located, while in others only a portion of the objects may be located (for example, to reduce a processing burden on the image generating device). The object locating unit 3810 may be operable to derive the location of objects based upon image processing techniques, user input, associated metadata, or any other suitable method. In some embodiments, the object is located while in others it is the surface of the object that is located.
The VDS defining unit 3820 is operable to define a VDS in dependence upon the location of the identified objects. As noted above, this may comprise setting one or more VDSs that intersect one or more of the one or more identified objects, or setting the one or more VDSs that approximate (or replicate) the surfaces of the identified objects.
The viewpoint detector 3930 is configured to detect a change in viewpoint of a viewer of a stereoscopic image generated by the image generation unit. The viewpoint detector 3930 may be configured to detect at least one of a pose of the viewer's head and a gaze direction of the viewer. In some examples, the viewpoint detector may be configured to detect the change in viewpoint based on a change in pose of an HMD that the viewer is wearing to view the virtual scene (which itself comprises the VDS onto which the stereoscopic image is projected). The change in pose of the HMD may be detected using the motion detectors described previously. Changes in the gaze of the viewer may be detected using e.g. one or more cameras, such as infra-red cameras, within the HMD.
The VDS configuration unit is operable to modify a configuration of a VDS, in response to the detected change in viewpoint of the object being displayed by that VDS. In some examples, the VDS may comprise a mesh and the configuration of the VDS may be modified by rotating and/or twisting the mesh. The VDS configuration unit may be operable to determine a rotation and/or twisting of the mesh that results in a portion of the stereoscopic image continuing to face the viewer, at the viewer's new viewpoint. Moreover, the VDS configuration unit may also be operable to control a deformation of the mesh, in addition to a rotation or twisting (i.e by controlling an overall 3D shape of the mesh and not just a complete or partial rotation thereof). In some examples, the stereoscopic image may be a live-action video of a performer and the VDS configure unit may be operable to twist an upper part of the corresponding mesh. As described above, this twisting may be such that the head of the performer remains facing (i.e. looking at) the viewer. Whilst the above examples have been described in relation to a human performer, it will be appreciated that the stereoscopic image may be a live-action video of any animated object.
The video signal comprises stereoscopic video content 4010, VDS position information 4520, VDS depth profile information 4030, shape modification information 404 and an additional metadata field 4050.
The stereoscopic video content 4010 comprises information that may be displayed to a viewer of a display device. For example, this may comprises pre-rendered images, or mesh and texture information.
The VDS position information 4020 defines the location of one or more VDSs in the scene represented by the stereoscopic video content 4010. This may comprise coordinates within the scene, for example, or an identification of which pixels correspond to the screen. Of course, any suitable method for defining the location may be used, such as linking the location of the VDS to that of an object within the scene.
The VDS depth profile information 4030 comprises information about the depth profile of the one or more VDSs. The depth profile of a VDS is a description of the shape of the VDS; for example, the depth profile may be a surface that is defined at a position corresponding to the VDS position information 4020. As described above, a depth profile may be defined using any suitable method rather than being limited to a function defining a surface or the like.
The VDS configuration information 4040 comprises information defining change information in respect of the configuration of at least one of the VDSs. The change in configuration of the at least one VDS is defined such that the change in configuration depends on a detected change in viewpoint of the object displayed by the at least one VDS. In some examples, the change in configuration corresponds to a rotating and/or twisting of a mesh. The rotation and/or twisting of the mesh may be defined such that at least part of the mesh is turned towards or to face the viewer, at the viewer's new respective viewpoint.
The additional metadata field 4050 may comprise any additional information that is useful for defining the one or more VDSs. For example, a list of object IDs could be provided so as to associate a VDS with an object in the scene represented by the stereoscopic video content 4010. Alternatively, or in addition, information about the motion of a VDS may be provided in this field 4040.
It will be appreciated that embodiments of the present invention may be implemented in hardware, programmable hardware, software-controlled data processing arrangements or combinations of these. It will also be appreciated that computer software or firmware used in such embodiments, and providing media for providing such software or firmware (such as storage media, for example a machine-readable non-transitory storage medium such as a magnetic or optical disc or a flash memory) are considered to represent embodiments of the present invention.
Number | Date | Country | Kind |
---|---|---|---|
1715555 | Sep 2017 | GB | national |
Filing Document | Filing Date | Country | Kind |
---|---|---|---|
PCT/GB2018/052671 | 9/19/2018 | WO | 00 |
Publishing Document | Publishing Date | Country | Kind |
---|---|---|---|
WO2019/063976 | 4/4/2019 | WO | A |
Number | Name | Date | Kind |
---|---|---|---|
5880711 | Tamada | Mar 1999 | A |
10559019 | Beauvais | Feb 2020 | B1 |
20100157433 | Mukawa | Jun 2010 | A1 |
20120019528 | Ugawa | Jan 2012 | A1 |
20120229604 | Boyce | Sep 2012 | A1 |
20140009636 | Lee | Jan 2014 | A1 |
20160232710 | Herman | Aug 2016 | A1 |
20160234482 | Bickerstaff | Aug 2016 | A1 |
20160327790 | Kasano | Nov 2016 | A1 |
20170316605 | Herman | Nov 2017 | A1 |
20180003981 | Urey | Jan 2018 | A1 |
20190011621 | Karafin | Jan 2019 | A1 |
Number | Date | Country |
---|---|---|
1731943 | Dec 2006 | EP |
3070513 | Sep 2016 | EP |
3171592 | May 2017 | EP |
Entry |
---|
International Search Report and Written Opinion for corresponding PCT Application No. PCT/GB2018/052671, 19 pages, dated Feb. 14, 2019. |
Combined Search and Examination Report for corresponding GB Application No. GB 1715555.7, 5 pages, dated Mar. 6, 2018. |
Number | Date | Country | |
---|---|---|---|
20200296354 A1 | Sep 2020 | US |