This disclosure relates to apparatus and methods.
When images are displayed to a user wearing a head mountable display (HMD), some arrangements can provide a composite representation of video material from multiple video sources. This can have the benefit of providing a user with significant amounts of material to provide interest and/or information.
It is in this context that the present disclosure arises.
The foregoing paragraphs have been provided by way of general introduction, and are not intended to limit the scope of the following claims. The described embodiments, together with further advantages, will be best understood by reference to the following detailed description taken in conjunction with the accompanying drawings.
Various aspects and features of the present disclosure are defined in the appended claims and within the text of the accompanying description.
Embodiments of the disclosure will now be described, by way of example only, with reference to the accompanying drawings, in which:
Referring now to the drawings, wherein like reference numerals designate identical or corresponding parts throughout the several views, embodiments of the present disclosure are described. In
Note that the HMD of
The HMD of
The HMD has associated headphone audio transducers or earpieces 60 which fit into the user's left and right ears 70. The earpieces 60 replay an audio signal provided from an external source, which may be the same as the video signal source which provides the video signal for display to the user's eyes.
The combination of the fact that the user can see only what is displayed by the HMD and, subject to the limitations of the noise blocking or active cancellation properties of the earpieces and associated electronics, can hear only what is provided via the earpieces, mean that this HMD may be considered as a so-called “full immersion” HMD. Note however that in some other embodiments the HMD is not a full immersion HMD, and may provide at least some facility for the user to see and/or hear the user's surroundings. This could be by providing some degree of transparency or partial transparency in the display arrangements, and/or by projecting a view of the outside (captured using a camera, for example a camera mounted on the HMD) via the HMD's displays, and/or by allowing the transmission of ambient sound past the earpieces and/or by providing a microphone to generate an input sound signal (for transmission to the earpieces) dependent upon the ambient sound.
A front-facing camera 122 (
In operation, a video signal is provided for display by the HMD. Although specific examples will be discussed below, in general terms this could be provided by an external video signal source 80 such as a video games machine or data processing apparatus (such as a personal computer), in which case the signals could be transmitted to the HMD by a wired or a wireless connection. Examples of suitable wireless connections include Bluetooth® connections. Audio signals for the earpieces 60 can be carried by the same connection. Similarly, any control signals passed from the HMD to the video (audio) signal source may be carried by the same connection. Furthermore, a power supply (including one or more batteries and/or being connectable to a mains power outlet) may be linked by a cable to the HMD. Note that the power supply and the video signal source 80 may be separate units or may be embodied as the same physical unit. There may be separate cables for power and video (and indeed for audio) signal supply, or these may be combined for carriage on a single cable (for example, using separate conductors, as in a USB cable, or in a similar way to a “power over Ethernet” arrangement in which data is carried as a balanced signal and power as direct current, over the same collection of physical wires). The video and/or audio signal may be carried by, for example, an optical fibre cable. In other embodiments, at least part of the functionality associated with generating image and/or audio signals for presentation to the user may be carried out by circuitry and/or processing forming part of the HMD itself. A power supply may be provided as part of the HMD itself.
Some embodiments of the disclosure are applicable to an HMD having at least one electrical and/or optical cable linking the HMD to another device, such as a power supply and/or a video (and/or audio) signal source. So, embodiments of the disclosure can include, for example:
If one or more cables are used, the physical position at which the cable enters or joins the HMD is not particularly important from a technical point of view. Aesthetically, and to avoid the cable(s) brushing the user's face in operation, it would normally be the case that the cable(s) would enter or join the HMD at the side or back of the HMD (relative to the orientation of the user's head when worn in normal operation). Accordingly, the position of the cables relative to the HMD in
Accordingly, the arrangement of
In the example of
Referring to
An alternative arrangement is shown in
In the case of an HMD in which the user's view of the external surroundings is entirely obscured, the mirror 210 can be a substantially 100% reflective mirror. The arrangement of
In the case where separate respective displays are provided for each of the user's eyes, it is possible to display stereoscopic images. An example of a pair of stereoscopic images for display to the left and right eyes is shown in
Note that the lateral displacements in
In some situations, an HMD may be used simply to view movies and the like. In this case, there is no change required to the apparent viewpoint of the displayed images as the user turns the user's head, for example from side to side. In other uses, however, such as those associated with virtual reality (VR) or augmented reality (AR) systems, the user's viewpoint needs to track movements with respect to a real or virtual space in which the user is located.
As mentioned above, in some uses of the HMD, such as those associated with virtual reality (VR) or augmented reality (AR) systems, the user's viewpoint needs to track movements with respect to a real or virtual space in which the user is located.
This tracking is carried out by detecting motion of the HMD and varying the apparent viewpoint of the displayed images so that the apparent viewpoint tracks the motion. The detection may be performed using any suitable arrangement (or a combination of such arrangements). Examples include the use of hardware motion detectors (such as accelerometers or gyroscopes), external cameras operable to image the HMD, and outwards-facing cameras mounted onto the HMD.
Turning to gaze tracking in such an arrangement,
In this example, an HMD 600 (with a display element 601) is provided with cameras 610 that are each arranged so as to directly capture one or more images of a respective one of the user's eyes using an optical path that does not include the lens 620. This may be advantageous in that distortion in the captured image due to the optical effect of the lens is able to be avoided. Four cameras 610 are shown here as examples of possible positions at which eye-tracking cameras may be provided, although it should be considered that any number of cameras may be provided in any suitable location so as to be able to image the corresponding eye effectively. For example, only one camera may be provided per eye or more than two cameras may be provided for each eye.
However it is considered that in a number of embodiments it is advantageous that the cameras are instead arranged so as to include the lens 620 in the optical path used to capture images of the eye. Examples of such positions are shown by the cameras 630. While this may result in processing being required to enable suitably accurate tracking to be performed, due to the deformation in the captured image due to the lens, this may be performed relatively simply due to the fixed relative positions of the corresponding cameras and lenses. An advantage of including the lens within the optical path may be that of simplifying the physical constraints upon the design of an HMD, for example.
Such an arrangement may be advantageous in that the cameras may be more easily arranged out of view of the user, for instance. Further to this, improvements to the accuracy of the eye tracking may be obtained due to the fact that the camera captures images from a position that is effectively (due to the reflection) along the axis between the user's eye and the display.
Whether an arrangement of the type shown in
As just one example, not to exclude other examples, the direction of gaze of an eye can be detected by detecting the location of the centre of the captured image of the pupil within the captured image of the cornea (whose outline is itself defined by a boundary with the sclera in the captured images). For example, a pupil centre which is central within a circular image of the cornea indicates a gaze straight ahead. Deviations of the captured pupil image in a particular direction from the central position indicate a gaze towards that direction.
Despite technical challenges including those discussed above, such tracking methods may be considered beneficial in that they allow a greater range of interactions for a user—rather than being limited to HMD viewing, gaze tracking may be performed for a viewer of a television, for instance.
Rather than varying only in the location in which cameras are provided, eye-tracking arrangements may also differ in where the processing of the captured image data to determine tracking data is performed.
By way of summary of the techniques to be described below,
Note that the method may be implemented by circuitry and/or by computer software which, when executed by a computer such as a processing device 910 to be described below, causes the computer to perform the method of
Although the presentation of multiple video sources as a composite video representation can provide a lot of interest and/or information simultaneously, it is at least potentially possible for a user to be overwhelmed or at least partially overloaded by the amount of information on the display. This can lead to a reduction in the user's attention to any individual video source and/or to user discomfort. It is an aim of embodiments of the present disclosure to at least partially alleviate such issues.
The arrangement in
While the connections shown in
The processing required to generate tracking information from captured images of the user's 800 eye or eyes may be performed locally by the HMD 810, or the captured images or results of one or more detections may be transmitted to an external device (such as the processing unit 830) for processing. In the former case, the HMD 810 may output the results of the processing to an external device for use in an image generation process if such processing is not performed exclusively at the HMD 810. In embodiments in which the HMD 810 is not present, captured images from the camera 840 are output to the processing unit 830 for processing.
As shown in
For example, the CPU 911 may be configured to generate tracking data from one or more input images of the user's eyes from one or more cameras, or from data that is indicative of a user's eye direction. This may be data that is obtained from processing images of the user's eye at a remote device, for example. Of course, should the tracking data be generated elsewhere then such processing would not be necessary at the processing device 910.
The GPU 912 may be configured to generate content for display to the user on which the eye tracking is being performed. Of course, such content generation processes may be performed elsewhere—for example, an HMD 930 may have an on-board GPU that is operable to generate content in dependence upon the eye tracking data.
The storage 913 may be provided so as to store any suitable information. Examples of such information include program data, content generation data, and eye tracking model data. In some cases, such information may be stored remotely such as on a server, and as such a local storage 913 may not be required—the discussion of the storage 913 should therefore be considered to refer to local (and in some cases removable storage media) or remote storage.
The input/output 914 may be configured to perform any suitable communication as appropriate for the processing device 910. Examples of such communication include the transmission of content to the HMD 930 and/or display 950, the reception of eye-tracking data and/or images from the HMD 930 and/or the camera 940, and communication with one or more remote servers (for example, via the internet).
As discussed above, the peripherals 920 may be provided to allow a user to provide inputs to the processing device 910 in order to control processing or otherwise interact with generated content. This may be in the form of button presses or the like, or alternatively via tracked motion to enable gestures to be used as inputs.
The HMD 930 may comprise a number of sub-elements, which have been omitted from
The camera 940 and display 950 may be configured in accordance with the discussion of the corresponding elements above with respect to
Example arrangements using the techniques of
Note that in
Note that as discussed above, in some examples the HMD may be a virtual reality HMD configured to obscure an ambient scene from the user. In other examples, the HMD may be an augmented reality HMD configured to present a representation of an ambient scene to the user (either by at least partial transparency of a front-facing portion (when worn) to allow the user wearing the HMD to view the ambient scene through that portion or by the use of a front-facing (when worn) camera to capture images of the ambient scene.
The video generator operates with respect to video images provided by the sources 1010 to generate video images for display by the HMD in the form of a composite or collage video representation of the set of two or more video sources. An example 1100 of such a composite representation is provided by
As mentioned above, the controller is configured to control the video generator to selectively vary a contribution of each of the set of two or more video sources to the composite video representation. Various example ways in which this variation can be achieved will be described further below with reference to
In some examples, the controller is configured to detect a number of the video sources (for example by communicating control signals with the video generator 1000) and to selectively vary the contribution of each of the set of two or more video sources to the composite video representation in response to the detected number. In some examples, the user may be considered to be capable of simultaneously comprehending or viewing no more than a threshold number of simultaneous video presentations within the composite representation 1100. This threshold number may be predetermined (for example, three) or may be derived by analysis of the user's behaviour (for example, by detecting that the user operates controls to shut down more than a particular number of video sources being simultaneously presented) or may be derived from user profile data, for example stored by the storage 913 and associated with the prevailing user of the HMD. In each of these techniques, however, the controller is configured to selectively vary the contribution of each of the set of two or more video sources to the composite video representation so as to limit a number of video sources contributing video material to the composite representation to a no more than a threshold number.
In other examples, the controller may be responsive to an indication or detection, for example provided by a control signal from the video generator 1000, of a total display area or at least respective display areas of the various representations of the video sources contributing to the composite representation 1100 and to selectively vary the contribution of each of the set of two or more video sources to the composite video representation in response to the detected respective display areas. For example, the controller may be configured to selectively vary the contribution of each of the set of two or more video sources to the composite video representation so as to limit a total display area of those video sources contributing video material to the composite representation to a no more than a threshold display area. Again, the threshold display area may be predetermined, may be derived from an analysis of user behaviour and/or may be associated with a user profile.
The thresholds discussed above may be considered as representations or proxies for limits on the user's ability to pay attention to the video presentation. Another possible detection which may also (or instead) contribute to a decision on the user's attention limit is as follows. In some examples, the apparatus comprises one or more user controls, the controller being responsive to user operation of the one or more user controls as an indication of detected user attention, and to selectively vary the contribution of each of the set of two or more video sources to the composite video representation in response to the detected user attention.
Similarly, as well or instead, the apparatus may comprise one or more sensors (such as the in-HMD camera or cameras discussed above with reference to
In further examples (as well or instead), the controller is responsive to user profile data indicative of user interests and to subject data associated with at least some of the set of two or more video sources, to selectively vary the contribution of the video sources in dependence upon a comparison of the user profile data and the subject data. Here, the controller may determine a video source to be reduced in prominence in response to a relative lack of overlap between the data indicative of user interests and the subject data.
Of course, it is not a requirement that the variation performed under the control of the controller 1020 involves only reducing a video source in prominence. Within a general approach by the controller to vary the composite representation 1100, one or more video sources could be reduced in prominence and one or more other video sources could be increased in prominence and/or reverted to a previously existing level of prominence.
Other examples are relevant to an augmented reality HMD of the type discussed above in which a camera such as a front-facing camera (such as the camera 122 in
Various techniques may be used to vary the presentation of a video source within the composite representation. Examples will be discussed with reference to
In a first example, with reference to
In another example, shown schematically in
In other examples, for example as shown in
Finally, with reference to
In so far as embodiments of the disclosure have been described as being implemented, at least in part, by software-controlled data processing apparatus, it will be appreciated that a non-transitory machine-readable medium carrying such software, such as an optical disk, a magnetic disk, semiconductor memory or the like, is also considered to represent an embodiment of the present disclosure. Similarly, a data signal comprising coded data generated according to the methods discussed above (whether or not embodied on a non-transitory machine-readable medium) is also considered to represent an embodiment of the present disclosure.
It will be apparent that numerous modifications and variations of the present disclosure are possible in light of the above teachings. It is therefore to be understood that within the scope of the appended claims, the technology may be practised otherwise than as specifically described herein.
Thus, the foregoing discussion discloses and describes merely exemplary embodiments of the present invention. As will be understood by those skilled in the art, the present invention may be embodied in other specific forms without departing from the spirit or essential characteristics thereof. Accordingly, the disclosure of the present invention is intended to be illustrative, but not limiting of the scope of the invention, as well as other claims. The disclosure, including any readily discernible variants of the teachings herein, defines, in part, the scope of the foregoing claim terminology such that no inventive subject matter is dedicated to the public.
Number | Date | Country | Kind |
---|---|---|---|
2202359.2 | Feb 2022 | GB | national |