This disclosure relates to virtual reality systems and methods.
The “background” description provided herein is for the purpose of generally presenting the context of the disclosure. Work of the presently named inventors, to the extent it is described in this background section, as well as aspects of the description which may not otherwise qualify as prior art at the time of filing, are neither expressly or impliedly admitted as prior art against the present disclosure.
A head-mountable display (HMD) is one example of a head-mountable apparatus for use in a virtual reality system in which an HMD wearer views a virtual environment. In an HMD, an image or video display device is provided which may be worn on the head or as part of a helmet. Either one eye or both eyes are provided with small electronic display devices.
Some HMDs allow a displayed image to be superimposed on a real-world view. This type of HMD can be referred to as an optical see-through HMD and generally requires the display devices to be positioned somewhere other than directly in front of the user's eyes. Some way of deflecting the displayed image so that the user may see it is then required. This might be through the use of a partially reflective mirror placed in front of the user's eyes so as to allow the user to see through the mirror but also to see a reflection of the output of the display devices. In another arrangement, disclosed in EP-A-1 731 943 and US-A-2010/0157433, a waveguide arrangement employing total internal reflection is used to convey a displayed image from a display device disposed to the side of the user's head so that the user may see the displayed image but still see a view of the real world through the waveguide. Once again, in either of these types of arrangement, a virtual image of the display is created (using known techniques) so that the user sees the virtual image at an appropriate size and distance to allow relaxed viewing. For example, even though the physical display device may be tiny (for example, 10 mm×10 mm) and may be just a few millimetres from the user's eye, the virtual image may be arranged so as to be perceived by the user at a distance of (for example) 20 m from the user, having a perceived size of 5 m×5 m.
Other HMDs, however, allow the user only to see the displayed images, which is to say that they obscure the real world environment surrounding the user. This type of HMD can position the actual display devices in front of the user's eyes, in association with appropriate lenses or other optical components which place a virtual displayed image at a suitable distance for the user to focus in a relaxed manner—for example, at a similar virtual distance and perceived size as the optical see-through HMD described above. This type of device might be used for viewing movies or similar recorded content, or for viewing so-called virtual reality content representing a virtual space surrounding the user. It is of course however possible to display a real-world view on this type of HMD, for example by using a forward-facing camera to generate images for display on the display devices.
Although the original development of HMDs and virtual reality was perhaps driven by the military and professional applications of these devices, HMDs are becoming more popular for use by casual users in, for example, computer game or domestic computing applications.
The foregoing paragraphs have been provided by way of general introduction, and are not intended to limit the scope of the following claims. The described embodiments, together with further advantages, will be best understood by reference to the following detailed description taken in conjunction with the accompanying drawings.
Various aspects and features of the present disclosure are defined in the appended claims and within the text of the accompanying description and include at least a head mountable apparatus such as a display and a method of operating a head-mountable apparatus as well as a computer program.
A more complete appreciation of the disclosure and many of the attendant advantages thereof will be readily obtained as the same becomes better understood by reference to the following detailed description when considered in connection with the accompanying drawings, wherein:
Referring now to
Note that the HMD of
The HMD of
The HMD has associated headphone audio transducers or earpieces 60 which fit into the user's left and right ears 70. The earpieces 60 replay an audio signal provided from an external source, which may be the same as the video signal source which provides the video signal for display to the user's eyes. A boom microphone 75 is mounted on the HMD so as to extend towards the user's mouth.
The combination of the fact that the user can see only what is displayed by the HMD and, subject to the limitations of the noise blocking or active cancellation properties of the earpieces and associated electronics, can hear only what is provided via the earpieces, mean that this HMD may be considered as a so-called “full immersion” HMD. Note however that in some embodiments the HMD is not a full immersion HMD, and may provide at least some facility for the user to see and/or hear the user's surroundings. This could be by providing some degree of transparency or partial transparency in the display arrangements, and/or by projecting a view of the outside (captured using a camera, for example a camera mounted on the HMD) via the HMD's displays, and/or by allowing the transmission of ambient sound past the earpieces and/or by providing a microphone to generate an input sound signal (for transmission to the earpieces) dependent upon the ambient sound.
A front-facing camera 122 may capture images to the front of the HMD, in use. A Bluetooth® antenna 124 may provide communication facilities or may simply be arranged as a directional antenna to allow a detection of the direction of a nearby Bluetooth transmitter.
In operation, a video signal is provided for display by the HMD. This could be provided by an external video signal source 80 such as a video games machine or data processing apparatus (such as a personal computer), in which case the signals could be transmitted to the HMD by a wired or a wireless connection 82. Examples of suitable wireless connections include Bluetooth® connections. Audio signals for the earpieces 60 can be carried by the same connection. Similarly, any control signals passed from the HMD to the video (audio) signal source may be carried by the same connection. Furthermore, a power supply 83 (including one or more batteries and/or being connectable to a mains power outlet) may be linked by a cable 84 to the HMD. Note that the power supply 83 and the video signal source 80 may be separate units or may be embodied as the same physical unit. There may be separate cables for power and video (and indeed for audio) signal supply, or these may be combined for carriage on a single cable (for example, using separate conductors, as in a USB cable, or in a similar way to a “power over Ethernet” arrangement in which data is carried as a balanced signal and power as direct current, over the same collection of physical wires). The video and/or audio signal may be carried by, for example, an optical fibre cable. In other embodiments, at least part of the functionality associated with generating image and/or audio signals for presentation to the user may be carried out by circuitry and/or processing forming part of the HMD itself. A power supply may be provided as part of the HMD itself.
Some embodiments of the disclosure are applicable to an HMD having at least one electrical and/or optical cable linking the HMD to another device, such as a power supply and/or a video (and/or audio) signal source. So, embodiments of the disclosure can include, for example:
(a) an HMD having its own power supply (as part of the HMD arrangement) but a cabled connection to a video and/or audio signal source;
(b) an HMD having a cabled connection to a power supply and to a video and/or audio signal source, embodied as a single physical cable or more than one physical cable;
(c) an HMD having its own video and/or audio signal source (as part of the HMD arrangement) and a cabled connection to a power supply; or
(d) an HMD having a wireless connection to a video and/or audio signal source and a cabled connection to a power supply.
If one or more cables are used, the physical position at which the cable 82 and/or 84 enters or joins the HMD is not particularly important from a technical point of view. Aesthetically, and to avoid the cable(s) brushing the user's face in operation, it would normally be the case that the cable(s) would enter or join the HMD at the side or back of the HMD (relative to the orientation of the user's head when worn in normal operation). Accordingly, the position of the cables 82, 84 relative to the HMD in
Accordingly, the arrangement of
In the example of
Referring to
An alternative arrangement is shown in
In the case of an HMD in which the user's view of the external surroundings is entirely obscured, the mirror 210 can be a substantially 100% reflective mirror. The arrangement of
In the case where separate respective displays are provided for each of the user's eyes, it is possible to display stereoscopic images. An example of a pair of stereoscopic images for display to the left and right eyes is shown in
Note that the lateral displacements in
In some situations, an HMD may be used simply to view movies and the like. In this case, there is no change required to the apparent viewpoint of the displayed images as the user turns the user's head, for example from side to side. In other uses, however, such as those associated with virtual reality (VR) or augmented reality (AR) systems, the user's viewpoint needs to track movements with respect to a real or virtual space in which the user is located.
Each Sony® Move® controller or a controller of a similar type suited to virtual reality experiences typically comprises one or more a mechanical or solid state sensors such as accelerometers and/or gyroscopes, enabling the controller to transmit telemetry to the games console indicative of the controller's placement in space. For example if such a controller detects acceleration, this information can be used by the controller or the games console to determine relative changes in speed and/or changes in position as well as or instead of monitoring acceleration itself. In this way, the games console can track the relative orientation, position and/or movement of the or each controller by use of the received telemetry.
Alternatively or in addition, such controllers may comprise a predetermined visible feature, such as a glowing sphere, pattern of lights, fiduciary marker, high contrast logo or the like. This feature can be detected in a video image captured by a video camera 302 operably coupled to the games console, so as to assist with tracking of a controller's position in space by use of image analysis. A single or monoscopic video camera may only be operable to track the controller's position on a 2D plane, whereas a dual or stereoscopic video camera (as shown) may be operable to track the controller's position in a 3D space.
The video displays in the HMD 20 are arranged to display images generated by the games console 300, and the earpieces 60 in the HMD 20 are arranged to reproduce audio signals generated by the games console 300. Note that if a USB type cable is used, these signals will be in digital form when they reach the HMD 20, such that the HMD 20 comprises a digital to analogue converter (DAC) to convert at least the audio signals back into an analogue form for reproduction.
Images from the camera 122 mounted on the HMD 20 are passed back to the games console 300 via the cable 82, 84. Similarly, if motion or other sensors are provided at the HMD 20, signals from those sensors may be at least partially processed at the HMD 20 and/or may be at least partially processed at the games console 300. The use and processing of such signals will be described further below.
The USB connection from the games console 300 also provides power to the HMD 20, according to the USB standard.
It will be appreciated that the localisation of processing in the various techniques described in this application can be varied without changing the overall effect, given that an HMD may form part of a set or cohort of interconnected devices (that is to say, interconnected for the purposes of data or signal transfer, but not necessarily connected by a physical cable). So, processing which is described as taking place “at” one device, such as at the HMD, could be devolved to another device such as the games console (base device) or the break-out box. Processing tasks can be shared amongst devices. Source signals, on which the processing is to take place, could be distributed to another device, or the processing results from the processing of those source signals could be sent to another device, as required. So any references to processing taking place at a particular device should be understood in this context. Similarly, where an interaction between two devices is basically symmetrical, for example where a camera or sensor on one device detects a signal or feature of the other device, it will be understood that unless the context prohibits this, the two devices could be interchanged without any loss of functionality.
As mentioned above, in some uses of the HMD, such as those associated with virtual reality (VR) or augmented reality (AR) systems, the user's viewpoint needs to track movements with respect to a real or virtual space in which the user is located.
This tracking is carried out by detecting motion of the HMD and varying the apparent viewpoint of the displayed images so that the apparent viewpoint tracks the motion.
Referring to
Consider the situation in which the user then moves his head to a new position and/or orientation 280. In order to maintain the correct sense of the virtual reality or augmented reality display, the displayed portion of the virtual environment also moves so that, at the end of the movement, a new portion 290 is displayed by the HMD.
So, in this arrangement, the apparent viewpoint within the virtual environment moves with the head movement. If the head rotates to the right side, for example, as shown in
In
The camera 322 is a video camera, capturing images at an image capture rate of, for example, 25 images per second. As each image is captured, it is passed to an image store 400 for storage and is also compared, by an image comparator 410, with a preceding image retrieved from the image store. The comparison uses known block matching techniques (so-called “optical flow” detection) to establish whether substantially the whole image has moved since the time at which the preceding image was captured. Localised motion might indicate moving objects within the field of view of the camera 322, but global motion of substantially the whole image would tend to indicate motion of the camera rather than of individual features in the captured scene, and in the present case because the camera is mounted on the HMD, motion of the camera corresponds to motion of the HMD and in turn to motion of the user's head.
The displacement between one image and the next, as detected by the image comparator 410, is converted to a signal indicative of motion by a motion detector 420. If required, the motion signal is converted by to a position signal by an integrator 430.
As mentioned above, as an alternative to, or in addition to, the detection of motion by detecting inter-image motion between images captured by a video camera associated with the HMD, the HMD can detect head motion using a mechanical or solid state detector 332 such as an accelerometer. This can in fact give a faster response in respect of the indication of motion, given that the response time of the video-based system is at best the reciprocal of the image capture rate. In some instances, therefore, the detector 332 can be better suited for use with higher frequency motion detection. However, in other instances, for example if a high image rate camera is used (such as a 200 Hz capture rate camera), a camera-based system may be more appropriate. In terms of
Other position or motion detecting techniques are of course possible. For example, a mechanical arrangement by which the HMD is linked by a moveable pantograph arm to a fixed point (for example, on a data processing device or on a piece of furniture) may be used, with position and orientation sensors detecting changes in the deflection of the pantograph arm. In other embodiments, a system of one or more transmitters and receivers, mounted on the HMD and on a fixed point, can be used to allow detection of the position and orientation of the HMD by triangulation techniques. For example, the HMD could carry one or more directional transmitters, and an array of receivers associated with known or fixed points could detect the relative signals from the one or more transmitters. Or the transmitters could be fixed and the receivers could be on the HMD. Examples of transmitters and receivers include infra-red transducers, ultrasonic transducers and radio frequency transducers. The radio frequency transducers could have a dual purpose, in that they could also form part of a radio frequency data link to and/or from the HMD, such as a Bluetooth® link.
As mentioned above in connection with
With reference to
In
The HMD 500 or an associated console is optionally associated with a user control 502 for operation by the user or wearer of that HMD. Similarly, a user control 512 is optionally associated with the HMD 510 or an associated console (not shown). An example of a user control is a controller such as the controller 330 described above and/or a camera 302. In examples, the user may hold one or two such controllers 330, one in each hand, in the manner shown in
A detector/image processor 504 operates with respect to the HMD 500, and a detector/image processor 514 operates with respect to the HMD 510. Some image processing functions may be carried out by a shared image processing resource 520, for example a server in data communication with both users' local systems. A respective computing device (such as a games console) of each user and/or a remote server may fulfil this role.
The two (or more, as the case may be) users share a common virtual environment. The virtual environment may contain environment features with respect to a coordinate system common to all users of the virtual environment. A simple example of a virtual environment is shown schematically in
Therefore, in some examples, communication is on a peer-to-peer basis between the respective PlayStation devices associated with the HMDs. In examples, each peer device (for example, an image processor (acting as an image generator) of each peer device) does the following:
In examples, at least some of these operations can be carried out by an image generator as part of the HMD/PlayStation device.
In other examples, a server-based network is used, such that a separate server provides coordination of the virtual environment. This could include receiving information from the HMD/PlayStation devices about operation of their respective controls and then broadcasting messages to each HMD/PlayStation device indicating facial and body configurations and other aspects. In further examples, a PlayStation device could render aspects of its own avatar and send the rendered image data to other HMD/PlayStation devices in the network. However, for economy of data traffic and potentially reduced latency, the peer-to-peer network discussed above is used in the present embodiments.
The arrangement of
An example 540 of an avatar's viewpoint is shown in
Note that in other arrangements the user may choose to move the user's virtual camera to another position with respect to the user's avatar, for example giving a third person view of the user, and hence visibility of their avatar's facial expression. An example is during movement (locomotion). However, the issues discussed above can occur during at least a first person view.
The generation of images of the virtual environment is handled by the image processors 504, 514 and/or 520. Various examples will now be discussed, all of which are functionally equivalent in terms of their end result, which is the generation of appropriate images for display at the HMDs 500, 510. In each instance, the images of the virtual environment are dependent upon aspects such as: the operation of user controls by the HMD users, optionally the use of microphones associated with the HMDs and the position or orientation of the HMDs, because those aspects determine the position, configuration, motion or actions of an avatar associated with the respective HMD user.
In some examples, the image generation is carried out by the shared image processing resource 520, for example a server. In this arrangement, each detector/image processor 504, 514 detects user control operation, optionally microphone operation, and HMD orientation and/or position, and sends data indicative of these aspects to the image processor 520, which then generates images of the virtual environment for display by each HMD. Note that generally speaking, the view provided to each HMD will be different, because it depends upon the respective position and orientation of the corresponding avatar.
In other examples, each HMD is associated with a respective image processor (for example, the image processor 504 associated with the HMD 500) which generates images for display by that HMD, in dependence upon data common to each HMD defining the virtual environment, and data received from detector/image processors associated with other HMDs defining aspects such as the current position and configuration of the respective avatars.
In other examples, each HMD is associated with a respective image processor (for example, the image processor 504 associated with the HMD 500) which generates image data relating to the avatar corresponding to that HMD, for use by other image processors to render images for display by their own respective HMD.
A detector 600 receives signals from the user control 502 and optionally also from the microphone 560 associated with that HMD. This corresponds to the operation (a) discussed above.
As mentioned above, an example of the user control 502 is one or a pair of Sony® Move® controllers, each held in one hand. The or each Move controller provides a set of buttons for operation by a thumb and another finger. As noted previously herein, the or each Move controller also has one or more sensors by which the controller's position and orientation in space can be detected. Activations of the buttons, and the detected position and/or orientation of the or each controller, are examples of signals sent to the detector 600. Audio signals representing audio information captured by the microphone 560 are other examples of signals that may be sent to the detector 600. Signals from the HMD 500 indicating the current orientation of the HMD are other examples of signals that may be sent to the detector 600. The detector 600 analyses the received signals to detect a current operational mode of the user controls, the HMD and/or the microphone as appropriate.
Based on the detection by the detector 600, in one embodiment of the present invention a hand configuration generator 610 generates a hand configuration for the respective avatar, providing an example of the use of one or more user controls, the image generator being responsive to operation of the user controls by the user to configure the hands of the avatar representing that user. For example this can be carried out by reference to a mapping table 620 which maps operations of the buttons on the Move controllers to respective hand configurations. In examples, the avatar has the same number of hands (two) as the user. In this case, the Move controller held by the user's left hand controls the configuration of the avatar's left hand, and the Move controller held by the user's right hand controls the configuration of the avatar's right hand. This is an example of the user controls comprising respective controls allowing separate configuration of each of the avatar's hands. The generator 610 contributes to the operations (b), (c) and (d) discussed above.
The mapping table 620 maps button operation on a controller to a hand configuration. For example, where the operation of two buttons is detected, this gives four possibilities, shown by the following example mapping table (where Button A and Button B are identifiers simply used to distinguish the two different buttons; for example, Button A might be the thumb-operated button and Button B the other-finger-operated button):
The latter three hand configurations (apart from the base configuration) are examples of active configurations assumed by that avatar hand in response to operation of the respective controls.
In some examples, the buttons need not be binary “on”/“off” buttons, but can provide an output signal depending upon a degree of operation by the user, for example how far the user presses a button, and/or how hard a user is pressing a button. The corresponding hand configuration or gesture can be exaggerated or diminished by the generator 610 in dependence on the degree of activation of the respective button. For example, in the case of a pointing gesture, the pointing finger may be extended to a greater or lesser extent depending on the degree of button activation, or more or fewer fingers may be pointed in dependence on the degree of button activation by the user.
Further buttons may be provided and used. For example, in addition to the two buttons discussed above, examples of the Move controller can provide other buttons. For example, a pair of buttons disposed either side of the thumb button discussed above can be mapped to commands to rotate the respective avatar in the virtual environment, or in other words, to turn the avatar to the left (for example, by operation of a button to the left of the main thumb button) or to the right (for example, by operation of a button to the right of the main thumb button). Other examples include providing avatar locomotion (movement from one place to another in the virtual environment) using buttons other than the two main buttons. A view (in the virtual environment) of the avatar's feet can indicate the direction that the avatar is facing; this can be useful to assist the real user in aligning his or her body with that of the real avatar, in instances in which the real user has turned his or her body during gameplay.
A generator 630 generates a facial configuration. This is mapped to the hand configuration by a mapping table 640. Here, there are two hands (in the present example, though of course the arrangements would work with one hand or more than two hands) each of which may have four states (from the example of the mapping table 620 given above) which leads to 16 (4×4) facial configurations controllable in dependence upon the hand configurations. An example of the mapping table 640 contains at least the information listed below in the discussion of
The generator 630 contributes to or provides the operations (b), (c) and (d) discussed above.
This arrangement therefore provides an example of the image generator being configured to generate respective facial configurations of the avatar for display to other users viewing the virtual environment as a mapping of the configuration of the hands of the avatar, so that the hand configuration of the avatar corresponding to a user provides an indication, to that user, of the facial configuration of the avatar corresponding to that user.
For example, the image generator is configured to generate a facial configuration in response to the current configuration of the avatar hands so that each permutation of avatar hand configurations corresponds to a respective facial configuration.
Once again, if the buttons are not binary on/off controls, the degree of operation by the user can map to variations in the degree of application of a facial configuration. For example, a set of control operations that are mapped to a “smile” can produce a mild smile in the case of a low level of activation of the respective buttons, but a broader smile in the case of a stronger level of activation of the respective buttons. Similarly, if one hand is set to a hand configuration mapped to a particular facial configuration, but the other hand is in the base configuration, the resulting facial configuration can be milder than if both hands were in the same activated hand configuration. Therefore, in examples, at least some of the facial configurations comprise at least two configuration versions; and the image generator is configured to select between versions of a facial configuration according to a degree of activation of the user controls.
In a similar manner, the microphone signal can be used to vary the degree of a facial configuration (such as mild smile—broad smile; mild frown—severe frown and the like). In the absence of an audio signal detected by the microphone, a milder facial configuration can be used. Where at least a threshold audio signal is detected, a more pronounced facial configuration can be used. In other examples, the same facial expression (such as a smile, or an open mouth or the like) can be used, but the mouth size can be varied or modulated in accordance with the microphone signal. These are therefore examples of the head mountable display comprising a microphone; at least some of the facial configurations comprising at least two configuration versions; and the image generator being configured to select between versions of a facial configuration according to audio information detected by the microphone.
Similarly, the orientation of the hands can be mapped to a facial configuration, and a transition from one orientation to another can be carried out gradually (for example, a user gradually rotating their hand to a thumbs-up position whilst holding the controller). This gradual change can be mapped to a gradual development of the corresponding facial configuration, from a mild instance of the gesture to a more pronounced instance of the configuration.
The above embodiment provides a first example of a means of mapping a user input to an avatar's facial configuration. In another embodiment of the present invention, alternatively or in addition to button presses, the position and/or orientation of the or each controller can be used to similarly signify a predetermined set of hand configurations.
In this case, a given hand configuration may be associated with a predefined but otherwise arbitrary position and/or orientation for the or each controller, in a manner analogous to how different letters of the alphabet are associated with predefined but otherwise arbitrary semaphore flag positions.
Alternatively or in addition, the position and/or orientation of the or each controller for some or all of the hand configurations may correspond to a pose that is frequently associated with a given facial configuration.
Hence holding the controllers out to each side of the body may select a hand configuration that in turn maps to a facial configuration indicating confusion. Meanwhile holding the controllers on either side of the user's head may select a hand configuration that in turn maps to a facial configuration indicating fear. Other examples will be apparent to the skilled person.
Alternatively or in addition to the position and/or orientation of the or each controller, motion such as in the form of a dynamic gesture (i.e. a detected change in position, velocity and/or acceleration) may be associated with a hand configuration. Hence for example holding two controllers with arms hanging down may (as a non-limiting example) be interpreted as a neutral or expressionless pose, but a subsequent brief lift and return of the controllers substantially in synchronisation with each other could be inferred to correspond to a shrug of the shoulders. Hence this dynamic gesture could map to a hand configuration that in turn corresponds to a facial configuration indicating indifference or acceptance. Similarly a rapid lateral movement of one controller corresponding to a punching action could map to hand configurations that in turn correspond to a facial configuration indicating anger. Other examples will be apparent to the skilled person.
In some other examples, a facial configuration can be mapped to a series of one or more hand movements. For example, a clap action repeated for a threshold number of occurrences within a particular time window (for example, three times in four seconds) could be mapped to a smile held for (say) ten seconds.
Meanwhile, phoneme based mouth animations can be used as a more advanced option than the simple mouth size variation discussed above.
In example arrangements, the orientation of each avatar hand can be controlled according to the orientation of the respective Move controller held by the user.
Optionally, the image generator is configured to generate an avatar body configuration in dependence upon that avatar's hand configuration. Again this can be by a mapping table and/or in dependence open the hand orientation. For example, both hands pointing downwards can be mapped to a stooped avatar body configuration. Both hands pointing upwards can be mapped to an avatar body configuration looking upwards. In examples, a simple body model can be used, for example a body having a pivot (within the animation) around waist level and with a fixed foot and leg position and orientation (in the absence of any locomotion or movement). The body orientation around this pivot can be controller as a mapping of hand configuration, and/or as a mapping of a detection of the (real) user's head orientation. In examples, the avatar body configuration is dependent upon the user's head (HMD) orientation. So, for example, if it is detected (by one or more orientation detectors associated with or forming part of the HMD) that the respective user has tilted his or her head forward, the avatar body pivots forward by an amount which can be dependent upon the forwards angle of the user's head. Similarly, if it is detected that the user has tilted his or her head to the right, the avatar's body can pivot about the pivot point to the right, and so on. Of course, more complex avatar joint or pivot arrangements can be used.
It will be appreciated therefore that instead of mapping button, position, orientation, and/or motion inputs to hand configuration and then associating avatar body poses with such hand configurations, these inputs may be mapped directly to physical configurations of the avatar, such as hand position, limb position, whether a limb is straight or bent, torso posture, and/or foot position in addition to or as an alternative to mapping to hand configuration.
Hence for given input configurations, the user's avatar may adopt partial or whole body poses such as a ‘heroic’ stance with hands on hips, an ‘inspiration’ or ‘question’ stance with one arm bent and its hand pointing up in the air, a ‘facepalm’ stance with the torso bent and hands placed over the face, and so on. In each case an associated facial configuration is selected as described previously herein.
Hence in such cases, the user can still infer their avatar's facial expression but from an inspection of their avatar's body pose, rather than from an inspection of hand configuration; although as noted above this may not be needed if for example an expression indicator or virtual mirror is otherwise displayed or accessible to the user.
It will further be appreciated that at least some of these whole or partial avatar configurations need not be static but could be animated, for example causing the avatar to jump up and down in an ‘excitement’ action, or stretch arms out wide in a ‘tired’ or yawning action.
Given the selected hand or whole or partial avatar body configurations and associated facial configurations, the generators 610, 630 act as an image generator to provide a rendered view of the virtual environment for local display to the user of the corresponding HMD, thereby providing the functions (e) and (f) discussed above.
detecting (at a step 700) operation of one or more user controls;
optionally, at a step 710, detecting an audio signal from a microphone associated with the HMD;
generating (at a step 720) a hand configuration and/or a whole or partial avatar pose configuration in response to operation of the user controls by the user, to configure the avatar representing that user;
generating (at a step 730) a respective facial configuration of the avatar for display to other users viewing the virtual environment that corresponds to the configuration of the hands and/or whole or partial avatar pose configuration, so that the configuration of the avatar corresponding to a user provides an indication, to that user, of the facial configuration of the avatar corresponding to that user. Optionally, the facial configurations may be selected from variants of a facial configuration in dependence upon the detected microphone input or controller inputs such as degree of trigger press, or hand rotation (for example, a degree of smile may respond to trigger pressure or hand angle); and
optionally, at a step 740, generating or modifying an avatar body configuration in dependence upon the user's head (HMD) orientation, for example to reflect where the user wants to direct their attention (e.g. when looking at a particular other avatar). The methods and techniques described herein recite the use of at least one controller for button input and/or orientation, position and/or movement input. As was described previously herein, the orientation, position and/or movement input can be determined from telemetry transmitted by the controller.
However as noted previously such a controller may have a visually distinctive component such as a coloured light that may be tracked visually by a camera 302. Hence alternatively or in addition to button presses and/or telemetry, a position and/or movement input of a controller may be obtained by analysis of video images from the camera 302 capturing a scene encompassing the or each controller held by the user. This analysis will typically be performed by the games console, but may optionally be performed at least in part by a processor of the breakout box or the HMD. As explained previously, depending on the capabilities of the camera (and/or available computational resources), a position and/or movement input based on the image analysis may be limited to a 2D plane or may include depth. It will be apparent that the example positions and poses described previously herein with respect to the position or movement of the controllers based on telemetry may be similarly detected by image analysis.
Such analysis may be based solely on the position of the visually distinctive component of each controller, or may be based on any number of appropriate visual cues such as the pose of the user holding the controllers and/or the position of the controllers relative to the user's body (as described previously, holding the controllers with arms outstretched away from the user's body could be indicative of confusion, whilst holding the controllers next to the user's face could be indicative of fear or surprise).
Hence optionally the positioning and/or posing of some or all of the user's body may similarly be detected. For example, a skeletal model may be generated based on detection of the user's head (for example, detection of lights on the HMD) and detection of the controllers in the user's hands.
It will be appreciated that by extension, such a skeletal model or other model of the user's body (torso and/or limbs etc) may not rely on visual recognition of the controllers in the user's hands at all, and instead directly detect the pose of the user's own body or part thereof. Consequently position, orientation and/or movement input may be provided by the user to the camera 302 without holding the or each controller, or whilst holding one controller in one hand but adopting poses made distinct by the position of the other hand and/or other body parts.
Regardless of how inputs are provided to the game console (i.e. whether via button/joystick/mouse/keyboard inputs, telemetry based inputs and/or video-based inputs), then as described previously herein, predetermined inputs correspond to predetermined configurations of the avatar's hand and/or a pose of some or all of the avatar's body (torso and/or limbs etc). Then in turn, a facial configuration (e.g. an emotional expression) is mapped to this predetermined configuration of hand and/or body.
As noted previously, one benefit of this is to enable the user, when in a first person view, to determine what expression their avatar is showing by an intuitive visual self-inspection of their avatar's hands and/or body whilst notionally immersed in the virtual world.
However as was also noted previously, other mechanisms to confirm their facial expressions are possible, and so this correspondence is not essential.
Consequently the chain of mappings between inputs, hand or body configuration, and facial configuration may vary, for example as follows.
Firstly, where the user input is statically pose based or dynamically gesture-based, the user's avatar will typically be arranged to at least partially mimic the user's actions in real time and so will mimic the gesture as it is made. However, once a specific predefined pose or gesture has been recognised by the games console or HMD as an instruction to change a configuration, it is not necessary for the avatar to hold the pose or gesture for as long as the avatar displays the corresponding facial configuration. Hence the input pose or gesture does not need to correspond to a subsequently ongoing hand or body configuration of the avatar, although of course this is still possible as described previously herein. Instead, the pose or gesture can map directly to a corresponding facial configuration.
Hence in an embodiment of the present invention, inputs based on orientation, position and/or motion that are used to specify a facial configuration can be recognised as part of an ongoing tracking of user inputs, and result in a corresponding change in facial configuration state; but the avatar's hands and/or body may at least partially continue to mimic some or all of the users inputs rather than similarly adopt a predetermined configuration.
Clearly where the facial expression itself comprises animations or associated poses (for example smiling and nodding, or laughing and giving a fist-pump), this may temporarily override or adapt the replication of user movement by the avatar for the duration of the animation. In such a case, however, the facial expression itself may persist for longer than such an animation, which may serve merely to punctuate a change in state of facial configuration.
Similarly when user inputs are button based, then where the facility to inspect one's own facial expression is provided through other means as described previously herein, then optionally the button input sequence does not need to correspond to a hand or body configuration, and instead there can be a direct mapping between predetermined input button selection(s) and corresponding facial expressions and/or animations.
The methods and techniques described herein have discussed a one-to-one correspondence between inputs and hand/body configurations and/or facial configurations. However it will be appreciated that a many-to-one mapping between inputs and facial configurations is possible. Hence for example several different physical gestures may map to an associated single facial expression. For example clapping, jumping up and down, and doing a twirl may all cause the avatar expression to be happy, whilst a thumbs down action, or putting one or more hands near the face may cause the expression of a user's avatar to be sad.
Conversely, it will be appreciated that a one-to-many mapping between inputs and facial configurations is possible. As was described previously herein, this may be due to certain inputs having a variable component such as trigger button pressure or hand orientation changing the extent of a smile. However, for a given button input selection or position, orientation and/or motion input, optionally one of several expressions may be selected in response to other factors.
For example, certain expressions may be age restricted and hence not available to some users. Similarly, some expressions may be gender specific, or the same expression may be triggered by different input poses or gestures depending on the gender of the user.
Meanwhile, the same input may result in the selection of a different facial configuration depending on the context of the virtual environment; for example if the avatar is holding a bottle of wine, then putting a hand forward is likely to be a friendly gesture to be accompanied by a smile, whereas if the avatar is holding an axe then putting a hand forward is likely to be an unfriendly gesture and could be accompanied by a frown. Consequently virtual objects and/or environments may comprise emotion preference metadata indicating what inputs should correspond to what facial configurations, so that the same input may result in a different facial expression depending on the overall environment and/or individual interactions with particular avatars or objects. Similarly, the time at which an input is provided may result in the selection of a different expression; for example holding one's arms out from one's side may result in a quizzical expression when performed during the day, but may result in a yawn or tiredness expression when performed after a threshold time in the evening.
Furthermore, it may be possible to change the avatar's facial configuration without any specific configuration input from the avatar's user. In particular, a time-out function may be provided that resets and avatar's face to a neutral expression after a predetermined period of time and/or after a predetermined event other than the inputs described previously herein. For example, an expression selected whilst another user is talking may be reset when or soon after that other user stops talking. Similarly, a facial expression selected by a user when with a particular other user or group may be reset when the user moves away from that other user or group by a threshold distance, or when a new person joins a group. The particular rules for these transitions may be chosen by a designer of a virtual environment, or via user settings.
Similarly, some environmental events may automatically trigger an associated facial expression in nearby avatars; for example a firework going off may trigger an expression of surprise or delight on the face of any avatar within a predetermined distance and/or on the face of any avatar looking approximately in the direction of the firework. Similarly if another user holds up a virtual camera, avatars within that camera's virtual field of view may all smile, or adopt an expression pre-selected by their respective users for such an occasion. Hence again such environmental events or objects may comprise emotion preference metadata indicating what facial expressions should be selected in response to interaction with the object, whether direct (interaction by the user) or indirect (e.g. in proximity to interaction by another user).
The methods and techniques discussed herein may be performed by programmable hardware executing computer software which, when executed by such a computer, causes the computer to execute the method defined above. The software may be provided by a non-transitory, machine-readable storage medium which stores the computer software.
In some examples, as discussed previously, the “local” image generator at or associated with an HMD is configured to generate images representing other avatars corresponding to other users interacting with the virtual environment. These avatars can be based upon information sent from the other users' apparatus indicating at least facial configurations. In this way the local image generator is configured to generate an image representing another avatar having a facial configuration dependent upon user control operations by the corresponding user.
When it comes to the eye direction of such a locally generated avatar, in some examples, an eye direction (gaze) detector at the other HMD can detect the direction and/or blinking of the (real) eyes of the HMD wearer (or the avatar's eyes) and send information indicating this direction to allow a corresponding direction and/or blink to be rendered for the avatar of that user. In other examples, such as examples using HMDs not having a gaze detection facility, the eye direction can be selected at the local apparatus rendering an image for display at a local HMD. In an example, the image generator is configured to generate an image representing another avatar having an eye direction dependent upon a position, relative to that avatar, of another avatar in the virtual environment engaging in an activity. Examples of such an activity can include one or more of: moving, talking, interacting with an NPC, fighting and so on. Note however that eye direction could be affected by other factors—for example, an avatar could simulate looking at a dramatic event happening nearby, such as a rock crashing down. So more generally, the avatar's eyes can be responsive to such general eye-look trigger items. Again as noted above, such objects (including other avatars) may comprise emotion preference metadata indicating which of the available facial configurations or actions of the avatar to use in response to the action of the object.
Illustrative examples of facial expressions triggered by hand configurations will now be discussed with reference to
Accordingly, in examples, the hand configuration of the avatar corresponding to a user provides an indication, to that user, of the facial configuration of the avatar corresponding to that user.
In some examples, the selected facial configuration can be overridden, or in other examples only the “neutral” facial configuration is overridden, by a facial configuration automatically derived from aspects of the settings or operational characteristics of the corresponding user's HMD. For example, if the user's microphone is muted (by operation of a user control) the neutral facial configuration can employ a schematic “X” shape instead of a neutral mouth. In another example, if the user has removed his or her HMD from the head (a feature which can be detected by current HMDs at the priority date) the avatar can be placed into a “sleeping” body and facial configuration.
It will be appreciated that in cases where the input comprises button selections, the user operating the controls can make abrupt changes from one button configuration to another. In some examples these can be represented by similarly abrupt transitions from one facial configuration to another, so that (subject to any minor processing delay) the facial configuration follows the prevailing button configuration. In other examples, an animated transition from one facial configuration can be used so as to provide a more gradual transition. This can be performed by animation techniques. In a simple example, a look-up table of transitional facial configurations can be provided so as to give a set of intermediate facial configurations, for example between each facial configuration in the table above and a neutral facial configuration. For example, if the buttons corresponding to “super happy” are pressed, the facial configuration can follow a sequence of little smile→slightly happy (from the table above)→larger smile→super happy (from the table above), for example over the course of 0.8 seconds in total. If the button configuration is then changed by the user to “super angry”, the facial configuration could return through that same sequence over (say) 0.8 seconds to neutral, and then progress through a similar sequence of “growing anger” to “super angry” over another 0.8 seconds. Clearly, the rate of change is a simple design feature and other rates of change could be used. It will be appreciated that such techniques may also be used to transition from any facial expression to any other facial expression, for example when changing expression state after detection of a particular physical pose or gesture.
The above examples and techniques have been illustrated with reference to a user's avatar, which is typically an object rendered within a virtual environment shared by a plurality of users. As such, applying these examples and techniques only to such avatars may potentially limit their use to such shared virtual environments. However, it is desirable for a user to share their experiences and emotions with a wider audience than those in the immediate vicinity of that user within a virtual space at a particular time.
For example, it would be preferable for a user to be able to express their emotions to a wider audience of friends, either within a network administered by the maker of the apparatus and/or more widely on social media, and/or to record their emotions for posterity in a video capture of a game experience, or in a vignette. Furthermore, when a user wishes to record or share their own view of the virtual environment, often their own avatar is not visible in the recorded view because the viewpoint is from a first-person perspective, or from a virtual camera over their shoulder. Similarly a user's avatar is not always facing towards another user, making it difficult for that other user to gauge an emotion expressed by the avatar's face. Hence again for these reasons it is desirable to be able to alternatively or in addition convey the user's emotions using the techniques described herein without direct reliance on being able to see the user's avatar's face.
Accordingly, in an embodiment of the present invention, a user can generate (select or create) a so-called emote sticker (a static or animated illustrated abstraction of an emotion) for circulation to one or more potential audiences using the techniques corresponding to those described previously herein.
The generation of an emote sticker can occur in parallel with the selection of an avatar's facial expression as described previously. Alternatively or in addition, an emote sticker can be generated separately to a user's avatar's expression (for example during the composition of a message, or when creating a separate social media post or status notice), thus potentially allowing a user to be ‘two faced’, expressing different emotions to different audiences, or simply generating an emote sticker summarising an experience independent of any other in-game events or configurations (such as a currently expressed avatar emotion).
In an embodiment of the present invention, a user performs an input action to indicate their desire to generate an emote sticker (for example pressing a controller button, adopting a pose, issuing voice command, or interacting with a trigger object in a virtual environment).
The user may then perform one or more physical gestures.
A lexicon of predetermined physical gestures may be used to select predefined emote stickers (either static or animated). Hence for example a user holding their hands in a ‘c’ configuration on either side of their mouth may result in a static or animated sticker of a character eating a hamburger, as per
The character may be pre-drawn (e.g. one of several pre-drawn characters, for example reflecting different genders/ages), or may optionally be generated to reflect the user's in-game avatar, if they have one. In this latter case, the pre-drawn character may be adjusted to use similar hair/skin/clothing colour to that of the user's avatar, or may have individual features adapted in a similar manner to (and potentially using the same parameters as) any prior customisation of the user's avatar itself.
Alternatively, the character may be derived from renderings of the user's avatar; for example, colours of skin, hair etc of the avatar can be averaged and optionally have their saturation and/or contrast increased, for use in a so-called cell-shaded 2D rendition of the avatar or a similar version that looks as though it is drawn flat. In addition, where the avatar has been customised, deviations from a neutral (pre-customisation) avatar model may be magnified to create a caricature of the avatar, giving the resulting emote sticker representation of the avatar a cartoon-like feel.
It will be appreciated that in principle this approach may optionally also be applied to pictures of the user themselves, for example as captured by video camera 302 during an optional registration process for using the emote stickers, to create a 2D version (or caricature based on deviations from mean facial features) of the user.
As noted above, a lexicon of predetermined physical gestures may be used to select predefined emote stickers (either static or animated). As noted above,
The predetermined physical gestures may have an approximately direct correlation with the key gesture within the sticker or animation; hence a user may adopt a pose similar to that shown in each sticker of
Alternatively or in addition, such poses may be mapped only to configurations of the user's hands, again as discussed previously. This may for example be advantageous when a user is in a virtual reality environment and may not wish to gesticulate whist unable to see what their hands might hit.
Similarly, such poses may be mapped to buttons or button sequences on one or more controllers, again as discussed previously.
As was noted previously herein, such stickers may be used within a shared virtual environment, for example as decals to tag a user's preferred area, or positioned above the user's avatar (when viewed by another user) to clearly indicate an emotion (for example when the avatar is not facing the other user, or the avatar is not sufficiently expressive to allow the clear conveyance of emotion; for example if the avatar is a monster, or robot).
Such a virtual environment may support a data convention for communicating which of the predefined emote stickers has been chosen, together with any customisation data for the character used in the sticker. For example, a number indicating the predefined emote sticker may be used, allowing look-up in a table to identify the relevant sticker, together with a flag indicating if other configuration data is being provided. A recipient apparatus can then look up the relevant sticker and render it from its own sticker asset data, optionally using any customisations of the character in the sticker as received.
In a similar manner, an app or plug-in for a different platform (such as for a mobile phone, and/or a social media platform) may also be provided with a look-up table of sticker numbers and the relevant sticker asset data needed to generate the stickers, optionally using any customisations of the character in the sticker as received. In this way stickers may be used to reach audiences outside the originating application (whether for example this is a VR application, a conventional game, or a messaging application provided with the apparatus OS).
Optionally stickers may similarly be provided in a manner that is not reliant on a co-operative app, plug-in or game to render the sticker itself. In this case static stickers may (also) be rendered as JPGs, GIFs or in any other common image format, to be used by any conventional application capable of displaying such images—such as for example web browsers and social networking apps.
Similarly animated stickers may (also) be rendered as animated GIFs, common video formats, WebGL widgets and the like, again to be used any conventional application capable of displaying such images—such as for example web browsers and social networking apps.
In this way, emote stickers can be exported/posted to other audiences. Hence such a sticker may be used to tag the user when they are shown in a ‘friends list’ of other friends in a network administered by the manufacturer of the apparatus or developer/publisher of the game or app (for example the PlayStation Network and its associated friends list capability), and/or may be posted to a social network such as Facebook® or Twitter®, optionally with accompanying text either pre-associated with the sticker or input by the user.
Similarly such a static sticker may be superposed on a screen-capture from the apparatus (for example of a frame from a videogame)—the sticker may be applied after the user explicitly generates one, or the last generated sticker may be automatically used or offered for use to the user. This allows the user to ‘brand’ or ‘tag’ their images with emotionally relevant character images. Likewise an animated sticker may be added to a video capture from the apparatus (again for example from a videogame). Hence respective frames of the sticker are added to respective frames of the video capture. Again this may occur after a user has specifically generated a sticker, or the last generated sticker may be used or offered for use to the user.
As will be described later herein, an animated sticker may typically have duration of between 1 and 30 seconds, and more typically around 10 seconds. Hence the sticker animation may loop when superposed on a longer video capture. Optionally the animation may pause on a key frame of the sticker, for example for 10-30 seconds, before looping again. Again optionally the apparatus may analyse the captured video for levels of motion, and only instigate or repeat an animation when the overall level of motion is below a predetermined threshold.
In any event, it is still necessary to generate the sticker, whether this is done ‘live’ by an enabled app, or for export as an image or animation as described previously.
In an embodiment of the present invention, graphical elements of the character in the sticker are positioned for rendering according to an underlying 2D skeleton structure. It will be appreciated that ‘skeleton’ in this case simply refers to structures having one or more degrees of freedom; hence whilst the skeleton may include arms, it may also include eyebrows—which are not conventionally understood to be part of a skeleton. Hence the underlying skeleton structure may equally be thought of as an underlying character puppet structure with a plurality of features each having one or more degrees of freedom, defined parametrically.
Hence a sticker may be expressed efficiently by defining the parameters of the underlying digital puppet; the apparatus then renders the sticker by building up the graphical elements of the character in order (e.g. from back to front), as applicable. Hence to render a face, a face view corresponding to the pose of the puppet is chosen (for example face-on, or looking left or looking right; optionally only one angle of view is supported); the face is then optionally selected, morphed, positioned, scaled or rotated as appropriate according to supported parameters, such as jaw angle. Alternatively or in addition, plural face components may be positioned according to supported parameters, such as upper and lower face portions that overlap and are positioned at a relative angle corresponding to the jaw angle. Similarly facial features such as the mouth, nose, eyes and eyebrows may be selected, assembled, positioned, scaled, rotated and/or morphed as applicable in a similar manner, depending on the supported parameters and their respective values.
In general, such a parametric description of the sticker will be more compact than a graphical rendering of the sticker itself. This can be beneficial for example in an on-line social environment where multiple animated stickers may be being exchanged frequently, as it reduces the network load and also speeds up the transmission of the emote sticker, which can be useful for rapid and fluid communication and reaction to events and conversations. It also means that the same ‘sticker’ (in terms of pose or animation) can easily use different character graphics once created.
The use of a digital puppet to define predetermined static and animated emote stickers also optionally enables user-created emote stickers.
A video image of the user may be captured by the video camera 302, and a user skeletal model may then be mapped to their pose in each frame (or just one for a static sticker). The values from this skeletal model may then optionally be smoothed by a filter, and/or optionally normalised, before being used to set a corresponding pose or sequence of poses in the emote sticker digital puppet. This may be done for example by generating a 2D projection of the user-skeletal model from the same direction as the notional direction of view for the sticker (for example face-on, to the left or to the right, as per the example above). Those parts of the projection that correspond to parts of the puppet (for example any or all of hand, arm, head, torso and leg positions) may then be used to set those parts of the puppet to a corresponding position.
In this way, overall posing and as applicable overall motion of the user can be translated to the emote sticker's underlying puppet data. Optionally this may be displayed back to the user in real-time to provide feedback.
Potentially, the user's facial expression may also be captured and used to set corresponding parameters of the puppet.
However, typically this is difficult to do reliably in many circumstances (for example due to variations in lighting, camera resolution/distance and the like) and may also require a laborious training or calibration process before it can be done. Furthermore, it may be impossible to capture expressions from the user's face in the case that they are wearing a VR head mounted display.
Hence in an embodiment of the present invention, the techniques described previously herein are used as appropriate to select an emotion for the sticker character.
Thus for example a user may use a predetermined physical gesture in the form of a hand configuration, orientation and/or position to select an emotion, as described previously herein in relation to an avatar. It will be appreciated that using a predetermined physical gesture such as a hand configuration can enable a user to separately control the captured overall motion of the sticker and the displayed emotion. Thus a user could equally record throwing their hands up in the air in joy, or in despair, depending for example on whether their hands are open or closed when doing so.
As is also described herein, a user may use not just their hands but other parts of their body to signify an emotion. Hence when creating an animation, optionally part of the user's pose or performance may be recognised as corresponding to a predetermined physical gesture indicative of a respective emotion; this predetermined physical gesture may use some or all of the user's body, as described previously herein.
Notably, there may be more predetermined physical gestures corresponding to emotional expressions than there are predetermined physical gestures corresponding to predefined emote stickers.
Hence whilst some performed actions can be used to trigger predefined animations with associated emotions, other performed actions can be recorded by the user, with an associated emotion being recognised automatically because the performed action corresponds to one of a lexicon of actions or poses associated with emotions.
Alternatively or in addition, recording of the physical pose/actions for the sticker could be recorded separately to specifying the facial expression (for example by recording a user's pose or performance first, and then receiving input indicating a desired expression second using any of the techniques described previously herein) so that the user is free to perform any physical action without concern for how the system may interpret it, and then specify an emotion using a specific predetermined physical gesture.
Animated stickers may be associated with a corresponding audio recording, for example made using a microphone in the video camera 302, or built into a handheld controller. Typically the audio recording will last as long as the animation cycle, but potentially and animation cycle and an audio loop could repeat according to respective periods. Potentially static stickers could also be associated with an audio recording, which could be triggered by interacting with the sticker within a suitably enabled app or other program.
As noted above, a user may interact with one or more handheld controller. This could make it difficult for a user to perform predetermined physical gestures that correspond to hand configurations of the kind illustrated herein. Accordingly in an embodiment of the present invention, interactions with buttons or joysticks of the or each controller, and/or sequential interactions with one or more buttons and/or joysticks, can be mapped to such predetermined physical gestures so that equivalent sticker/emotion selection can be performed using hand configurations that also serve to interact with the controller.
In a summary embodiment of the present invention, a virtual reality apparatus comprises an image generator 300 (such as a PlayStation® games console) to generate images representing a user of a virtual environment, the image generator being responsive to selection of a predetermined input by the user; a video camera (302) for use in image-based tracking of a user by the image generator; in which the predetermined input selected by the user is the performing of a predetermined physical gesture in view of the video camera; and the image generator is arranged to generate a static or animated emote sticker responsive to the predetermined physical gesture of the user for display to other users (for example other users viewing the virtual environment, or friends on social media).
In an instance of the summary embodiment, the virtual environment includes an avatar representation of the user positioned within the virtual environment so that the user's viewpoint of the virtual environment substantially corresponds to the viewpoint of the avatar corresponding to that user; and the image generator is arranged to generate respective facial configurations of the avatar responsive to the predetermined physical gesture of the user for display to other users viewing the virtual environment.
In an instance of the summary embodiment, the predetermined input further comprises one or more selected from the list consisting of pressing one or more predetermined buttons; adopting a predetermined pose while holding a control device 330; and performing a predetermined physical gesture while holding a control device 330; Again it will be appreciated that any suitable combination of such inputs may be considered.
Suitable control devices include but are not limited to one or more selected from the list consisting of a control device 330 comprising one or more buttons (such as a DualShock 4® controller, or a Playstation Move® controller); a control device 330 operable to transmit telemetry data indicative of position, orientation and/or motion (again such as a DualShock 4® controller, or a Playstation Move® controller); and a control device 330 comprising a predetermined visible feature for use in image-based tracking (again such as a DualShock 4® controller, or a Playstation Move® controller), as described previously herein.
It will be appreciated that any suitable combination of such user controls may be considered; as noted previously, a user may hold one PlayStation Move® controller that provides telemetry back to the console, whilst both the Move controller and the user are tracked by video camera 302, with the combined data of the telemetry and image analysis determining a pose and hence a corresponding emote sticker and optionally avatar facial configuration, which in turn may be modified by an extent of pressure on a trigger button of the controller.
In an instance of the summary embodiment, the predetermined input selected by the user corresponds to a configuration of the avatar's hands, and the image generator is arranged to generate respective hand configurations of the avatar for display to other users viewing the virtual environment. In this case therefore, both the hands and face of the avatar are adapted in response to the input, in conjunction with any other changes.
In an instance of the summary embodiment, each predetermined physical gesture corresponds to a single predefined static or animated emote sticker. Alternatively or in addition, plural predetermined physical gestures correspond to a single predefined static or animated emote sticker, as described previously herein.
In an instance of the summary embodiment, a predetermined input is associated with the selection of a predetermined static or animated emote sticker, as described previously. However, optionally the selection of an emotional expression in the sticker is in dependence upon one or more selected from the list consisting of metadata associated with the virtual environment; metadata associated with a virtual object in the virtual environment with which the user's avatar is interacting; metadata associated with another avatar with which the user's avatar is interacting; and the time of day, as described previously herein. Hence the same gesture (e.g. placing one's hands in the air) may result in selection of an animated sticker or a character similarly putting their hands in the air—but if the current virtual environment is a bedroom, or is set at night time, then the character may yawn, whilst if the current virtual environment is a nightclub, then the character may smile or appear to sing.
In an instance of the summary embodiment, a tracking processor is adapted to track a skeletal model of the user as they perform a predetermined physical gesture, and at least partially transpose the skeletal model of the user onto a skeletal model of the emote sticker, as described previously. In this way, the user's performance provides digital puppetry for the sticker, which replicates the user's motions to the extent chosen (subject to any smoothing and/or mapping to different body morphologies, etc).
In such a case, a predetermined emotion may be selected for the emote sticker in response to at least part of the performed predetermined physical gesture. Hence the user's performance may include one or more recognised gestures corresponding to emotional expressions, allowing the system to set the facial expression of the character in the sticker accordingly, as described previously herein.
In an instance of the summary embodiment, an object in the virtual environment is associated with emotion preference metadata indicative of a preferred emotion, and the image generator is responsive to interaction by the user with the object to configure the face of the emote sticker in accordance with the emotion preference metadata, as described previously herein.
In an instance of the summary embodiment, the image generator is configured to generate images representing other avatars corresponding to other users interacting with the virtual environment. Similarly, the image generator is configured to generate an image representing another avatar having a facial configuration dependent upon user control operations by the corresponding user. In these cases, the image generator may be configured to generate an image representing another avatar having an eye direction dependent upon a position, relative to that avatar, of an eye-look trigger item.
In an instance of the summary embodiment, the apparatus comprises a head mountable display 20.
Meanwhile in another summary embodiment of the present invention, a method of generating images representing a virtual environment, for display to a user by a head mountable display to be worn by that user, the virtual environment including an avatar representation of the user positioned within the virtual environment so that the user's viewpoint of the virtual environment substantially corresponds to the viewpoint of the avatar corresponding to that user, comprises the steps of detecting operation of one or more user controls with which predetermined inputs may be selected by the user, the generating of images being responsive to selection of a predetermined input by the user to correspondingly configure a face of the avatar representing that user; and generating respective facial configurations of the avatar for display to other users viewing the virtual environment.
Instances of this summary embodiment corresponding to the operation of the apparatus in the preceding summary embodiment are also clearly understood to be within the scope of the appended claims.
As noted previously herein, such methods and techniques may be implemented by conventional hardware suitably adapted as applicable by software instruction or by the inclusion or substitution of dedicated hardware.
Thus the required adaptation to existing parts of a conventional equivalent device may be implemented in the form of a computer program product comprising processor implementable instructions stored on a non-transitory machine-readable medium such as a floppy disk, optical disk, hard disk, PROM, RAM, flash memory or any combination of these or other storage media, or realised in hardware as an ASIC (application specific integrated circuit) or an FPGA (field programmable gate array) or other configurable circuit suitable to use in adapting the conventional equivalent device. Separately, such a computer program may be transmitted via data signals on a network such as an Ethernet, a wireless network, the Internet, or any combination of these or other networks.
It will be apparent that numerous modifications and variations of the present disclosure are possible in light of the above teachings. It is therefore to be understood that within the scope of the appended claims, the disclosure may be practised otherwise than as specifically described herein.
Number | Date | Country | Kind |
---|---|---|---|
1604203.8 | Mar 2016 | GB | national |
1619809.5 | Nov 2016 | GB | national |
1702700.4 | Feb 2017 | GB | national |
Filing Document | Filing Date | Country | Kind |
---|---|---|---|
PCT/GB2017/050649 | 3/10/2017 | WO | 00 |