The present disclosure relates to an information processing device, an information processing method, and a program.
In recent years, with development of information processing technologies, various technologies related to an input for manipulating devices have been researched and developed. Specifically, there are technologies related to inputs in 2-dimensional spaces using touch screens or the like and there are technologies related to inputs in 3-dimensional spaces based on recognized gestures or the like of users.
Here, there are cases in which manipulations intended by users are not recognized by devices. For example, when users perform manipulations without being conscious of regions in which manipulations are recognized (hereinafter also referred to as manipulation regions), some or all of the manipulations are performed outside of the manipulation regions in some cases. In these cases, the manipulations performed outside of the manipulation regions are not recognized by devices. As a result, processes related to manipulations intended by users may not be performed and processes related to different manipulations from the manipulations intended by the users may be performed.
In contrast, Patent Literature 1 discloses an image processing device that displays a black image indicating a region other than a manipulation region and a captured image in which a user performing a manipulation is shown and superimposes an image for notifying the user that the user approaches a region other than the manipulation region on the black image.
However, in the technology of the related art typified in Patent Literature 1, there is concern of manipulability deteriorating in a case in which a manipulation region and a display region in which a virtual object related to a manipulation is viewed are different. For example, in a device such as the above-described touch screen, a display region and a manipulation region are matched. Therefore, a user can confirm a user manipulation while viewing the display region. However, in a device in which a gesture manipulation or the like is recognized, a display region and a manipulation region are different. Therefore, a user may not confirm a user manipulation while viewing the display region. As a result, there is concern of the user performing a manipulation erroneously.
Accordingly, the present disclosure proposes a structure capable of suppressing deterioration in manipulability in a case in which a display region and a manipulation region are different.
According to the present disclosure, there is provided an information processing device including: an input unit configured to obtain distance information changed in accordance with a distance between a manipulator recognized in a different manipulation region from a display region in which a virtual object is viewed and an input recognition region in which an input by the manipulator in the manipulation region is recognized; and an output control unit configured to control an output related to the virtual object on a basis of the distance information.
In addition, according to the present disclosure, there is provided an information processing method including, by a processor: obtaining distance information changed in accordance with a distance between a manipulator recognized in a different manipulation region from a display region in which a virtual object is viewed and an input recognition region in which an input by the manipulator in the manipulation region is recognized; and controlling an output related to the virtual object on a basis of the distance information.
In addition, according to the present disclosure, there is provided a program causing a computer to realize: an input function of obtaining distance information changed in accordance with a distance between a manipulator recognized in a different manipulation region from a display region in which a virtual object is viewed and an input recognition region in which an input by the manipulator in the manipulation region is recognized; and an output control function of controlling an output related to the virtual object on a basis of the distance information.
According to the present disclosure, as described above, it is possible to provide a structure capable of suppressing deterioration in manipulability in a case in which a display region and a manipulation region are different. Note that the effects described above are not necessarily limitative. With or in the place of the above effects, there may be achieved any one of the effects described in this specification or other effects that may be grasped from this specification.
Hereinafter, (a) preferred embodiment(s) of the present disclosure will be described in detail with reference to the appended drawings. Note that, in this specification and the appended drawings, structural elements that have substantially the same function and structure are denoted with the same reference numerals, and repeated explanation of these structural elements is omitted.
Note that the description will be made in the following order.
1. Embodiment of present disclosure
1.1. System configuration
1.2. Configuration of device
1.3. Process of device
1.4. Summary of embodiment of present disclosure
1.5. Modification examples
2. Hardware configuration of information processing device according to embodiment of present disclosure
An information processing system according to an embodiment of the present disclosure and an information processing device realizing the information processing system will be described.
First, a configuration of the information processing system according to the embodiment of the present disclosure will be described with reference to
As illustrated in
The information processing device 100 controls projection of the projection device 300 using a measurement result of the measurement device 200. Specifically, the information processing device 100 recognizes a part of the body of a user serving as a manipulator from the measurement result supplied from the measurement device 200. Then, the information processing device 100 controls an aspect of the projection by the projection device 300 on the basis of a positional relation between the recognized part of the body and a region in which an input by the manipulator is recognized in a manipulation region (hereinafter also referred to as an input recognition region). For example, the information processing device 100 controls a projection position or the like of a virtual object 20 which the projection device 300 is caused to project on the basis of a position of a hand of the user measured by the measurement device 200 and a position of an input recognition region 50A. Note that the manipulation region corresponds to a measurement range of the measurement device 200. The details of the information processing device 100 will be described below.
The measurement device 200 measures a surrounding situation of the measurement device 200. Specifically, the measurement device 200 measures a phenomenon in which a position or a state of an object near the measurement device 200, for example, a user, a manipulator, or the like, is grasped. Then, the measurement device 200 supplies information obtained through the measurement (hereinafter also referred to as measurement information) to the information processing device 100 as a measurement result. For example, the measurement device 200 is a depth sensor and can measure a positional relation between a part of the body (for example, a hand) on which a marker is mounted and a nearby object (that is, positions of the part of the body and the nearby object in a 3-dimensional space) by mounting the marker on the part of the body of the user. The measurement information may be 3-dimensional image information. Note that the measurement device 200 may be an inertial sensor mounted on the user or a ring type or wristband type wearable device that includes the inertial sensor. In addition, the measurement device 200 may be installed in any way as long as the manipulation region can be ensured. For example, the measurement device 200 may be provided on a table of which an upper surface illustrated in
The projection device 300 projects an image on the basis of instruction information of the information processing device 100. Specifically, the projection device 300 projects the image supplied from the information processing device 100 to an instructed place. For example, the projection device 300 projects the virtual object 20 to the projection region 10 illustrated in
As illustrated in
On the other hand, in the related art, there is a technology in which a manipulation region and a display region are coincident with each other. For example, in a device such as a smartphone or a tablet terminal that includes a touch screen, a display screen corresponding to a display region and an input unit corresponding to a manipulation region are integrated. In addition, in a device that receives a manipulation input when a user touches a graphical user interface (GUI) projected to a wall surface or the like, the GUI corresponding to a display region and a region corresponding to a manipulation region are coincident with each other. Therefore, when a manipulator enters a field of view of a user viewing a display screen, the user can confirm a manipulation performed by the user while viewing the display screen. Accordingly, the user can intuitively perform a manipulation.
In the related art, however, in a case in which a manipulation region and a display region are each located at different places, there is concern of manipulability deteriorating. For example, as illustrated in
Accordingly, the present disclosure proposes the information processing system capable of suppressing deterioration in manipulability in a case in which a display region and a manipulation region are different and the information processing device 100 that realizes the information processing system.
Next, a configuration of the information processing device 100 according to an embodiment of the present disclosure will be described with reference to
As illustrated in
The communication unit 102 communicates with an external device of the information processing device 100. Specifically, the communication unit 102 receives a measurement result from the measurement device 200 and transmits projection instruction information to the projection device 300. For example, the communication unit 102 communicates with the measurement device 200 and the projection device 300 in conformity with a wired communication scheme. Note that the communication unit 102 may communicate in conformity with a wireless communication scheme.
The recognition unit 104 performs a recognition process on the basis of the measurement result of the measurement device 200. Specifically, the recognition unit 104 recognizes a manipulator on the basis of the measurement information received from the measurement device 200. As a manipulator, there is a part of the body of the user. For example, the recognition unit 104 recognizes a finger, a hand, or the like of the user on the basis of 3-dimensional image information obtained from the measurement device 200. Note that a manipulator recognized by the recognition unit 104 may be a manipulation device such as a pen type or a stick type manipulation device.
Further, the recognition unit 104 recognizes a position of the manipulator. Specifically, the recognition unit 104 recognizes the position of the manipulator in a manipulation region in which the manipulator is recognized. For example, the recognition unit 104 recognizes the position of the manipulator in a 3-dimensional space of the manipulation region on the basis of a distance between the measurement device 200 and the manipulator measured by the measurement device 200. Note that the recognition unit 104 may recognize the position of the manipulator on the basis of a size of the recognized manipulator.
In addition, the recognition unit 104 recognizes a manipulation region. Specifically, the recognition unit 104 recognizes the manipulation region and an input recognition region included in the manipulation region on the basis of the measurement result of the measurement device 200. For example, the manipulation region and the input recognition region correspond to an object, and thus the recognition unit 104 recognizes the manipulation region and the input recognition region by recognizing the object. In the example of
The input recognition region corresponds to the object, as described above. The input recognition region is not limited to a flat surface such as the upper surface of the table 2 illustrated in
In addition, the recognition unit 104 recognizes a manipulation target of a manipulator. Specifically, the recognition unit 104 recognizes a manipulation target which is an actual object in a display region. For example, the recognition unit 104 recognizes the switch 1A which is in the projection region 10, as illustrated in
The distance information generation unit 106 serves as an input unit and generates distance information changed in accordance with a distance between a manipulator and an input recognition region. Specifically, the distance information generation unit 106 calculates the distance between the manipulator and the input recognition region on the basis of a recognition result of the recognition unit 104. For example, the distance information generation unit 106 calculates a distance between a position of the recognized manipulator and a position of the input recognition region in a 3-dimensional space. Then, the distance information indicating the calculated distance is supplied to the control unit 108. Note that the distance information may be an index or the like corresponding to the distance between the manipulator and the input recognition region in addition to the above-described calculated value. In addition, the distance information may be information regarding a distance between the manipulator and the input recognition region in a specific direction. For example, the distance information may be information regarding a distance between the manipulator and the input recognition region in a depth direction for the user.
The control unit 108 generally controls a process of the information processing device 100. Specifically, the control unit 108 serves as an output control unit and controls an output related to a virtual object on the basis of the distance information. More specifically the control unit 108 controls display of at least one first virtual object (hereinafter also referred to as a first feedback image or a first FB image) viewed in the display region on the basis of the distance information.
First, the control unit 108 controls display or non-display of the first FB image on the basis of the distance information and a threshold of the distance information. Specifically, in a case in which a value of the distance information is equal to or less than a first threshold determined in advance, the control unit 108 causes the projection device 300 to project the first FB image when the first FB image has not been projected. The control unit 108 causes the projection device 300 to continue the projection of the first FB image when the first FB image is being projected. Conversely, in a case in which the value of the distance information greater than a first threshold determined in advance, the control unit 108 causes the projection device 300 to stop projecting the first FB image when the first FB image is being projected, and causes the projection device 300 not to project the first FB image when the first FB image is not projected. In addition, the display position of the first FB image is a position on the display region corresponding to the position of the manipulator with regard to the manipulation region.
In addition, in a case in which the distance between the manipulator and the input recognition region is equal to or less than a second threshold, the control unit 108 may cause the projection device 300 to stop the projection of the first FB image. For example, the second threshold is a value of the distance between the manipulator and the input recognition region which is 0 or is considered to be 0. When the value of the distance information is equal to or less than the second threshold, the control unit 108 causes the projection device 300 to stop the projection of the first FB image. That is, when the manipulator reaches the input recognition region, the first FB image vanishes.
Note that display and non-display of the first FB image may be switched while the display is changed for a predetermined time. For example, the control unit 108 performs control such that transparency of the first FB image decreases over time when the first FB image is projected. In addition, the control unit 108 performs control such that the transparency of the first FB image increases over time when the projection of the first FB image is stopped. In addition, the change in the display is not limited to the change in the transparency. Gradation, hue, saturation, brightness, luminance, or the like may be changed or the first FB image may be partially be deleted for the change. In addition, the change or non-change in the display may be switched in accordance with a setting set by the user.
The control unit 108 controls a display aspect of the first FB image on the basis of the distance information. Here, the display of the first FB image is controlled in various patterns. Hereinafter, display control patterns of the first FB image will be described with reference to
The control unit 108 controls an aspect of the first FB image related to a display range of the first FB image on the basis of the distance information supplied from the distance information generation unit 106. As an aspect of the first FB image to be controlled, there is a size of the first FB image.
In addition, as an aspect of the first FB image to be controlled, there is a shape of the first FB image.
In addition, the control unit 108 controls display of the first FB image with a shape corresponding to a manipulator on the basis of the distance information. As the first FB image with the shape corresponding to the manipulator, there is an image evoking the shadow of the manipulator (hereinafter also simply referred to as the shape of the manipulator). Then, the control unit 108 controls depth of the shadow of the manipulator on the basis of the distance information.
Note that the control unit 108 controls display of the second virtual object corresponding to the manipulator (hereinafter also referred to as a manipulator image) and the first FB image may be displayed on the basis of the manipulator image. As the manipulator image, there is a generated image such as a captured image or an illustration image of a manipulation device or a part of the body of the user performing a manipulation. Specifically, the first FB image with a shape corresponding to the manipulator is decided on the basis of the manipulator image. For example, the manipulator image may be the virtual object 20 corresponding to a hand of the user illustrated in
In addition, the control unit 108 controls a virtual effect on the first FB image on the basis of the distance information. Here, the virtual effect on the first FB image means a virtual effect that has no influence on a display range of a size, a shape, or the like of the first FB image. As the control of the virtual effect on the first FB image, there is transparency of the first FB image. The description will be made with reference to
In addition, as control of the virtual effect on the first FB image, there is a color of the first FB image. The description will be made with reference to
In addition, as control of the virtual effect on the first FB image, there is gradation of the first FB image. The description will be made with reference to
In addition, the control unit 108 controls display density of the first FB image on the basis of the distance information. The description will be made with reference to
In addition, the control unit 108 controls a positional relation between the plurality of first FB images on the basis of the distance information. Specifically, the control unit 108 controls a distance between the plurality of first FB images in accordance with the distance information.
In addition, the control unit 108 controls display of the first FB image expressing a value of the distance information on the basis of the distance information. Specifically, the control unit 108 controls display of an indicator indicating the value of the distance information.
In addition, the control unit 108 may control a manipulator image as the first FB image. Specifically, the control unit 108 controls display of manipulator image on the basis of the distance information. The display control of the manipulator image may be substantially the same as a part of the display control of the above-described first FB image.
The control unit 108 further controls display of the first FB image on the basis of information regarding a manipulation target (hereinafter also referred to as manipulation target information). As the manipulation target information, there is information regarding an aspect of a manipulation target. For example, the control unit 108 controls the display of the first FB image on the basis of information regarding a size or a shape of the manipulation target. Further, display control of the first FB image based on the manipulation target information will be described with reference to
First, the control unit 108 acquires the manipulation target information before display of the first FB image starts. For example, when it is decided that the first FB image is displayed, the control unit 108 acquires information of a size, a shape, a color, or the like of a manipulation target recognized as the manipulation target information from the recognition unit 104.
Subsequently, after the control unit 108 decides display control of the first FB image (hereinafter also referred to as first display control) based on the distance information, the control unit 108 decides display control of the first FB image based on the manipulation target information (hereinafter also referred to as second display control). Specifically, the control unit 108 decides a size of the first FB image on the basis of the size of the manipulation target acquired from the recognition unit 104 in addition to the display control of the first FB image decided on the basis of the distance information. For example, in an example of Case 1 in
Note that when the size of the first FB image is decided in accordance with the size of the manipulation target, a virtual effect of the first FB image may be changed instead of the change in the size of the first FB image or together with the change in the size of the first FB image in a case in which the size of the first FB image is smaller than a predetermined size. For example, the control unit 108 may increase the transparency of the first FB image while reducing the first FB image to the predetermined size.
In addition, in a case in which the manipulation target is a virtual object which is not an actual object, the manipulation target may be extended. For example, when a virtual object of an icon such as a magnifying glass is displayed and a manipulation of selecting the icon is recognized, the manipulation target may be extended and displayed. Thus, it is possible to prevent the first FB image from becoming small.
(4. Display Control of Second FB Image after Input Recognition)
The control unit 108 further controls display of a third virtual object according to a manipulation target by a manipulator (hereinafter also referred to as a second feedback image or a second FB image) on the basis of the recognition of an input by the manipulator. Specifically, when the value of the distance information reaches a second threshold or less (that is, the manipulator reaches the input recognition region), the control unit 108 performs a process corresponding to an input by the manipulator assuming that the input is performed by the manipulator. In addition, when the value of the distance information reaches a second threshold or less, the control unit 108 causes the projection device 300 to project the second FB image in accordance with the manipulation target. Further, the display control of the second FB image will be described with reference to
The control unit 108 determines whether the value of the distance information reaches the second threshold or less. For example, the control unit 108 periodically determines whether the value of the distance information reaches the second threshold or less while the first FB image is projected.
When it is determined that the value of the distance information reaches the second threshold or less, the control unit 108 performs a process on the manipulation target corresponding to the input by the manipulator. For example, when it is determined that the value of the distance information reaches the second threshold or less, the control unit 108 switches a switch 1B illustrated in
In addition, the control unit 108 causes the projection device 300 to project the second FB image. For example, when it is determined that the value of the distance information reaches the second threshold or less, the control unit 108 causes the projection device 300 to project a second FB image 40A illustrated in the right drawing of
Note that recognition of the input by the manipulator may be realized using information other than the distance information. For example, the control unit 108 controls display of the second FB image on the basis of a pressure on the input recognition region from the manipulator. For example, the control unit 108 controls display or non-display of the second FB image in accordance with information regarding whether there is a pressure supplied from a pressure sensor or the like installed in the input recognition region as the measurement device 200. Further, a display aspect of the second FB image may be controlled in accordance with information regarding the degree of the pressure supplied from the pressure sensor.
In addition, the control unit 108 may control the display of the second FB image on the basis of a change in the shape of the manipulator. For example, the control unit 108 controls display or non-display of the second FB image in accordance with a change or non-change in the shape of a finger of the user recognized by the recognition unit 104 (for example, a change or the like to a shape applying a force). Further, the display aspect of the second FB image may be controlled in accordance with the degree of change in the shape.
Next, a process of the information processing device 100 will be described with reference to
The information processing device 100 acquires the distance information regarding the distance between the position of the manipulator and the input recognition region (step S402). Specifically, the recognition unit 104 recognizes the manipulator and the input recognition region on the basis of a measurement result of the measurement device 200 and the distance information generation unit 106 generates the distance information regarding the distance between the recognized manipulator and input recognition region.
Subsequently, the information processing device 100 determines whether the value of the distance information is equal to or less than the first threshold (step S404). Specifically, the control unit 108 determines whether the value of the generated distance information is equal to or less than the first threshold.
When it is determined that the value of the distance information is equal to or less than the first threshold (YES in step S404), the information processing device 100 acquires the manipulation target information (step S406). Specifically, when it is determined that the value of the distance information is equal to or less than the first threshold, the recognition unit 104 recognizes the aspect of the manipulation target and supplies the aspect of the manipulation target to the control unit 108.
Subsequently, the information processing device 100 decides control content of the first FB image on the basis of the distance information and the manipulation target information (step S408). Specifically, the control unit 108 decides the first display control on the first FB image on the basis of the distance information and further decides the second display control on the basis of the aspect of the manipulation target recognized by the recognition unit 104.
Subsequently, the information processing device 100 causes a display device to display the first FB image (step S410). Specifically, the control unit 108 causes the projection device 300 to project the first FB image by performing the decided first display control and second display control.
Subsequently, the information processing device 100 determines whether the value of the distance information is equal to or less than the second threshold (step S412). Specifically, the control unit 108 determines whether the value of the distance information is equal to or less than the second threshold after the projection of the first FB image.
When it is determined that the value of the distance information is equal to or less than the second threshold (YES in step S412), the information processing device 100 causes the display device to display the second FB image (step S414). Specifically, when it is determined that the value of the distance information is equal to or less than the second threshold, the control unit 108 causes the projection device 300 to project the second FB image.
Subsequently, the information processing device 100 determines whether the value of the distance information is greater than the second threshold (step S416). Specifically, the control unit 108 determines whether the value of the distance information is greater than the second threshold after the projection of the second FB image.
When it is determined that the value of the distance information is greater than the second threshold (YES in step S416), the information processing device 100 stops the display of the FB images (step S418). Specifically, when it is determined that the value of the distance information is greater than the second threshold, the control unit 108 causes the projection device 300 to stop the projection of the first and second FB images. Note that when the second FB image is projected in step S414, the control unit 108 may cause the projection device 300 to stop the projection of the first FB image.
In this way, according to the embodiment of the present disclosure, the information processing device 100 controls an output related to the virtual object on the basis of the distance information changed in accordance with a distance between a manipulator recognized in a different manipulation region from a display region in which the virtual object is viewed and an input recognition region in which an input by the manipulator in the manipulation region is recognized.
In the related art, in a case in which the display region and the manipulation region are provided in different places, it was difficult for a user to grasp a situation of a manipulation using a manipulator while the user views the display region. Therefore, the user performed a manipulation by the manipulator while confirming a manipulation result in the display region, and thus an intuitive manipulation interfered.
In contrast, according to the embodiment of the present disclosure, a reference for a distance between the position of a manipulator and a position at which an input of a manipulation by the manipulator is recognized (hereinafter also referred to as a sense of distance related to a manipulation) can be presented to a user. Therefore, the user can intuitively manipulate the manipulator while viewing the display region. Accordingly, it is possible to suppress deterioration in manipulability in a case in which the display region and the manipulation region are different. In particular, since it is more difficult for the user to grasp the sense of distance of a manipulation in a depth direction to the user in a 3-dimensional space than a manipulation in different directions (for example, vertical and horizontal directions), the information processing device 100 is beneficial. In addition, since an intuitive manipulation is possible, it is possible to accelerate a skill of an input manipulation.
In addition, the output includes display of at least one first virtual object. Therefore, it is possible to enable the user viewing the display region to ascertain the sense of distance related to a manipulation. In addition, since the sense of vision facilitates understanding of more information at a time than other senses, the sense of distance is enabled to grasp the sense of distance related to a manipulation at a relatively short time.
Note that the output may be an output other than the display of the first FB image. Specifically, the output may be a sound or a tactile vibration. For example, when the value of the distance information is equal to or less than the first threshold, occurrence of a sound such as a mechanical sound or a voice or a tactile vibration is started. In addition, when the value of the distance information is equal to or less than the second threshold, the sound or the tactile vibration is stopped. Note that kinds of sounds or tactile vibrations may be switched instead of stopping the sound or the tactile vibration. In addition, a sound or a tactile vibration may be used together with the display of the first FB image. For example, in a case in which light is inserted into a projection region or there is concern of projection by the projection device 300 interfering, a sound or a tactile vibration based on the distance information may be output instead of or together with the display of the first FB image. In this case, even in a situation in which visibility of the first FB image may not be ensured, the user is enabled to grasp the sense of distance related to a manipulation.
In addition, the information processing device 100 controls an aspect of the first virtual object related to a display range of the first virtual object on the basis of the distance information. Therefore, when the display range of the first FB image is changed, the user is enabled to be aware of a change in the sense of distance related to a manipulation. Accordingly, it is possible to promote an improvement in precision of the manipulation by the user.
In addition, the aspect of the first virtual object includes a size or a shape of the first virtual object. Therefore, by changing the size or the shape of the first FB image in accordance with the distance information, the user is enabled to intuitively understand the sense of distance related to a manipulation.
In addition, the information processing device 100 controls a positional relation between the plurality of first virtual objects on the basis of the distance information. Therefore, for example, by changing the distances of the plurality of first FB images in accordance with the distance information, the user is enabled to understand the sense of distance related to a manipulation intuitively and more accurately.
In addition, the information processing device 100 controls a virtual effect on the first virtual object on the basis of the distance information. Therefore, by using the virtual effect of not changing the display range of the first FB image, the sense of distance related to a manipulation can be presented to the user even in a case in which the sufficient display range may not be ensured in the display of the first FB image. Accordingly, visibility of the display region and conveyance of the sense of distance related to the manipulation can be caused to be compatible.
In addition, the control of the virtual effect on the first virtual object includes control of transparency, gradation, or a color of the first virtual object. Therefore, the user is enabled to effectively recognize the sense of distance related to a manipulation without changing the display range.
In addition, the information processing device 100 controls display density of the first virtual object on the basis of the distance information. Therefore, by controlling magnitude of the display density in accordance with the distance information, the user is enabled to intuitively understand the sense of distance related to a manipulation.
In addition, the first virtual object includes a virtual object with a shape corresponding to the manipulator. Therefore, the user is enabled to easily understand that the first FB image is related to the manipulator manipulated by the user.
In addition, the first virtual object includes a virtual object expressing a value of the distance information. Therefore, the sense of distance related to an accurate manipulation by the user can be presented to the user. Accordingly, it is possible to further accelerate a skill of an input manipulation.
In addition, the information processing device 100 further controls display of the second virtual object corresponding to the manipulator, and the first virtual object includes a virtual object displayed on the basis of the second virtual object. Therefore, by displaying the manipulator image to correspond to the manipulator, the user is enabled to easily grasp a place corresponding to the manipulation by the user in the display region. Further, by displaying the first FB image on the basis of the manipulator image, it is possible to alleviate the sense of discomfort of the user to the first FB image. Accordingly, the user can naturally understand the sense of distance related to the manipulation.
In addition, the display region includes a region including a manipulation target by the manipulator, and the information processing device 100 further controls the display of the first virtual object on the basis of information regarding the manipulation target. Therefore, for example, by controlling the display of the first FB image in accordance with the aspect of the manipulation target, it is possible to prevent one of visibility of the first FB image and visibility of the manipulation target from deteriorating one-sidedly. Accordingly, maintenance or improvement in manipulability of the manipulation target and grasping of the sense of distance related to a manipulation can be caused to be compatible.
In addition, the information processing device 100 controls display or non-display of the first virtual object on the basis of a threshold of the distance information. Therefore, only in a situation in which the user is estimated to desire the sense of distance related to the manipulation, the first FB image can be caused to be displayed. Accordingly, it is possible to present the user with the sense of distance related to the manipulation while suppressing deterioration in visibility of the display region.
In addition, the information processing device 100 controls display of a third virtual object in accordance with a manipulation target by the manipulator on a basis of recognition of an input by the manipulator. Therefore, it is possible to present the user with the fact that an input is received. Accordingly, the user can move to a subsequent manipulation or stop the manipulation, and thus it is possible to improve usability.
The embodiment of the present disclosure has been described above. Note that the embodiment of the present disclosure is not limited to the above-described example. Hereinafter, first to sixth modification examples of the embodiment of the present disclosure will be described.
According to a first modification example of the embodiment of the present disclosure, the first FB image may be displayed without displaying the manipulator image. Specifically, the control unit 108 causes only the first FB image to be displayed in the display region. Hereinafter, display control patterns of the first FB images according to the modification example will be described with reference to
The control unit 108 controls an aspect and a visual effect of the first FB image related to the display range of the first FB image on the basis of the distance information. Specifically, the size of the first FB image and transparency of the first FB image are controlled.
In addition, the control unit 108 controls the shape of the first FB image on the basis of the distance information.
In addition, the control unit 108 controls display of the first FB image such as the shadow of the manipulator on the basis of the distance information.
In addition, the control unit 108 controls a distance between the plurality of first FB images and display of other first FB images indicating a relation between the plurality of first FB images in accordance with the distance information.
In addition, the control unit 108 controls display of the first FB image expressing a value of the distance information on the basis of the distance information.
In this way, according to the first modification example, the information processing device 100 causes the display device to display the first FB image without displaying the manipulator image. Therefore, it is possible to improve the visibility of the first FB image. In addition, in a case in which the manipulation target is in the display region, it is possible to also improve the visibility of the manipulation target.
According to a second modification example of the embodiment of the present disclosure, the information processing device 100 may control the display of the first FB image on the basis of a position of manipulation target. Specifically, the control unit 108 causes the display device to display the first FB image, eluding the manipulation target. Further, a process of the modification example will be described with reference to
The control unit 108 causes the display device to display the first FB image when the value of the distance information is equal to or less than the first threshold. For example, as illustrated in the left drawing of
Subsequently, the control unit 108 causes the first FB image to be moved in accordance with movement of a manipulator. For example, the control unit 108 causes the projection device 300 to change a projection position of the first FB image 30H to a position on the display region in accordance with the movement of the finger of the user. Note that the virtual object 20 is also caused to be moved similarly to the first FB image 30H.
In addition, in a case in which it is determined that the first FB image overlaps the manipulation target in a movement destination, the control unit 108 decides the position of the first FB image to a position separate from the manipulation target. For example, in a case in which the virtual object 20 is caused to be moved in accordance with the movement of the manipulator, as illustrated in the right drawing of
In this way, according to the second modification example, the information processing device 100 controls display of the first FB image on the basis of the position of the manipulation target. Therefore, the first FB image can be displayed at the position at which the manipulation target does not overlap the first FB image. Accordingly, the visibility of the first FB image and the visibility of the manipulation target can be caused to be compatible and it is possible to more reliably suppress the deterioration in the manipulability.
Note that the information processing device 100 may select the first FB image on the basis of a color, a shape, or a texture of the manipulation target or the display region. For example, the control unit 108 causes the projection device 300 to project the first FB image with a color, a shape, or a texture which is not the same as or analogous to the manipulation target or the display region. In this case, it is possible to ensure the visibility of the first FB image.
According to a third modification example of the embodiment of the present disclosure, the number of manipulators for manipulating a manipulation target may be plural. Specifically, the control unit 108 controls the display of the first FB image in accordance with the number of manipulators. For example, the control unit 108 causes the display device to display the number of first FB images corresponding to the number of manipulators. Further, a process according to the modification example will be described with reference to
The control unit 108 causes the display device to display the first FB image with regard to each of a plurality of manipulators recognized by the recognition unit 104. For example, when the recognition unit 104 recognizes two fingers (the thumb and the forefinger) illustrated in
Further, the control unit 108 may control the display of the second FB image on the basis of manipulations on the plurality of manipulators. Specifically, when inputs by the plurality of manipulators are recognized, the control unit 108 causes the display device to display the second FB image and controls an aspect of the second FB image in response to continuously input manipulations. Display control of the second FB image based on manipulations on the plurality of manipulators will be described in detail with reference to
When the values of the distance information with regard to the plurality of manipulators are equal to or less than the second threshold, the control unit 108 causes the display device to display the second FB image. For example, when a thumb and a forefinger of a user are manipulators and each value of the distance information related to the manipulators reaches the second threshold or less, a second FB image 40B with a triangular shape illustrated in
Subsequently, the control unit 108 controls a display aspect of the second FB image on the basis of a manipulation performed when the value of the distance information is equal to or less than the second threshold. For example, when a manipulation of rotating a hand is performed with two fingers, which are manipulators, being fixed with respect to the hand, as illustrated in the right drawing of
In this way, according to the third modification example, the information processing device 100 controls the display of the first FB image in accordance with the number of manipulators. Therefore, in a complicated manipulation using a plurality of manipulators, the sense of distance related to a manipulation can be presented to the user. Accordingly, it is possible to improve manipulability in the complicated manipulation.
In addition, the information processing device 100 controls the display of the second FB image on the basis of the manipulations by the plurality of manipulators. Therefore, the user can manipulate the manipulation target in which a manipulation more complicated than a manipulation by a single manipulator is requested. Accordingly, it is possible to extend an application range of the information processing device 100.
According to a fourth modification example of the embodiment of the present disclosure, the information processing device 100 controls display of the first FB image on the basis of information regarding a user manipulating a manipulator (hereinafter also referred to as user information). As the user information, there is information regarding of a skill level of the manipulation by the user (hereinafter also referred to as skill level information). For example, the recognition unit 104 identifies the user performing a manipulation and the control unit 108 acquires the skill level information of the identified user. Then, the control unit 108 controls the display of the first FB image on the basis of the skill level information. Further, a process according to the modification example will be described in detail with reference to
The recognition unit 104 recognizes the user performing the manipulation in the manipulation region. For example, the recognition unit 104 acquires image information in which the user performing the manipulation in the manipulation region is shown from the measurement device 200 via the communication unit 102. Then, the recognition unit 104 recognizes the user using a face recognition technology on the basis of the acquired image information. Note that the method of recognizing the user is not limited thereto and any other method may be used.
Subsequently, the control unit 108 acquires the skill level information of the recognized user. For example, the control unit 108 calculates the skill level of the manipulation from a manipulation history of the user recognized by the recognition unit 104. Note that the skill level information may be stored in the information processing device 100 or may be stored in an external device.
Then, the control unit 108 determines an aspect of the first FB image on the basis of the acquired skill level information. Specifically, the control unit 108 causes the display device to display the first FB image with a size in accordance with the skill level information. For example, the control unit 108 causes the projection device 300 to project the larger first FB image as the calculated or acquired skill level is higher. As illustrated in
In addition, as the user information, there is identification information of the user. For example, the control unit 108 controls the aspect of the first FB image depending on an individual user identified by the recognition unit 104. In addition, as the user information, there is attribute information of the user. For example, the control unit 108 controls the aspect of the first FB image depending on an age, a sex, an address, nationality, or the like of the user recognized by the recognition unit 104.
In this way, according to the fourth modification example, the information processing device 100 further controls the display of the first FB image on the basis of information related to the user manipulating the manipulator. Here, a manipulation behavior generally differs for each user for which a skill level, a nature, or the like of the user is a factor. Therefore, as in the modification example, by causing the first FB image to be changed in accordance with the user, it is possible to supply the sense of distance related to a manipulation appropriate for the individual user.
According to a fifth modification example of the embodiment of the present disclosure, the first threshold of the distance information may be variable. Specifically, the control unit 108 controls the first threshold on the basis of the distance information. For example, in a case in which a difference between the value of the distance information and the first threshold is within a predetermined range, the control unit 108 changes the first threshold to a value at which the difference between the value of the distance information and the first threshold stretches. A process of the modification example will be described with reference to
First, control of the first threshold before the display of the first FB image will be described with reference to Case 3 of
Thereafter, when the value of the distance information reaches the lowered first threshold or less, the control unit 108 causes the display device to display the first FB image. For example, when the first threshold is lowered to the first threshold 60B and the finger of the user is subsequently further lowered to the first threshold 60B or less, an input by the finger of the user is recognized. That is, the first FB image is projected to a projection region.
Subsequently, control of the first threshold after the display of the first FB image will be described with reference to Case 4 of
Thereafter, when the value of the distance information reaches a value greater than the raised first threshold or more, the control unit 108 causes the display device to stop the display of the first FB image. For example, when the first threshold is raised to the first threshold 60C and the finger of the user is subsequently further raised to the first threshold 60C, the projection of the first FB image is stopped.
Note that the control unit 108 may control the first threshold on the basis of information regarding the manipulator (hereinafter also referred to as manipulator information). As the manipulator information, there is information regarding an aspect of a manipulator. Specifically, the control unit 108 controls the first threshold on the basis of a size of the manipulator. For example, the first threshold is set so that the first threshold is higher as the finger of the user is larger. In addition, the control unit 108 may control the first threshold on the basis of a speed of movement of the manipulator. For example, the first threshold is set so that the first threshold is higher as a movement speed or acceleration of the finger of the user is faster.
In addition, the control unit 108 may control the first threshold on the basis of the user information. Specifically, the control unit 108 may control the first threshold on the basis of the attribute information of the user. As the attribute information of the user, there is information regarding an age, a sex, handedness, or the like of the user. For example, the first threshold is set so that the first threshold is higher as the age of the user is higher.
Further, a process of controlling the first threshold on the basis of the manipulator information or the user information will be described with reference to
The information processing device 100 acquires the manipulator information or the user information (step S502). Specifically, the recognition unit 104 recognizes the manipulator and an aspect of the manipulator which is in the manipulation region and supplies information regarding the aspect of the recognized manipulator to the control unit 108. In addition, the recognition unit 104 recognizes a user which is in the manipulation region using a face recognition technology or the like and supplies identification information or the like of the recognized user to the control unit 108.
Subsequently, the information processing device 100 decides the first threshold on the basis of the manipulator information or the user information (step S504). Specifically, the control unit 108 changes the first threshold from a reference value on the basis of information regarding the acquired aspect of the manipulator. In addition, the control unit 108 changes the first threshold from the reference value on the basis of the acquired identification information of the user and the attribute information of the user to be acquired. Of course, the first threshold may not be changed as a result in some cases.
Since the processes (steps S506 to S522) after the first threshold is decided are substantially the same as the processes described with reference to
In addition, the plurality of first thresholds of the distance information may be provided. Specifically, the control unit 108 selects the first threshold to be used for a display control process for the first FB image from the plurality of first threshold values on the basis of the manipulator information or the user information. For example, in a case in which a speed of movement of the manipulator is equal to or greater than a predetermined value, the control unit 108 selects the first threshold higher than the reference value (that is, a distance between the manipulator and the input recognition region is long). Then, when the value of the distance information reaches the selected first threshold or less, the control unit 108 causes the projection device 300 to project the first FB image.
In this way, according to the fifth modification example, the information processing device 100 controls the first threshold on the basis of the distance information. Therefore, by changing the firs threshold depending on a situation, it is possible to optimize a display time of the first FB image. Accordingly, the user is enabled to grasp the sense of distance related to the manipulation more reliably and visibility of the display region can be prevented from deteriorating due to the display of the first FB image.
In addition, the information processing device 100 controls the first threshold on the basis of the information regarding the manipulator. Here, as the manipulator is larger and the movement of the manipulator is faster, a time taken until the manipulator reaches the input recognition region is shorter, and thus the display time of the first FB image is shorter. Accordingly, by setting the first threshold in accordance with the manipulator, it is possible to optimize the display time of the first FB image. Thus, the user is enabled to reliably grasp the sense of distance related to the manipulation.
In addition, the information processing device 100 controls the first threshold on the basis of information regarding the user. Here, even in a case in which the first FB image is presented, a time taken for the user to grasp the sense of distance related to the manipulation generally differs. For example, as the age of the user is higher, a time take to grasp the sense of distance related to the manipulation is longer. Accordingly, by setting the first threshold value in accordance with the user, it is possible to optimize the display time of the first FB image for the user. Thus, the user is enabled to reliably grasp the sense of distance related to the manipulation.
In addition, the information processing device 100 selects the first threshold to be used for the display control of the first FB image from the plurality of first thresholds. Therefore, the process of controlling the first threshold value can be further simplified than in a case in which the single first threshold varies. Accordingly, it is possible to accelerate the process of controlling the first threshold, and thus it is possible to suppress a time lag between the manipulation by the user and the display of the first FB image.
According to a sixth modification example of the embodiment of the present disclosure, the number of users performing manipulations may be plural. Specifically, the control unit 108 performs display control of the first FB image based on distance information on each of manipulators manipulated by the plurality of users. For example, the recognition unit 104 recognizes the plurality of users performing manipulations in manipulation regions and recognizes each manipulator manipulated by each user. The distance information generation unit 106 generates distance information of each of the recognized manipulators. The control unit 108 causes the projection device 300 to project each first FB image on the basis of whether the value of the generated distance information is equal to or greater than the first threshold with regard to each manipulator.
In addition, the control unit 108 causes the display device to display each first FB image and each manipulator image in accordance with each manipulator related to the user. Specifically, each first FB image in an identification aspect in accordance with each user is displayed. For example, the first FB image with a color, a pattern, a shape, or a figure in accordance with each user is projected. Further, the identification aspect of the first FB image may be decided on the basis of information regarding the recognized user. For example, the color of the first FB image may be decided from a color or the like of a cloth which the recognized user wears.
In addition, the manipulator image with the identification aspect in accordance with each user is displayed. For example, the control unit 108 acquires an image of each of the manipulators related to the plurality of users from a captured image and causes the projection device 300 to project each of the acquired images. Note that the manipulator image may be a generated image such as an illustration image, as described above. In this case, the generation image corresponding to each user is used.
Note that only in a case in which the manipulator image in accordance with the manipulator related to the user is not displayed, the first FB image in accordance with the manipulator related to the user may be displayed.
In this way, according to the sixth modification example, the information processing device 100 performs the display control of the first FB image based on the distance information on each of the manipulators manipulated by the plurality of users. Therefore, even in a case in which the plurality of users simultaneously perform manipulations in the manipulation regions provided in different places from the display region, the deterioration in the manipulability of each user can be prevented from deteriorating by presenting the sense of distance related to the manipulation to each user. In addition, by displaying the first FB image so that the first FB image can be individually identified by each user, each user can perform the manipulation without being confused.
The information processing devices 100 according to an embodiment of the present disclosure have been described above. The processing performed by the information processing device 100 is achieved by cooperatively operating software and hardware of the information processing device 100 described below.
The processor 132 functions as an arithmetic processing device, and achieves the functions of the speech recognition unit 104, a distance information generation unit 106, and the control unit 108 in the information processing device 100 by cooperatively operating with various programs. The processor 132 causes various logical functions of the information processing device 100 to operate, by using a control circuit to execute programs stored in the memory 134 or another storage medium. For example, the processor 132 may be a central processing unit (CPU), a graphics processing unit (GPU), a digital signal processor (DSP), or a system on chip (SoC).
The memory 134 stores a program, a calculation parameter, and the like used by the processor 132. For example, the memory 134 includes a random access memory (RAM), and transiently stores programs used for executing the processor 132, parameters or the like that change as appropriate when executing the processor 132. In addition, the memory 134 includes a read only memory (ROM), and functions as the storage unit by using the RAM and the ROM. Note that, an external storage device may be used as a part of the memory 134 via the connection port 150 or the communication device 152.
Note that, the processor 132 and the memory 134 are connected to each other via an internal bus including a CPU bus or the like.
The bridge 136 connects buses. Specifically, the bridge 136 connects the internal bus and the bus 138. The internal bus is connected to the processor 132 and the memory 134. The bus 138 is connected to the interface 140.
The input device 142 is used by a user for operating the information processing device 100 or inputting information to the information processing device 100. For example, the input device 142 includes an input means to which the user inputs information, and an input control circuit that generates an input signal on the basis of the user input and outputs the generated input signal to the processor 132. Note that, the input means may be a mouse, a keyboard, a touchscreen, a switch, a lever, a microphone, or the like. By operating the input unit 142, the user of the information processing device 100 can input various kinds of data into the information processing device 100 and instruct the information processing device 100 to perform a processing operation.
The display device 144 is used to notify a user of information. For example, the output device 144 may be a device such as a liquid crystal display (LCD) device, an organic light emitting diode (OLED) device, a projector, a speaker, or a headphone, or may be a module configured to output information to such a device.
Note that, the input device 142 or the output device 144 may include an input/output device. For example, the input/output device may be a touchscreen.
The storage device 146 is a data storage device. The storage device 146 may include a storage medium, a recording device that records data on the storage medium, a reading device that reads data from the storage medium, a deletion device that deletes data recorded on the storage medium, and the like. The storage device 146 stores various kinds of data or a program to be executed by the CPU 132.
The drive 148 is a reader/writer for a storage medium, and is incorporated in or externally attached to the information processing device 100. The drive 148 reads information stored in a removable storage medium that is mounted, such as a magnetic disk, an optical disc, a magneto-optical disk, or semiconductor memory, and outputs the information to the memory 134. The drive 148 is also capable of writing information to the removable storage medium.
The connection port 150 is a port used to directly connect apparatuses to the information processing device 100. For example, the connection port 150 may be a USB (Universal Serial Bus) port, an IEEE1394 port, or a SCSI (Small Computer System Interface) port. In addition, the connection port 150 may be an RS-232C port, an optical audio terminal, an HDMI (registered trademark) (High-Definition Multimedia Interface) port, or the like. Data may be exchanged between the information processing device 100 and an external apparatus by connecting the external apparatus to the connection port 150.
The communication device 152 mediates communication between the information processing device 100 and an external device, and functions as the communication unit 102. Specifically, the communication device 152 establishes communication in accordance with a wireless communication scheme or a wired communication scheme. For example, the communication device 152 establishes wireless communication in accordance with a cellular communication scheme such as Wideband Code Division Multiple Access (W-CDMA) (registered trademark), WiMAX (registered trademark), Long-Term Evolution (LTE), or LTE-A. Note that, the communication device 152 may establish wireless communication in accordance with any wireless communication scheme like a short-range wireless communication such as Bluetooth (registered trademark), near-field communication (NFC), wireless USB, or TransferJet (registered trademark), or a wireless local area network (LAN) such as Wi-Fi (registered trademark). In addition, the communication device 152 may establish wired communication such as signal line communication or wired LAN communication.
Note that, the information processing device 100 does not have to include a part of the structural elements described with reference to
As described above, according to the embodiment of the present disclosure, the reference of a distance between the position of a manipulator and a position at which an input of a manipulation by the manipulator is recognized (hereinafter also referred to a sense of distance related to the manipulation) can be presented to a user. Therefore, the user can intuitively manipulate the manipulator while viewing the display region. Accordingly, it is possible to suppress deterioration in manipulability in a case in which the display region and the manipulation region are different. In particular, since it is more difficult for the user to grasp the sense of distance of a manipulation in a depth direction to the user in a 3-dimensional space than a manipulation in different directions (for example, vertical and horizontal directions), the information processing device 100 is beneficial. In addition, since an intuitive manipulation is possible, it is possible to accelerate a skill of an input manipulation.
The preferred embodiment(s) of the present disclosure has/have been described above with reference to the accompanying drawings, whilst the present disclosure is not limited to the above examples. A person skilled in the art may find various alterations and modifications within the scope of the appended claims, and it should be understood that they will naturally come under the technical scope of the present disclosure.
For example, in the foregoing embodiment, the example in which the distance information is information regarding a distance in the depth direction to the user has been mainly described, but the present technology is not limited to the example. For example, the distance information may be one direction among the vertical and horizontal directions to the user or another specific direction.
In addition, the foregoing embodiment, the example in which the measurement device 200 is a depth sensor has been mainly described, but the measurement device 200 may be a visible-light imaging device, a laser ranging sensor, an ultrasonic sensor, or the like.
In addition, the foregoing embodiment, the example in which a virtual object such as the first FB image is projected by the projection device 300 has been mainly described, but the virtual object may be displayed by another display device in which a manipulation region and a display region are different. For example, the virtual object may be displayed by a head-up display (HUD) in which light of an external world image is penetrated and an image is displayed on a display unit or image light related to an image is projected to the eyes of a user, a head mount display (HMD) in which a captured external world image and an image are displayed, or the like.
In addition, in the foregoing embodiment, the example in which the manipulation region and the input recognition region are recognized by the recognition unit 104 has been described, but the manipulation region and the input recognition region may be acquired in accordance with another method. For example, information such as positions, areas, or the like of the manipulation region and the input recognition region may be stored in advance in the information processing device 100 or may be acquired from an external device through communication. In addition, the manipulation region and the input recognition region may be fixed.
In addition, the information processing system according to the embodiment of the present disclosure may be a client-server type of system or a cloud service type of system. For example, the information processing device 100 may be a server installed in a different place from the places in which the measurement device 200 and the projection device 300 are installed.
Further, the effects described in this specification are merely illustrative or exemplified effects, and are not limitative. That is, with or in the place of the above effects, the technology according to the present disclosure may achieve other effects that are clear to those skilled in the art from the description of this specification.
Further, not only a process in which steps shown in the flowcharts of the above embodiments are performed in a time-series manner in accordance with a described sequence but also a process in which the steps are not necessarily processed in a time-series manner but are executed in parallel or individually is included. Also, it is self-evident that even steps processed in a time-series manner can be appropriately changed in sequence depending on circumstances.
In addition, a computer program for causing hardware built in the information processing device 100 to exhibit functions equivalent to those of the above-described respective logical elements of the information processing device 100 can also be produced. Furthermore, a storage medium in which the computer program is stored is also provided.
Additionally, the present technology may also be configured as below.
(1)
An information processing device including:
an input unit configured to obtain distance information changed in accordance with a distance between a manipulator recognized in a different manipulation region from a display region in which a virtual object is viewed and an input recognition region in which an input by the manipulator in the manipulation region is recognized; and
an output control unit configured to control an output related to the virtual object on a basis of the distance information.
(2)
The information processing device according to (1), in which the output includes display of at least one first virtual object.
(3)
The information processing device according to (2), in which the output control unit controls an aspect of the first virtual object related to a display range of the first virtual object on the basis of the distance information.
(4)
The information processing device according to (3), in which the aspect of the first virtual object includes a size or a shape of the first virtual object.
(5)
The information processing device according to any one of (2) to (4), in which the output control unit controls a positional relation between the plurality of first virtual objects on the basis of the distance information.
(6)
The information processing device according to any one of (2) to (5), in which the output control unit controls a virtual effect on the first virtual object on the basis of the distance information.
(7)
The information processing device according to (6), in which the control of the virtual effect on the first virtual object includes control of transparency, gradation, or a color of the first virtual object.
(8)
The information processing device according to any one of (2) to (7), in which the output control unit controls display density of the first virtual object on the basis of the distance information.
(9)
The information processing device according to any one of (2) to (8), in which the first virtual object includes a virtual object with a shape corresponding to the manipulator.
(10)
The information processing device according to any one of (2) to (9), in which the first virtual object includes a virtual object expressing a value of the distance information.
(11)
The information processing device according to any one of (2) to (10),
in which the output control unit further controls display of a second virtual object corresponding to the manipulator, and
the first virtual object includes a virtual object displayed on a basis of the second virtual object.
(12)
The information processing device according to any one of (2) to (11),
in which the display region includes a region including a manipulation target of the manipulator, and
the output control unit further controls the display of the first virtual object on a basis of information regarding the manipulation target.
(13)
The information processing device according to any one of (2) to (12), in which the output control unit further controls the display of the first virtual object in accordance with a number of the manipulators.
(14)
The information processing device according to any one of (2) to (13), in which the output control unit further controls the display of the first virtual object on a basis of information regarding a user manipulating the manipulator.
(15)
The information processing device according to any one of (2) to (14), in which the output control unit controls display or non-display of the first virtual object on a basis of a threshold of the distance information.
(16)
The information processing device according to (15), in which the output control unit controls the threshold on the basis of the distance information.
(17)
The information processing device according to (15) or (16), in which the output control unit controls the threshold on a basis of information regarding the manipulator or information regarding a user manipulating the manipulator.
(18)
The information processing device according to any one of (2) to (17), in which the output control unit further controls display of a third virtual object in accordance with a manipulation target of the manipulator on a basis of recognition of an input by the manipulator.
(19)
An information processing method including, by a processor:
obtaining distance information changed in accordance with a distance between a manipulator recognized in a different manipulation region from a display region in which a virtual object is viewed and an input recognition region in which an input by the manipulator in the manipulation region is recognized; and
controlling an output related to the virtual object on a basis of the distance information.
(20)
A program causing a computer to realize:
an input function of obtaining distance information changed in accordance with a distance between a manipulator recognized in a different manipulation region from a display region in which a virtual object is viewed and an input recognition region in which an input by the manipulator in the manipulation region is recognized; and
an output control function of controlling an output related to the virtual object on a basis of the distance information.
Number | Date | Country | Kind |
---|---|---|---|
2016-108409 | May 2016 | JP | national |
Filing Document | Filing Date | Country | Kind |
---|---|---|---|
PCT/JP2017/014910 | 4/12/2017 | WO | 00 |