This application claims priority to Japanese Patent Application No. 2020-211166 filed on Dec. 21, 2020, incorporated herein by reference in its entirety.
The present specification discloses a display system, a display device, and a program for displaying an augmented reality (AR) image.
A display device using augmented reality technology has been known. For example, in Japanese Unexamined Patent Application Publication (Translation of PCT Application) No. 2016-522485 (JP 2016-522485 A), an augmented reality image in which a real object such as an action figure that is a toy is replaced with a virtual object such as a virtual action figure with animation is displayed.
The augmented reality technology as described above enables remote communication. In a facility such as a theme park, for example, with the augmented reality technology, it is possible to produce an effect as if a person at a remote location was traveling around the facility, by superimposing an image of the person at the remote location on a captured image of the facility.
The present specification discloses a display system, a display device, and a program capable of displaying an augmented reality image that matches the concept of the surroundings.
The present specification discloses a display system. The display system includes a display device and a server. The display device includes an image recognition unit, a display control unit, a display unit, and a position information acquisition unit. The image recognition unit recognizes a person image from an image captured at a remote location. The display control unit is configured to superimpose a decoration image on the person image and generate an augmented reality image in which the person image that has been decorated is superimposed on scenery of a real world. The display unit is configured to display the augmented reality image. The position information acquisition unit acquires a position of the position information acquisition unit. The server includes a storage unit and an extraction unit. In the storage unit, a plurality of kinds of the decoration images is stored. The extraction unit extracts, from the storage unit, the decoration image based on the position acquired by the position information acquisition unit, as the decoration image to be superimposed on the person image.
According to the above configuration, a virtual image in which the person image of the remote location is decorated based on the position of the display device can be superimposed on the display unit. This makes it possible to produce an effect that the person image of the remote location is decorated in accordance with the concept of the facility in which the display device is located.
In the above configuration, the display device may be disposed in a facility configured based on a specific theme. In this case, in the storage unit, a character image defined as a character in the facility is stored as the decoration image.
According to the above configuration, it is possible to produce an effect that the person at the remote location plays a character in the facility.
In the above configuration, an identifier that is imageable may be displayed on a device in the facility. In this case, the display device includes an imager configured to image an inside of the facility. When the image recognition unit recognizes the identifier in an in-facility captured image captured by the imager, the display control unit sets an image area including the identifier as a superimposed area of the person image that has been decorated.
According to the above configuration, since the superimposed area of the person image of the remote location in the in-facility captured image is defined, unnatural superimposition such as superimposition of the person image in the air is suppressed.
In the above configuration, the display system may include a first caller configured to make a call in the facility, and a second caller configured to make a call with the first caller at the remote location.
According to the above configuration, it is possible to have a conversation with the person superimposed in the augmented reality image in the facility.
The present specification also discloses a display device. The display device includes an image recognition unit, a display control unit, a display unit, a position information acquisition unit, a storage unit, and an extraction unit. The image recognition unit recognizes a person image from an image captured at a remote location. The display control unit is configured to superimpose a decoration image on the person image and generate an augmented reality image in which the person image that has been decorated is superimposed on scenery of a real world. The display unit is configured to display the augmented reality image. The position information acquisition unit acquires a position of the position information acquisition unit. In the storage unit, a plurality of kinds of the decoration images is stored. The extraction unit extracts, from the storage unit, the decoration image based on the position acquired by the position information acquisition unit, as the decoration image to be superimposed on the person image.
The present specification also discloses a program. The program causes a computer to function as an image recognition unit, a display control unit, a display unit, a position information acquisition unit, a storage unit, and an extraction unit. The image recognition unit recognizes a person image from an image captured at a remote location. The display control unit is configured to superimpose a decoration image on the person image and generate an augmented reality image in which the person image that has been decorated is superimposed on scenery of a real world. The display unit is configured to display the augmented reality image. The position information acquisition unit acquires a position of the position information acquisition unit. In the storage unit, a plurality of kinds of the decoration images is stored. The extraction unit extracts, from the storage unit, the decoration image based on the position acquired by the position information acquisition unit, as the decoration image to be superimposed on the person image.
With the display system, the display device, and the program disclosed in the present specification, it is possible to display an augmented reality image that matches the concept of the surroundings.
Features, advantages, and technical and industrial significance of exemplary embodiments of the present disclosure will be described below with reference to the accompanying drawings, in which like signs denote like elements, and wherein:
Configuration of Complex Entertainment Facility
The complex entertainment facility 10 includes a plurality of theme parks 14 to 18. The theme park refers to a facility having a concept based on a specific theme (subject) and including facilities, events, scenery, and the like that are comprehensively organized and produced based on that concept. For example, the theme parks 14 to 18 are connected by connecting passages 20A, and users can come and go between the theme parks 14 to 18 through the connecting passages 20A.
The complex entertainment facility 10 includes theme parks having different themes. For example, the complex entertainment facility 10 includes an amusement park 14, an aquarium 16, and a zoo 18 as the theme parks.
Characters are set for each of the theme parks 14 to 18 based on their respective themes. The characters are set so as to match the theme and the concept of each of the theme parks 14 to 18. For example, for the amusement park 14, characters such as a clown and a go-kart are set. For example, for the aquarium 16, characters such as a dolphin, goldfish, and a shark are set. Further, for example, for the zoo 18, characters such as an elephant, a lion, and a panda are set.
Images of these characters (hereinafter, appropriately referred to as character images) are used as decoration images for decorating a person image captured at a remote location, as will be described later. The character image data of the theme parks 14 to 18 is stored in a decoration image storage unit 82 of a server 70 (see
In each of the theme parks 14 to 18, in-park passages 20B to 20D are provided. The in-park passages 20B to 20D are connected to the connecting passages 20A, and a vehicle 90 can travel through these passages.
A beacon transmitter 22 is provided along the connecting passages 20A and the in-park passages 20B to 20D of the complex entertainment facility 10. A plurality of transmitters 22 are provided, for example, at equal intervals. As will be described later, when a beacon receiver 37 (see
The display device 30B and an imager 35B are disposed in the house 92 at the remote location distant from the complex entertainment facility 10. The display device 30B and the imager 35B are operated by, for example, a person 94 in the house 92. The imager 35B and the display device 30B only need to be placed in such an environment that the imager 35B and the display device 30B can be operated by the person 94 at the remote location and can communicate with the display device 30A and the imager 35C. For example, the imager 35B and the display device 30B may be placed outdoors, instead of being placed in the house 92.
As described in detail below, a display system according to the present embodiment causes the display device 30A carried by a visitor of the complex entertainment facility 10 to display an image of the person 94 at the remote location. Examples of the visitor to the complex entertainment facility 10 include a family, and examples of the person 94 at the remote location include a guardian included in the family and on assignment at the remote location, an elderly relative having mobility difficulties, and the like.
Further, in the display system according to the present embodiment, when displaying the image of the person 94 at the remote location on the display device 30A, image processing of superimposing a decoration image on the person image is executed. For the decoration image, a character image set for the theme park, out of the theme parks 14 to 18, where the display device 30A is located is extracted.
Configuration of Devices of Display System
Here, focusing on the function that the image of the person 94 at the remote location is decorated with the decoration image and the decorated image is superimposed on the scenery of the complex entertainment facility 10 that is the scenery of the real world, the display system only needs to include at least the display device 30A, the imager 35B, and the server 70. Configuration of Display Device 30B
The display device 30B is operated and used by the person 94 at the remote location (see
The display device 30B also includes a display unit 46B and a display control unit 45B that controls a display image on the display unit 46B. Further, the display device 30B includes an input unit 47B for inputting information. For example, the display device 30B is provided with a touch panel display in which the input unit 47B and the display unit 46B are integrated. The display device 30B is also provided with a caller 32B (second caller) that can make a call with the display device 30A from the remote location. Further, the display device 30B includes an input-output controller 39B that manages input and output of information.
The display device 30B can wirelessly communicate with the display device 30A via a communication line, a base station of a telecommunications carrier, or the like (not shown). The data to be transmitted includes, for example, voice data sent from the person 94 at the remote location to the caller 32B.
The imager 35B is disposed near the display device 30B and can be operated by the person 94 at the remote location. The imager 35B includes an imaging device such as a complementary metal oxide semiconductor (CMOS) imaging device or a charge coupled device (CCD) imaging device. The imager 35B can capture a still image and a moving image.
The imager 35B can capture an image of the scenery of the real world that includes the person 94 at the remote location. For example, the imager 35B is set apart from the person 94 by a predetermined distance in order to capture the whole body image of the person 94 at the remote location in the field of view.
Further, the imager 35B may be a so-called RGB-D camera having a function of measuring the distance of the subject from the imager 35B in addition to a function of imaging the real world. As the function of measuring the distance, for example, the imager 35B is provided with a distance measuring mechanism using infrared rays, in addition to the above-mentioned imaging device.
The imager 35B is connected to a communication line, and data of the image including the person 94 at the remote location (hereinafter, appropriately referred to as remote location captured image) is transmitted to the display device 30A. In response to this, a display unit 46A of the display device 30A can display a person image 100 (see
Configuration of Imager 35C
The imager 35C is disposed in the complex entertainment facility 10. The imager 35C includes, similar to the imager 35B, an imaging device such as a CMOS imaging device or a CCD imaging device. The imager 35C can capture a still image and a moving image. Further, the imager 35C is provided with a distance measuring mechanism using infrared rays.
The imager 35C can wirelessly communicate with the display device 30B at the remote location via a communication line and a base station of a telecommunications carrier or the like (not shown), and the captured still image data or moving image data is transmitted to the display device 30B.
With reference to
Further, an augmented reality (AR) marker 35C1 that is an imageable identifier is indicated on the imager 35C. The AR marker 35C1 is printed on the surface of the imager 35C. The AR marker 35C1 is an identifier that is used when an augmented reality image is displayed, and indicates, for example, a display area of a virtual image to be superimposed. For example, when the inside of the vehicle 90 is imaged, the image area including the AR marker 35C1 is set as the superimposed area of the virtual image.
In this way, by displaying the AR marker 35C1 that is the identifier in the complex entertainment facility 10 in an imageable state, it is possible to set the superimposed area of the person image 100 of the remote location (see
Configuration of Server
With reference to
The server 70 includes an input unit 71 such as a keyboard and a mouse, a central processing unit (CPU) 72 serving as an arithmetic device, and a display unit 73 such as a display. The server 70 also includes a read-only memory (ROM) 74, a random access memory (RAM) 75, and a hard disk drive (HDD) 76 as storage devices. Further, the server 70 includes an input-output controller 77 that manages input and output of information. These components are connected to an internal bus 78.
The server 70 includes a facility map storage unit 80, a decoration image storage unit 82, and a decoration image extraction unit 85.
The facility map storage unit 80 stores map information of the complex entertainment facility 10. For example, the facility map storage unit 80 stores position information of the passages (connecting passages 20A and in-park passages 20B to 20D) and facilities in the complex entertainment facility 10. Specifically, the facility map storage unit 80 stores plan view data of the complex entertainment facility 10 that is associated with the position information. The position information includes longitude and latitude information using the GPS function and position information using the beacon function.
The decoration image storage unit 82 stores the decoration image data including virtual objects, out of the augmented reality images that are displayed on the display device 30A.
The decoration image data stored in the decoration image storage unit 82 may be three-dimensional (3D) model data of a decoration object that is a virtual object. The 3D model data includes, for example, 3D image data of the decoration object, and the 3D image data includes shape data, texture data, and motion data.
A plurality of kinds of decoration image data is stored in the decoration image storage unit 82 for each of the theme parks 14 to 18. For example, 10 to 100 kinds of decoration image data for one theme park is stored in the decoration image storage unit 82. The decoration image data is individually provided with an identification code of a corresponding theme park, out of the theme parks 14 to 18. Further, a unique identification code is provided to each piece of the decoration image data.
The decoration image data is, for example, a character image defined as a character for a theme park, out of the theme parks 14 to 18. For example, the decoration image 110 provided with the identification code corresponding to the amusement park 14 includes an image of a large ball for ball riding. Further, the decoration image 110 provided with the identification code corresponding to the aquarium 16 includes an image of an arch of a school of fish.
In
The decoration image extraction unit 85 determines which of the theme parks 14 to 18 of the complex entertainment facility 10 the display device 30A is located in, based on the current position information acquired by a position information acquisition unit 50 (see
Configuration of Display Device 30A
With reference to
The display device 30A may be a portable device. For example, the display device 30A is a smartphone provided with an imaging device and a display unit, or a glasses-type head-mounted display (HMD).
The display device 30A can be divided into a video see-through display (VST display) and an optical see-through display (OST display) from a viewpoint of the mode of displaying scenery of the real world. In the VST display, an imager such as a camera captures an image of scenery of the real world, and the captured image is displayed on the display unit. On the other hand, in the OST display, scenery of the real world is visually recognized through a transmissive display unit such as a half mirror, and a virtual object is projected onto the display unit.
The display device 30A provided with an imager 35A (see
In the embodiment below, as shown in
The system memory 40A is a storage device used by an operating system (OS) executed by the CPU 31A. The storage device 41A is an external storage device, and stores, for example, a program for displaying a virtual reality image (AR image), which will be described later.
The imager 35A is, for example, a camera device mounted on a smartphone, and can capture an image of the scenery of the real world as a still image or a moving image. The imager 35A includes an imaging device such as a CMOS imaging device or a CCD imaging device. Further, the imager 35A may be a so-called RGB-D camera having a function of measuring the distance from the imager 35A in addition to a function of imaging the real world. As the function of measuring the distance, for example, the imager 35A is provided with a distance measuring mechanism using infrared rays, in addition to the above-mentioned imaging device.
The GPU 42 is an arithmetic device for image processing, and is mainly operated when image recognition described later is performed. The frame memory 43 is a storage device that stores an image captured by the imager 35A and subjected to computation by the GPU 42. The RAMDAC 44 converts the image data stored in the frame memory 43 into analog signals for the display unit 46A that is an analog display.
The GPS receiver 36 receives GPS signals that are positioning signals from a GPS satellite 24 (see
Here, both the GPS receiver 36 and the beacon receiver 37 have overlapping position estimation functions. Therefore, the display device 30A may be provided with only one of the GPS receiver 36 and the beacon receiver 37.
The input unit 47A can input an activation instruction and an imaging instruction to the imager 35A. For example, the input unit 47A may be a touch panel integrated with the display unit 46A.
The display control unit 45A can generate an augmented reality image (AR image) in which an image of a virtual object is superimposed on scenery of the real world and display the AR image on the display unit 46A. For example, the display control unit 45A superimposes the virtual image on the image area of the AR marker 35C1 (see
Further, as the functional blocks, the display device 30A includes the position information acquisition unit 50 and an image recognition unit 58. The display device 30A includes a learned model storage unit 59 as a storage unit. These functional blocks are composed of the CPU 31A, the system memory 40A, the storage device 41A, the GPU 42, the frame memory 43, and the like.
The position information acquisition unit 50 acquires information on the current position of the position information acquisition unit 50 from at least one of the GPS receiver 36 and the beacon receiver 37 in
The image recognition unit 58 receives the image data captured by the imagers 35A and 35B and performs image recognition. The image recognition includes recognition of objects in the captured image and estimation of the distance of each object from the display device 30A and the imager 35B. In such image recognition, the captured image data includes, for example, a color image data obtained by imaging the scenery of the real world as well as distance data of each object in the color image data from the imagers 35A and 35B, as described above.
The image recognition unit 58 recognizes the captured image using the learned model for image recognition stored in the learned model storage unit 59. The learned model storage unit 59 stores, for example, a neural network for image recognition that has been trained by an external server or the like. For example, outdoor image data containing the complex entertainment facility 10, in which each object in the image has been segmented and annotated, is prepared as training data. In addition, training data for recognizing the person image from the captured image is also prepared. Using the training data, a multi-level neural network is formed that has machine-learned by supervised learning, and is stored in the learned model storage unit 59. This neural network may be, for example, a convolutional neural network (CNN).
As will be described later, by using the learned model, the image recognition unit 58 can recognize and extract the person image from the captured image captured by the imager 35B at the remote location. Further, by using the learned model, the image recognition unit 58 can recognize the AR marker 35C1 that is the identifier displayed in the complex entertainment facility 10.
Augmented Reality Image Display Flow
The display flow illustrated in
In
With reference to
The image recognition unit 58 performs image recognition on the received in-facility captured image (S10). The image recognition includes recognition of the AR marker 35C1 that is the identifier included in the in-facility captured image. The recognition also includes segmentation and annotation of objects in the in-facility captured image. Further, in the image recognition, the distance of each object from the display device 30A is obtained.
The image recognition unit 58 determines whether the AR marker 35C1 is recognized in the captured image (S12). When the AR marker 35C1 is not recognized, the flow ends. On the other hand, when the AR marker 35C1 is recognized in the captured image, the image recognition unit 58 tracks the AR marker 35C1 for a predetermined period (performs so-called marker tracking), and determines whether the AR marker 35C1 is continuously included in the captured image for the predetermined period (S14). The predetermined period may be, for example, five seconds or more and 10 seconds or less.
When the AR marker 35C1 disappears from the captured image during the predetermined period, it is considered to be a so-called unintended reflection, and therefore, generation of the augmented reality image activated by the AR marker 35C1 is not carried out. That is, the display of the augmented reality image on the display unit 46A is suspended. On the other hand, when the AR marker 35C1 is continuously included in the captured image for the predetermined period, the image recognition unit 58 sets the image area of the AR marker 35C1 as the superimposed area of the person image 100 (see
Further, the position information acquisition unit 50 acquires the current position of the display device 30A. This current position information is transmitted to the server 70 (S16). When the server 70 receives the current position information of the display device 30A, the decoration image extraction unit 85 checks which position of the complex entertainment facility 10 the current position of the display device 30A corresponds to, from the park map data stored in the facility map storage unit 80. Further, the decoration image extraction unit 85 determines whether the current position of the display device 30A is included in any of the theme parks 14 to 18 (S18).
When the current position of the display device 30A is not included in any of the theme parks 14 to 18, the decoration image extraction unit 85 notifies the display device 30A that the current position is not included in the theme parks 14 to 18 (S20). For example, when the vehicle 90 (see
Upon receiving the out-of-park notification, the image recognition unit 58 of the display device 30A acquires the data of the captured image of the remote location from the imager 35B at the remote location. Further, the image recognition unit 58 recognizes the person image from the acquired remote location captured image (S22).
The image recognition unit 58 extracts the person image 100 (see
As illustrated in
Returning to step S18, when the decoration image extraction unit 85 of the server 70 determines that the current position of the display device 30A is included in any of the theme parks 14 to 18, the decoration image extraction unit 85 extracts the decoration image 110 (see
For example, the decoration image extraction unit 85 extracts the data of the decoration image 110 with the identification code of the theme park, out of the theme parks 14 to 18, in which the display device 30A is staying. When a plurality of kinds of decoration images is stored for each of the theme parks 14 to 18, an appropriate image is extracted from the decoration images with the identification code of the same theme park.
Further, the decoration image extraction unit 85 transmits the extracted decoration image to the display device 30A (S28). Further, the image recognition unit 58 of the display device 30A recognizes the person image from the captured image of the remote location that is acquired from the imager 35B at the remote location and extracts the recognized image (S30). The display control unit 45A of the display device 30A executes a decoration process of superimposing the decoration image 110 acquired from the decoration image extraction unit 85 on the person image 100 of the remote location that is acquired from the image recognition unit 58 (S32).
For example, in this decoration process, the display control unit 45A associates body parts of the person image 100 with body parts of the decoration image. For example, when the decoration image 110 is an image of the head of an animal, the image recognition unit 58 estimates the head of the person image 100 by image recognition of the person image 100. Further, the display control unit 45A defines the head region of the person image 100 as the superimposed region of the decoration image 110.
As illustrated in
Other Example of Display Device
In the above-described embodiment, the display device 30A is exemplified by a smartphone that is a video see-through display. However, the display device 30A according to the present embodiment is not limited to this form. For example, as is the head-mounted display (HMD) as illustrated in
In this case, the display device 30A includes the imager 35A, a half mirror 114 corresponding to the display unit 46A, a projector 116 corresponding to the display control unit 45A and the image recognition unit 58, and a sensor unit 112 corresponding to the position information acquisition unit 50.
The half mirror 114 may be, for example, the lenses of eyeglasses or goggles. The half mirror 114 allows light (image) from the real world to be transmitted to the user. The projector 116 disposed above the half mirror 114 projects an image of the virtual object onto the half mirror 114. Thus, it is possible to display an augmented reality image in which the person image 100 of the remote location and the person image 100 on which the decoration image 110 is superimposed are superimposed on the scenery in the complex entertainment facility 10 that is the scenery of the real world.
Other Example of Display Device
In the above-described embodiment, the augmented reality image display flow of
In
Other Example of Identifier
In the above-described embodiment, the AR marker 35C1 is provided to the surface of the imager 35C as the identifier for the display device 30A to generate an augmented reality image, but the display system according to the present embodiment is not limited to this form. For example, a so-called markerless AR method in which the AR marker 35C1 is not provided to the imager 35C (see
Number | Date | Country | Kind |
---|---|---|---|
2020-211166 | Dec 2020 | JP | national |