This application claims the benefit of Korean Patent Application No. 10-2011-0057559, filed on Jun. 14, 2011, which is hereby incorporated by references as if fully set forth herein.
The present invention relates to a technology of exhibiting mixed reality, and more particularly, an apparatus and method for exhibiting mixed reality based on a print medium, which provides the integration of virtual digital contents and the print medium in reality.
As well-known in the art, many researches are ongoing on a technology for augmenting information in the real world by adding virtual contents to an object in the real world. Technical development has been achieved by various approaches, starting from a virtual reality technology of mainly representing virtual reality to an augmented reality technology of adding virtual information based on the real world and a mixed reality technology of attempting to appropriately mix reality with virtual reality.
Especially, as terminals, such as smart phones, having an improved computing capability and a camera function, are widely used, a mobile augmented reality (AR) technology has been on the rise. The mobile AR technology is providing various services, such as adding virtual information required by a user to an ambient environment during movement. However, most of mobile AR technologies merely provide both an actual image and virtual information through a display device mounted in a terminal. Thus, the user still feels imaginarily the virtual information existing within the terminal, and an input method for providing the virtual information is still performed through the general operation in the terminal.
In addition, with the introduction of user equipment, such as a mobile phone having a small projector attached thereto, an attempt to use the projector as a new display device is being made. This is also utilized as a service to provide a large screen, without limit to a small display screen, for allowing many persons to watch a movie and share information, and the like.
A new service concept using the output function of the projector and the input function of the camera has been introduced, as Sixth Sense, by Massachusetts Institute of Technology (MIT). According to this concept, user's hand gestures are input as camera images for use and information as a new display or a part of an actual object is added to an image projected through the projector, such that digital information that is integrated with information about the actual object can be provided to the user as if they are originally one information. For example, when a user views a paper with a picture printed thereon, the user can view not only the printed picture of the paper but also a video of the picture through an image being projected in real time. In addition, changed flight information is additionally exhibited on printed flight information within the ticket, thereby expressing virtuality of the digital information to be more realistic.
Due to recent development of technologies, a projector and a camera which are reduced in size are mounted in a mobile device. Thus, a system for providing various services by fabricating the small projector and camera in a wearable form is being introduced, and also a system for allowing the small projector and camera to be usable during movement by fabricating them in a portable form is being developed. The use of those systems enables digital information to be exhibited or displayed on a real-world object other than a screen of a digital terminal, and also allows for creation of new services. However, the portable type system as introduced above has a limitation of concentrating on exhibiting digital information and direct user interactions by using a projected region itself as a new display area, rather than creation of new contents through integration between information provided by an actual object and virtual information.
Further, the wearable type system such as the Sixth Sense is employing a method of attaching markers with specific colors onto user's fingers and attaching a separate sensor onto an actual object, which may lower practical utilization.
In view of the above, the present invention provides an apparatus and method for exhibiting mixed reality based on a print medium, which provides the integration of virtual digital contents and the print medium in reality.
Further, the present invention provides an apparatus and method for exhibiting mixed reality based on a print medium, which provides a space for digital contents exhibition and a space for a user's input command within an actual reality space to allow an intuitive user input command.
Further, the present invention also provides an apparatus and method for exhibiting mixed reality based on a print medium, which are capable of allowing recognition of a user's input command and an output of digital contents without a separate marker.
In accordance with an aspect of the present invention, there is provided an apparatus for exhibiting mixed reality based on a print medium, which includes: a command identification module configured to identify a hand gesture of a user performed on a printer matter in the print medium to recognize a user input command corresponding to the hand gesture; and a content reproduction module configured to provide a digital content corresponding to the printed matter onto a display area on the print medium.
Preferably, the command identification module includes: a pattern image output unit configured to generate a pattern image on the printed medium, the hand gesture causes a change in the pattern image; an image acquiring unit configured to capture an image of the surface of the printed medium, wherein the captured image has the pattern image included therein, and wherein the hand gesture causes the change in the pattern image; and a command recognizing unit configured to detect the pattern image caused by the hand gesture to recognize the user input command corresponding to the hand gesture.
Preferably, the pattern image includes an image in a grid form projected onto the print medium at a preset period.
Preferably, the pattern image includes an infrared image invisible to the user.
Preferably, the command identification module further includes a command model database that stores a plurality of command models corresponding to hand gestures representative of user input commands; and wherein the command recognition unit is configured to match the hand gesture with the command models to find out a command model corresponding to the hand gesture.
Preferably, the command identification module further comprises an environment recognizing unit that is configured to analyze the captured image of the print medium to find the display area appropriate for presenting the digital content.
Preferably, the environment recognizing unit is further configured to collect display environment information from the captured image of the print medium, the display environment information including at least one of information related to size, brightness, flat state or distorted state of the display area.
Preferably, the apparatus further includes a content management module that is configured to format the digital content based on the display environment information of the display area and provide the digital content to the content reproduction module.
Preferably, the content reproduction module includes an image correction unit that is configured to correct the image of the digital content based on the display environment information.
In accordance with another aspect of the present invention, there is provided a method for exhibiting mixed reality based on a print medium, which includes: generating a pattern image onto the print medium, wherein a hand gesture of a user is interacted with a printed matter in the printed medium to produce a user input command; identifying the hand gesture causing a change in the pattern image to recognize the user input command; and projecting digital content corresponding to the printed matter onto a display area of the print medium depending on the user input command.
Preferably, the generating a pattern image onto the print medium includes projecting an image in a grid form onto the print medium at a preset period.
Preferably, the identifying the hand gesture includes: capturing an image of the print medium, the captured image including the pattern image; detecting the change in the pattern image caused by the hand gesture; and
matching the hand gesture with a plurality of command models to find out a command model corresponding to the user input command.
Preferably, the pattern image includes an infrared image invisible to the user.
Preferably, the method further includes: analyzing the captured image to find a display area appropriate for reproducing the digital contents on the print medium.
Preferably, the method further includes: collecting display environment information including at least one of information related to size, brightness, flat state and distorted state of the display area.
Preferably, the method further includes: formatting the digital content based on the collected display environment information.
Preferably, the method further includes: correcting an image of the digital content reproduced on the display area based on the collected display environment information.
The above and other objects and features of the present invention will become apparent from the following description of embodiments, given in conjunction with the accompanying drawings, in which:
Hereinafter, embodiments of the present invention will be described in detail with the accompanying drawings.
As shown in
In the embodiment, the print medium may includes such as a story book, an illustrated book, a magazine, an English language teaching material, an encyclopedia, a paper or the like. The printed medium has printed matters thereon such as printed words, printed pictures or images, or the like. When a user interacts with a printed matter on the printed medium or an image projected onto the printed medium using the hand gesture, virtual digital content corresponding to the printed matter may be reproduced or represented onto a certain area on the printed medium in real world. The command identification module 100 includes an environment recognizing unit 110, a pattern image output unit 120, an image acquiring unit 125, a command recognizing unit 130, and a command model database 140.
The pattern image output unit 120 projects a pattern image on the surface of the print medium at a preset period or in a consecutive manner. The pattern image projected onto the surface of the print medium has the form of stripe patterns or the form of a grid pattern as shown in
It is preferable that the pattern images are invisible to a user not to interfere with the visibility of the printed matters in the printed medium for which the projected pattern will be confusing. Hence, there may be a limitation on the number of pattern images capable of being projected onto the printed medium per unit time.
Therefore, the pattern image output unit 320 may be implemented with a structured-light 3D scanner which projects a specific pattern of infrared light onto the surface of the print medium or a diffraction grating which forms specific patterns of infrared light by means of diffraction of laser beams. The infrared pattern image is invisible to a user, and therefore the number of pattern images capable of being inserted per time may rarely be limited. Further, if it is necessary to project many pattern images in order to increase a performance of identifying the respective hand gestures, the use of an extremely high frame rate pattern image may satisfy the need.
The image acquiring unit 125 captures an image of the surface of the print medium depending on a preset period at which the pattern image output unit 120 projects the pattern image. The captured image includes the pattern image on which a hand gesture of a user is performed on the printed matter in the print medium. For an infrared pattern image, the image acquiring unit 125 may be implemented as an infrared camera for capturing an infrared pattern image projected onto the print medium. The captured image is then provided to the environment recognizing unit 110 and the command recognizing unit 130.
The environment recognizing unit 130 analyzes the captured image of the print medium to find a display area for presenting digital content corresponding to the printed matter exerted by the hand gesture on the print medium. The environment recognizing unit 130 also collects display environment information including at least one or all of information relating to size, brightness, a flat state or a distorted state related to the display area. That is, the environment recognizing unit 110 collects in advance required display environment-related information, such as whether or not the display area is flat or whether or not the display area is distorted, for presenting digital content in reality through a projection.
The command model database 140 stores a plurality of command models corresponding to the hand gestures representative of the user's input commands.
When the print matter in the print medium, onto which the pattern image is projected, is touched by the hand gesture, it may cause a change in the pattern image. The command recognizing unit 130 detects the change in the pattern image by the hand gesture to recognize the input of a user's command corresponding to the hand gesture. More specifically, when the command recognizing unit 130 detects the change in the pattern image, it matches the hand gesture with the command models to find out a command model corresponding to the hand gesture, which becomes the user's input command. The hand gesture may include, for example, underlining on a word included in the print medium on which the pattern image has been projected or pointing vertexes of a picture included in the print medium with a finger, which will be discussed with reference
As shown in
As shown in
The content management module 200 controls selection, creation, modification and the like of the digital content corresponding to the printed matter in the print medium depending on the user input command recognized by the command identification module 100. The content creation module 200 includes a content creation unit 210 and a content database 220.
The content creation unit 210 reconstructs the digital content corresponding to the printed matter based on the display environment information collected by the command identification module 100. The digital content to be displayed on the display area in the print medium may fetched from the local content database 220 or provided from an external server 250 via a network. The digital content provided from the local content database 220 or the external server 250 may have a structure which is improper to the display environment. In this case, the content creation unit 210 may modify, format or reconstruct the digital content to be compatible with the display environment, such as the size of the display area or the like.
The content database 220 stores user interfaces that frequently used by the user and digital content to be displayed on the display area in the print medium.
The content reproduction module 300 projects the digital content onto the display area in the print medium. The content reproduction module 300 includes a content output unit 310 and an image correction unit 320. The content output unit 310 projects the digital content provided by the content management module 200 onto the display area of the print medium. For example, the content output unit 310 is implemented as a projector, which projects digital content onto the display area in the print medium in reality to reproduce the images of the digital content. In addition, the content output unit 310 may adjust a focus of the projector, a projection direction of the projector and the like to avoid a visibility-related problem when projecting the digital content onto the display area. The image correction unit 320 corrects the images of the digital content projected by the content output unit 310 based on the display environment information. Color and brightness of the image of the digital content may be changed depending on the display environment information. The image correction unit 320 corrects the image of the digital content to be actually displayed in advance because exhibition of the color or brightness of the image of the digital content may actually change depending on features of the display area of the print medium. Further, when the display area on which the image of the digital content is projected is not flat, distortion in the image of the digital content may be caused. Hence, the image correction unit 320 corrects the image of the digital content to be projected in advance by performing geometric correction of the image.
As shown in
Optionally, the apparatus may be configured such that the projector 312 inserts or overlaps a pattern image directly into or with an image of the digital content projected by the projector 312. In this case, the scanner 314 may be omitted from the apparatus for exhibiting mixed reality based on a print medium of the embodiment of the present invention.
On the right part in
First, in step S401, the pattern image output unit 110 projects a pattern image such as the grid image 350 onto a surface of a print medium 370 as shown in
A user may then issues a user input command by taking a specific gesture on a printed matter in the print medium with a finger as described with reference to
Then, the environment recognizing unit 110 analyzes the captured image for the print medium to find a display area appropriated for exhibiting digital content corresponding to a printed matter such as word, picture, image, etc. selected by the hand gesture, and collect display environment information including at least one or all of information about size, brightness, flat state or distorted state of the display area. For example, the environment recognizing unit 110 identifies color distribution in the captured image, and as shown in
The command recognizing unit 125 detects the change in the pattern image and matches the hand gesture with the command models stored in the command model database 140, thereby recognizing a user input command based on the matching result in step S405. Such the hand gesture corresponding to the user input command may be any one of the hand gestures shown in
When the user input command and the environment information are recognized through step S405, the content output unit 310 obtains digital content corresponding to the selected printed matter from the content creation module 200 in step S407.
The image correction unit 110 then reconstructs or formats the digital content based on the display environment information in step S409. For example, the image correction unit 320 changes colors and brightness of the digital content to be provided by the content output unit 310 based on the display environment information. Colors and/or brightness desired to be actually given may be differently reproduced depending on features of the display area on which the digital content is projected. Thus, the image correction unit 320 corrects such colors and/or brightness in advance. Also, when the display area to be projected is not flat, image distortion may be caused. Hence, it is compensated in advance by a geometric correction of the image of the digital content.
Next, the content output unit 310 controls the output of the digital content in step S411 to exhibit the digital content 730 on the display area 720 in the print medium 710 as shown in
The method for exhibiting mixed reality based on a print medium in accordance with the embodiment of the present invention as described above may be recorded with a computer program. Codes and code segments constituting the computer program may be easily inferred by a programmer in the art. Further, the computer program may be stored in a computer-readable storage medium that can be read by a computer, and read and executed by a computer, the apparatus for exhibiting mixed reality based on a print medium in accordance with the embodiment of the present invention, or the like, thereby implementing the method for exhibiting mixed reality based on a print medium. The computer-readable storage medium includes a magnetic recording medium, an optical recording medium, and a carrier wave medium.
In accordance with the embodiment of the present invention, a printed matter on a print medium and a virtual digital content may be integrated with each other, so as to be displayed on a display area on the print medium in the real world, thus allowing for an intuitive user input. Further, recognition of a hand gesture of a user and a reproduction of the virtual digital content may be utilized without a separate marker or sensing device.
Thus, the mixed reality exhibiting apparatus in accordance with the embodiment may be used in mobile equipment as well as the existing projector system. The virtual digital content may be exhibited directly onto the printed matter in the real world, which may provide a user with a new experience, increase utilization of a real-world object such as a print medium and digital content, and enhance reuse of content.
In addition, the integration of reality information and virtual information with a real-world medium may allow for correspondence of an information exhibition space. Also, a user interaction may be performed between virtual digital information and a printed matter of the real-world medium, thereby allowing for correspondence with a user input space. Moreover, use of a simplified effective input/output method which can be actually used as well as being conceptually designed may result in improvement of user convenience.
While the invention has been shown and described with respect to the particular embodiments, the present invention is not limited thereto. It will be understood by those skilled in the art that various changes and modification may be made.
Number | Date | Country | Kind |
---|---|---|---|
10-2011-0057559 | Jun 2011 | KR | national |