The present invention relates to an information processing technology for processing a picked up image.
Various technologies for selecting a content in response to a situation at present of a user and automatically reproducing the content have been proposed heretofore. For example, PTL 1 discloses a reproduction apparatus which determines a psychological state of a user from biological information of the user and selects music on the basis of the psychological state.
Meanwhile, attention is paid to a technology called augmented reality (AR) which presents additional information in a superposed relationship on a real space to a user, and a head mounted display unit (hereinafter referred to as “HMD”) which can present a stereoscopic image having a depth has become popular. Although various types are available for the HMD, one of the HMDs is an optical transmission type HMD which uses a holographic device, a half mirror or the like to present a virtual stereoscopic image to a user and allow the user to view a manner outside the HMD in a see-through fashion through the HMD.
[PTL 1] Japanese Patent Laid-Open No. 2006-146630
Although PTL 1 adopts a fresh approach of utilizing a psychological state of a user in order to select music to be reproduced, actually there is a situation that it is not easy to determine a psychological state of a user with a high degree of accuracy. Therefore, it is desired to develop a technology for appropriately deciding a situation of a user at present and providing a service at a suitable timing to the user. The inventor of the present invention has found out the possibility that, especially by incorporating such a technology as described above into a wearable computer, for example, into a computer including an optical transmission type HMD, a situation of a user at present can appropriately decide and a suitable service may be provided. Further, the inventor of the present invention has conceived a user interface which can be handled readily by a user by utilizing a characteristic of a mounted type display unit such as a HMD.
The present invention has been made in view of such a subject as described above, and it is an object of the present invention to provide an information processing technology which can provide a service suitably to a user and a user interface application which can be handled readily by a user.
In order to solve the subject described above, an information processing apparatus according to an aspect of the present invention includes a recording unit configured to record an event list in which time information and physical solid image information are associated with an event, an image pickup unit configured to pick up an image of a real space, a control section configured to determine whether or not an event starting condition is satisfied, and an execution unit configured to process an application. The control section includes an image processing portion configured to determine whether an image corresponding to the physical solid image information recorded in the recording unit is included in a picked up image picked up within a time zone specified by the time information recorded in the recording unit, a condition determination portion configured to determine that the event starting condition is satisfied if it is determined that the image corresponding to the physical solid image information is included in the picked up image, and an instruction portion configured to instruct the execution unit to execute processing of the application. The execution unit starts processing of the application associated with the event whose starting condition is satisfied.
Another aspect of the present invention is an application execution method. The method includes a step of acquiring a picked up image obtained by picking up a real space, a step of referring to an event list in which time information and physical solid image information are associated with an event to determine whether an image corresponding to the physical solid image information included in the event list is included in the picked up image picked up within a time zone specified by the time information included in the event list, a step of determining, when it is determined that an image corresponding to the physical solid image information is included in the picked up image, that an event starting condition is satisfied, and a step of starting processing of an application associated with the event whose starting condition is satisfied.
A further aspect of the present invention is an information processing apparatus. This apparatus is an information processing apparatus which presents a virtual object in a superposed relationship with a real space and includes a mounted type display unit configured to display the virtual object so as to be observed by a user in the real space, an image pickup unit configured to pick up an image of the real space, an image processing section configured to specify a real physical solid image included in the picked up image, a recording unit configured to record the specified real physical solid image or a mark image corresponding to the real physical solid image as a physical solid image, and a display processing section configured to display the physical solid image on the mounted type display unit.
It is to be noted that also arbitrary combinations of the components described above and those obtained by converting representations of the present invention between methods, apparatus, systems, computer programs, data structures, recording media and so forth are effectively applied as different modes of the present invention.
With the present invention, an information processing technology which can provide a service suitably to a user and a user interface application which can be handled readily by a user can be provided.
Information presented to a user in the augmented reality (AR) technology is called annotation and is visualized using virtual objects of various forms such as a text, an icon, or an animation. However, where the AR technology is utilized for advertisement, a virtual object for advertisement is displayed in a superposed relationship on a wall face of a building, a signboard or the like in a real space. Meanwhile, where the AR technology is used for road guide, a virtual object, a map or the like for road guide is displayed in a superposed relationship on a physical solid or on a road which serves as a mark in the real world.
In such an AR application as described above, a marker is registered first, and then an image of a real physical solid corresponding to the marker is picked up to start up a process (service) associated with the marker. The inventor of the present invention has paid attention to the fact that, if a user can individually register a marker, then the application can be customized and processed by the user, and has come to develop a user interface for allowing a user to register a marker readily and simply. Further, the inventor of the present invention has paid attention also to the fact that, by providing, after a registered marker and a real physical solid image are compared with each other, an application of a different type to the user in addition to a visual annotation, the possibility of the AR technology can be expanded. In the following, the present invention is described in connection with an embodiment thereof.
In the present embodiment, the display apparatus 12 is an optical transmission type HMD configured from a half mirror, and the user can view the real space in a see-through fashion through the display apparatus 12 and can further view a video (virtual object) created by the information processing apparatus 10. The information processing apparatus 10 may otherwise create a stereoscopic image. The display apparatus 12 may be a display apparatus which uses a holographic element to project a video on a light guide plate or a display apparatus of the projection type which forms a virtual image so as to allow a video to be viewed. It is to be noted that the display apparatus 12 may otherwise be a video transmission type HMD and may display, while it displays a real space image picked up by the image pickup unit 14, a virtual object created by the information processing apparatus 10 in a superposed relationship with the real space image. In this manner, the information processing apparatus 10 may be an information processing apparatus of the mounted type which presents the real space as an environment of the user to the user and displays a virtual object in a peripheral environment of the real space.
In the present embodiment, the “real physical solid” signifies a substance which exists in the real space, and the “virtual object” signifies an object created by the information processing apparatus 10. Where the display apparatus 12 is an optical transmission type HMD, a real physical solid can be observed in a see-through fashion through the optical transmission type HMD by the user. It is to be noted that, where an image of a real physical solid is cut out from a picked up image and the information processing apparatus 10 displays the cutout image at an arbitrary position of the display apparatus 12, the real physical solid image is handled as virtual object.
The display apparatus 12 has a form of a pair of glasses, and an image for the right eye is displayed on the right glass while an image for the left eye is displayed on the left glass. Consequently, the user can observe a stereoscopic image. It is to be noted that the display apparatus 12 in the embodiment may not necessarily have a mechanism for providing a stereoscopic image but may have only one glass for one eye.
The image pickup unit 14 is provided between the two eyeglass type display members so as to be disposed in the middle of the forehead when the information processing apparatus 10 is mounted on the user and picks up an image of the real space included in the field of view of the user. Although the angle of view of the image pickup unit 14 preferably coincides with or equivalent to the angle of view of a human being, the angle of view of the image pickup unit 14 in the information processing apparatus 10 of the type depicted in
The housing 18 plays a role of a glasses frame of the information processing apparatus 10 in the form of glasses and accommodates therein various modules used by the information processing apparatus 10. The modules used by the information processing apparatus 10 include a module for implementing an optical transmission type HMD, a module for creating a three-dimensional video (virtual object), a communication module for carrying out a communication process by the Bluetooth (registered trademark) protocol, the IEEE802.11 protocol, or the mobile communication protocol, a module for sound outputting, an electronic compass, an acceleration sensor, an inclination sensor, a global positioning system (GPS) sensor, an illuminance sensor and so forth. The modules mentioned are exemplary, and the information processing apparatus 10 need not necessarily include all of the modules mentioned. Which one or ones of the modules are to be incorporated may be determined in response to scenes assumed to be used by the information processing apparatus 10.
While the information processing apparatus 10 depicted in
In the present embodiment, the information processing apparatus 10 provides a user interface which can be handled readily by the user and has a function of providing a content in response to a behavior of the user. It is to be noted that a content provided to the user may be that retained in a recording unit of the information processing apparatus 10 or may be distributed from an external apparatus such as a content server or a terminal apparatus. In the following, an environment in which a content is distributed from an external apparatus to the information processing apparatus 10 is described.
An outline of functions of the information processing apparatus 10 is described. It is to be noted that the following description is given taking a case in which the display apparatus 12 is an optical transmission type HMD with which the user can view the real space in a see-through fashion as an example. The information processing apparatus 10 incorporates an AR application and has a function of displaying a virtual object on the display apparatus 12.
The information processing apparatus 10 can process an application in response to the situation of the user. Here, the process of an application includes to start up and execute an application in response to the situation of the user and to execute a content in response to the situation of the user in a particular application. For example, the former includes to start up and execute an application for the notification of congestion information of roads, delay information of trains, today's weather forecast or the like before the user goes to work in the morning, to start up and execute an application for reproducing music while the user is going to work after leaving the own home and so forth. Meanwhile, the latter includes, for example, to reproduce music for good awakening before the user goes to work in the morning, streaming reproduce a radio program of a foreign language for study of the foreign language when the user is going to work and so forth. The information processing apparatus 10 of the embodiment retains an event list which associates events, time information, and physical solid image information with each other, and decides, if an image corresponding to the physical solid image information is included in a real space image picked up by the image pickup unit 14 within a time zone specified by the time information, that a starting condition for the event is satisfied and then starts processing of an application associated with the event.
In order to implement the functions described above in a high efficiency, the information processing apparatus 10 provides a user interface for creating a schedule table for one day of the user. Generally, in a schedule table, start time and end time of an event are set in an associated relationship with the event. However, in the schedule table in the present embodiment, physical solid image information for determining establishment of an event starting condition is registered.
Although the event list may be same as the schedule table created by the user, preferably it is created by working the schedule table as hereinafter described in detail. The information processing apparatus 10 which is a video observation device includes the image pickup unit 14. If the image pickup unit 14 picks up an image of a real physical solid corresponding to physical solid image information registered in the event list, then it is decided that a starting condition for the event associated with the physical solid image information is satisfied.
In the following, a user interface for implementing a basic operation for making it possible for the user to create a schedule table simply and easily is described.
The input acceptance unit 100 accepts information inputted from the image pickup unit 14 or the microphone 20. From the microphone 20, sound information is inputted, and the input acceptance unit 100 includes a speech recognition function and recognizes speech from the user to create operation information. The created operation information is provided to processing modules such as the display processing section 120, the image processing section 110, and/or the schedule registration section 150.
Further, if the input acceptance unit 100 receives picked up images picked up by the image pickup unit 14, then it detects a gesture of the user included in the picked up images, creates operation information from the gesture and provides the created operation information to processing modules such as the display processing section 120, the image processing section 110, and/or the schedule registration section 150. In the present embodiment, since the image pickup unit 14 is disposed in the proximity of the middle of the forehead of the user, a movement of a physical solid whose gesture is recognized by the input acceptance unit 100 is a movement of a hand or a movement of a finger in the proximity of the direction of the line of sight of the user. It is to be noted that the input acceptance unit 100 may register a physical solid of a predetermined shape as a gesture recognition physical solid such that, when the user moves the physical solid within the angle of view of the image pickup unit 14, the input acceptance unit 100 detects a gesture. In the following, an example wherein the input acceptance unit 100 detects a movement of a hand or a finger of the user who is a gesture recognition physical solid to create operation information is described. The input acceptance unit 100 starts the gesture recognition process when it detects a start of a gesture inputting mode and ends the gesture recognition process when it detects an end of the gesture inputting mode.
If the input acceptance unit 100 recognizes a movement of a hand in the gesture recognition region 30 as described above and the movement satisfies the starting condition for a gesture inputting mode, then the input acceptance unit 100 starts a gesture recognition process. It is to be noted that the following description is given under the assumption that the “hand” includes the entire hand or a finger or fingers of the hand.
After the gesture inputting mode is started, a menu screen image is displayed in the AR display region of the display apparatus 12. Consequently, the user is allowed to view the menu screen image displayed on the display apparatus 12 while viewing the real space. It is to be noted that a virtual object displayed in the AR display region such as the menu screen image is positioned at a corresponding position in the gesture recognition region 30 in the real space.
A relationship between the AR display region of the display apparatus 12 and the gesture recognition region 30 of the picked up image is described. The AR display region is a region in which the user views a virtual image in a superposed relationship in the real space which spreads as the background of the display apparatus 12. The user would carry out an inputting operation using a hand or a finger to the displayed virtual object. Since this inputting operation must be imaged by the image pickup unit 14, the virtual object must be virtually disposed in the angle of view of the image pickup unit 14 and further must be virtually disposed in the gesture recognition region 30. Therefore, while the AR display region of the display apparatus 12 and the gesture recognition region 30 of the picked up image have spatial coordinate systems independent of each other, they must be superposed with each other and preferably are the same image region. The following description is given under the assumption that the AR display region and the gesture recognition region 30 have the same region in the real space.
The input acceptance unit 100 determines that a click gesture is carried out when a hand moves back and forth in the direction of the optical axis of the image pickup unit 14, namely, in the depthwise direction. Where the image pickup unit 14 is configured from a stereo camera, the input acceptance unit 100 can detect a movement of a hand in the depthwise direction from a parallax amount between the images of the hand. On the other hand, where the image pickup unit 14 is configured from a monocular camera, it is possible to detect a variation of the magnitude of the hand to detect a movement of the hand in the depthwise direction. It is to be noted that the information processing apparatus 10 may include a distance sensor to detect a movement of the hand in the depthwise direction from measurement values of the distance sensor. If the input acceptance unit 100 detects a click gesture successively twice, then it determines that a double click operation is carried out.
If the input acceptance unit 100 detects a click gesture in the gesture recognition region 30, then it transmits information representing that a click operation is carried out to the processing modules such as the display processing section 120, the image processing section 110, and/or the schedule registration section 150 together with two-dimensional AR display coordinate values at which the click operation is carried out. The input acceptance unit 100 converts the two-dimensional coordinate values at which the click gesture is carried out in the gesture recognition region 30 into two-dimensional AR display coordinate values. The input acceptance unit 100 has a function of converting two-dimensional coordinate values on a picked up image into two-dimensional AR display coordinate values in this manner and transmitting the two-dimensional AR display coordinate values to the processing modules. In the following description, such conversion and transmission of the coordinate values are omitted.
When a hand moves vertically, horizontally, and obliquely on an AR coordinate system, the input acceptance unit 100 determines that a drag gesture is carried out. If the input acceptance unit 100 detects a drag gesture in the gesture recognition region 30, then it transmits information representing that a drag operation is carried out to the processing modules such as the display processing section 120, the image processing section 110, and/or the schedule registration section 150. It is to be noted that the drag operation is carried out for a virtual object selected by a click operation. Accordingly, the input acceptance unit 100 specifies a virtual object for which a click gesture is carried out and then transmits, when a drag gesture is carried out for the virtual object, drag operation information for the virtual object to a predetermined processing module or modules. It is to be noted that the input acceptance unit 100 recognizes an end of a drag operation when a click gesture is carried out for the virtual object.
The input acceptance unit 100 decides that a zoom-in gesture is carried out when the distance between two fingers is expanded on the AR coordinate system. If the input acceptance unit 100 detects a zoom-in gesture in the gesture recognition region 30, then it transmits information representing that a zoom-in operation is carried out to the processing modules such as the display processing section 120, the image processing section 110, and/or the schedule registration section 150. It is to be noted that the zoom-in operation is carried out for a virtual object selected by a click operation. Accordingly, the input acceptance unit 100 specifies a virtual object for which a click gesture is carried out and transmits, when a zoom-in gesture is carried out for the virtual object, zoom-in operation information for the virtual object to the predetermined module or modules. It is to be noted that the input acceptance unit 100 recognizes an end of the zoom-in operation when a click gesture for the virtual object is carried out.
The input acceptance unit 100 decides that a zoom-out gesture is carried out when the distance between two fingers is reduced on the AR coordinate system. If the input acceptance unit 100 detects a zoom-out gesture in the gesture recognition region 30, then it transmits information representing that a zoom-out operation is carried out to the processing modules such as the display processing section 120, the image processing section 110, and/or the schedule registration section 150. It is to be noted that the zoom-out operation is carried out for a virtual object selected by a click operation. Accordingly, the input acceptance unit 100 specifies a virtual object for which a click gesture is carried out and transmits, when a zoom-out gesture is carried out for the virtual object, zoom-out operation information for the virtual object to the predetermined module or modules. It is to be noted that the input acceptance unit 100 recognizes an end of the zoom-out operation when a click gesture for the virtual object is carried out.
If the gesture inputting mode is ended, then the menu screen image or the like displayed in the AR display region of the display apparatus 12 disappears.
It is to be noted that, as hereinafter described, in the interface in the present embodiment, by selecting an item included in the menu screen image, a menu screen image of a lower layer is opened, and a process relating to an item included in the lower layer menu screen image is executed. In this manner, various processes are carried out successively. Therefore, when an ending condition for the gesture inputting mode relating to
Further, since the ending condition relating to
In this manner, the information processing apparatus 10 provides a user interface for handling a gesture of the user as an operation input. It is to be noted that the input acceptance unit 100 may detect a gesture other than the examples described above, for example, a turning gesture of a hand to create turning operation information of a virtual object. In this manner, the input acceptance unit 100 can detect a movement of a gesture recognition object to create operation information of the user. It is to be noted that, as described hereinabove, the input acceptance unit 100 may analyze the voice inputted from the microphone 20 to create operation information of the user. For example, if the user utters “click,” then the input acceptance unit 100 accepts the utterance information from the microphone 20 and creates and transmits click operation information to the predetermined processing module or modules.
The information processing apparatus 10 of the present embodiment creates an event list from a schedule table in which a schedule of events within one day of the user is set. Although details of the event list are hereinafter described, if the information processing apparatus 10 determines establishment of a starting condition for an event on the basis of the event list, then the information processing apparatus 10 carries out processing of an application associated with the event. The starting condition for the event is that, within a time zone of the event registered in the event list, an image of a real physical solid corresponding to physical solid image information associated with the event is picked up by the image pickup unit 14. In order to allow such processing, the event list is configured by associating time information and physical solid image information with events.
In order to create an event list, also a schedule table on which the event list is based is set by the user by associating time information and physical solid image information with events. In the following, an example of a procedure for creating such a schedule table as just described is described. In the following example, the user would first collect physical solid images for being associated with events, and then carries out, when a schedule table is to be created, a work of specifying the time information of the events and associating the events and the collected physical solid image information with each other.
If the input acceptance unit 100 detects a gesture depicted in
If the user moves a finger to a position at which it overlaps with the menu screen image 32 displayed on the display apparatus 12 and then carries out a click gesture twice in a region on the space in which “schedule application” is displayed, then the input acceptance unit 100 sends coordinate values of the AR display region of the display apparatus 12 and double click operation information to the display processing section 120. If the menu displaying portion 122 detects from the coordinate values of the AR display region that the menu item “schedule application” is selected, then it displays a menu screen image of a lower layer associated with the “schedule application” in the AR display region.
If the user carries out a double click gesture on the displaying region of the “registration of event starting object” on the menu screen image 34 displayed on the display apparatus 12, then the input acceptance unit 100 sends double click operation information and the AR display region coordinate values to the image processing section 110 and the display processing section 120. It is to be noted that “on the displaying region” signifies on a display region in the virtual space represented by the display apparatus 12. If the menu displaying portion 122 detects that the menu item “registration of event starting object” is selected, then it displays a menu screen image of a lower layer associated with the item on the display apparatus 12, and the image processing section 110 starts a process for specifying a real physical solid image included in the picked up image.
If the user carries out a double click gesture on the display region of the “freehand rendering,” then an inputting mode by freehand is started. If the user operates, in the inputting mode, a finger in such a manner as to surround a physical solid to be cut out, then a locus displaying portion 130 displays the locus of the finger as a virtual object on the display apparatus 12 in such a manner as to follow up the movement of the finger. A start and an end of inputting of “freehand rendering” are determined by click gestures of the user. Accordingly, if the user performs a click gesture once at a start point of a free curve and then performs a click gesture after the finger is moved to surround the physical solid, then the free curve is settled.
After the real physical solid specification portion 112 extracts a real physical solid image, it may color the outer profile of the real physical solid image or the overall real physical solid image to notify the user that the extraction process is completed. If the user confirms that the extraction process is completed appropriately, then the user would carry out a double click gesture on the display region of “registration” of the menu screen image 36. Then, the input acceptance unit 100 sends double click operation information to the image processing section 110, and a physical solid image determination portion 114 records the extracted real physical solid image as a physical solid image 144 into the recording unit 140. The physical solid image 144 is used as basic information for determining satisfaction of an event starting condition as hereinafter described.
The physical solid image determination portion 114 may record a mark image corresponding to the extracted real physical solid image as the physical solid image 144 into the recording unit 140. In the recording unit 140, a mark data 142 including a mark image corresponding to a physical solid is recorded. The mark data 142 includes a plurality of mark images prepared for the physical solid. Here, the physical solid is the clock, and the mark data 142 includes mark images of various types of clocks. Table clocks include various variations such as, in terms of the shape, vertically elongated clocks, horizontally elongated clocks, and round clocks, in terms of the display method, analog clocks and digital clocks, or in terms of the color, blue clocks, black clocks and so forth. The mark data 142 is prepared so as to cover all of the variations. The physical solid image determination portion 114 may extract a mark image same as or similar to a real physical solid image extracted by the real physical solid specification portion 112 from the mark data 142 and record the extracted mark image as the physical solid image 144 into the recording unit 140.
In particular, if the user performs a double click gesture on the display region of “image recognition” of the menu screen image 36 in the state in which the locus 38 surrounds the clock, then the physical solid image determination portion 114 extracts a mark image same as or similar to the real physical solid image from the mark data 142. Then, the extracted mark image is displayed for user confirmation in the AR display region. If the user looks at the mark image and confirms that the mark image is same as or similar to the clock surrounded by the locus 38, then the user would carry out a double click gesture on the display region of “registration” on the menu screen image 36. Consequently, the physical solid image determination portion 114 records the mark image as the physical solid image 144 into the recording unit 140. The decision of the sameness or similarity may be carried out in accordance with a degree of coincidence calculated from characteristic amounts of the real physical solid image and the mark image. For example, if the physical solid image determination portion 114 extracts a plurality of similar mark images from the mark data 142, then the mark images are displayed for user confirmation in the AR display region so as to allow selection thereof by the user.
It is to be noted that, although it is indicated that the mark data 142 includes mark images of various variations of table clocks, it includes mark images relating to various other objects such as, for example, desks, chairs, trains, buildings, and doors. Preferably, a mark image same as or similar to a picked up real physical solid image is suitably extracted.
In this manner, the physical solid image determination portion 114 determines a real physical solid image or a mark image as the physical solid image 144 and records the physical solid image 144 into the recording unit 140. The user can record a plurality of physical solid images 144 into the recording unit 140 by repeating the work just described.
As described above, in the application for reproducing a content in response to a situation of a user, the physical solid image 144 is a condition for selecting and reproducing music by being picked up. Accordingly, in order to create an event for executing the reproduction application, it is necessary for one physical solid image 144 to be registered in the event list without fail, and by registering a plurality of physical solid images 144, the possibility that one of the physical solid images 144 may be picked up is enhanced, which makes it possible to detect creation of an event with a high degree of accuracy. To this end, preferably the user records many physical solid images 144 into the recording unit 140 in the work for the “registration of event starting object” so that a plurality of physical solid image 144 can be registered for one event. It is to be noted that the physical solid image determination portion 114 may record the entire picked up image picked up by the image pickup unit 14 as the physical solid image 144 into the recording unit 140.
While the foregoing is directed to an example wherein the user designates a real physical solid by a free curve, also it is possible for the user to directly designate a real physical solid. If the user carries out a double click gesture on the display region of the “direct designation” on the menu screen image 36, then a selection mode of a real physical solid by direct designation is started. If the user carries out, in this mode, a click gesture for a real physical solid, then the real physical solid specification portion 112 specifies the real physical solid image selected by the user. In particular, the real physical solid specification portion 112 extracts a real physical solid including space coordinate values for which the click operation is carried out on the picked up image. When the real physical solid image is extracted, the real physical solid specification portion 112 may color the outer profile of the real physical solid image or the overall real physical solid image to notify the user that the extraction process is completed. If a double click gesture is carried out on the display region of “registration” of the menu screen image 36 and the input acceptance unit 100 sends double click operation information to the image processing section 110, then the physical solid image determination portion 114 records the extracted real physical solid image as the physical solid image 144 into the recording unit 140. It is to be noted that the physical solid image determination portion 114 may record a mark image corresponding to the extracted real physical solid image as the physical solid image 144 into the recording unit 140 as described hereinabove.
While the “free hand rendering” and “direct designation” modes are modes in which the user manually selects a real physical solid, the information processing apparatus 10 of the present embodiment prepares also a mode for selecting a real physical solid automatically.
If the user carries out a double click gesture on the display region of “automatic object selection” of the menu screen image 36, then an automatic selection mode of a real physical solid is started. In this mode, the real physical solid specification portion 112 specifies a real physical solid image included in the picked up image. At this time, the real physical solid specification portion 112 preferably selects a real physical solid image having a characteristic factor with respect to a surrounding environment in the picked up image. The characteristic factor here may be that the object has a high contrast to the background color, that the object has a shape greater than the surroundings or the like. For example, in regard to the example of
It is to be noted that, where a mark image is recorded as the physical solid image 144, the physical solid image determination portion 114 confirms whether or not a mark image corresponding to a real physical solid image specified by the real physical solid specification portion 112 exists. For example, if a mark image corresponding to the chair in front of the table is not included in the mark data 142, then since a mark image of the chair cannot be recorded into the recording unit 140, the physical solid image determination portion 114 preferably determines so that a slanting line image may not be superposed on the chair. Therefore, creation of a slanting line image by the real physical solid specification portion 112 is preferably carried out after confirmation of the presence of a corresponding mark image by the physical solid image determination portion 114 is carried out.
Further, since the selection process of a real physical solid in the “automatic object selection” mode is carried out irrespective of the will of the user, the user may be able to decide whether or not a real physical solid image specified by the real physical solid specification portion 112 or a mark image of the real physical solid image is to be recorded as the physical solid image 144. For example, it is assumed the chair in front of the table does not exist at the place (in the living room) but exists by chance at the time. Since the physical solid image 144 configures a condition for starting of an event, preferably it is an image of a physical solid which exists on a routine basis. The user can exclude the chair from a candidate for the physical solid image 144 by carrying out a double click gesture on the region to which slanting lines for the chair are applied.
In this manner, by selecting an item of the menu screen image 36, the user can record the physical solid image 144 into the recording unit 140 in various modes.
Here, an example of behaviors within a weekday of the user who is a system engineer is indicated. The user would wake up in the morning, take a breakfast, arrange dressing and go to work in the company. After the user arrives at the company, the user would sit in front of the desk and carry out a programming work while looking at the screen of a personal computer. When a lunch break comes, the user would take a lunch, and after the lunch break ends, the user would begin a work in the afternoon. When leaving time comes, the user would turn off the power of the personal computer and go home. After returning home, the user would take a bath and take a dinner and then enjoy a television program, and when bedtime comes, the user would go to bed. Such actions within one day as just described can be divided into various events.
a) 7:00 to 8:00 wake-up event
The user wakes up, has breakfast and arranges dressing.
b) 8:00 to 9:00 going-to-work event
The user leaves the home, gets on a train and goes to the company.
c) 9:00 to 12:00 work event
The user works in the morning.
d) 12:00 to 13:00 lunch break event
The user goes out of the company and takes a lunch.
e) 13:00 to 17:30 work event
The user works in the afternoon.
f) 17:30 to 18:30 return home event
The user leaves the company and returns home.
g) 18:30 to 19:30 taking-bath event
The user takes a bath.
h) 19:30 to 20:30 dinner event
The user takes a dinner.
i) 20:30 to 23:00 relax event
The user relaxes enjoying a television program or the like to refresh the user itself.
j) 23:00 sleep event
The user goes to bed.
This behavior table is a typical one in which a situation of the user is specified by an event name and time information. The information processing apparatus 10 of the present embodiment carries out supporting of the user to create such a behavior table as described above as a schedule table.
The information processing apparatus 10 of the present embodiment can process an application in response to a situation of the user. As described hereinabove, the processing of an application includes to start up and execute an application in response to a situation of the user and to execute a content in response to the situation of the user in a particular application. In order to facilitate understandings, in the following description, an example is described wherein the information processing apparatus 10 selects and reproduces a content in response to a situation of the user in a sound reproduction application.
If an event starting condition is satisfied, then the information processing apparatus 10 specifies a situation of the user and reproduces a content associated with the event. Since the starting condition of the event includes that a real image corresponding to physical solid image information associated with the event is picked up by the image pickup unit 14, it is necessary for the user to record a physical solid image 144 relating to the event in advance into the recording unit 140.
Thus, the user registers, for each of the events a) to j), a physical solid image 144 which defines a starting condition for the event into the recording unit 140.
a) Wake-Up Event
The user would wake up from the bed in the morning and wear the information processing apparatus 10 and then take a breakfast in the living room. After the breakfast, the user would change into outside clothes and read a newspaper until departure time comes. The user would record a physical solid image 144 representing the wake-up event into the recording unit 140. This recording process is carried out by such a technique as described hereinabove in connection with the menu items of the menu screen image 36. However, if the user goes to the living room without fail in the wake-up event, then the physical solid image 144 of a real physical solid (clock, table or the like) existing in the living room may be recorded into the recording unit 140. On the other hand, if the user reads a newspaper without fail, then a physical solid image 144 of a newspaper may be recorded into the recording unit 140. In this manner, preferably the user registers a physical solid image 144 which may be imaged by the image pickup unit 14 with high possibility in the wake-up event in advance. Further, in order that a start of the event is detected at an early stage, preferably the user registers a physical solid image 144 which may be imaged at an early stage of the event with high possibility in accordance with a behavior in the event. In the wake-up event, since the user goes immediately to the living room after wake-up, preferably a physical solid image 144 of a real physical solid existing in the living room is recorded in the recording unit 140. This similarly applies also to the other events.
b) Going-to-Work Event
When the going-to-work time comes, the user would take a work bag, open the door of the own home and go out. The user would go to the railway station by bus and go by train from the railway station, and then get off at the nearest railway station to the company and walk from the nearest railway station to the company. The user would record, for example, a physical solid image 144 of the door of the own home or a physical solid image 144 of the outside environment when the door of the own home is opened into the recording unit 140.
c) Work Event
After the user arrives at the company, the user would sit on the own seat and carry out a programming work while observing the display unit and the keyboard. Thus, the user would record a physical solid image 144, for example, of a building in which the company is situated, the entrance of the building, the display unit on the desk, or the keyboard into the recording unit 140.
d) Lunch Break Event
When the end time of the work in the morning comes, the user would leave the company and take a lunch in a diner and then return to the company before the start time in the afternoon. The user would record a physical solid image 144, for example, of the entrance of the company building or a physical solid on the way to the diner into the recording unit 140.
e) Work Event
After the user returns to the company, the user would sit on the own seat and carry out a programming work while observing the display unit and the keyboard again. The user would record a physical solid image 144, for example, of the company building, the entrance of the company building, the display unit on the own desk, or the keyboard into the recording unit 140.
f) Return Home Event
After the work end time comes, the user would leave the company holding the work bag, walk to the railway station, get on the train, leave the train at the nearest railway station to the own home and then return to the home by bus. The user would record a physical solid image 144, for example, of the work bag, the entrance of the company building, or a physical solid on the way to the railway station into the recording unit 140.
g) Taking-Bath Event
After the user returns home, the user would place the work bag in the own room and go to the bath room holding changing clothes. In the bath room, the information processing apparatus 10 is dismounted. After the user uses the bath, the user would wear pajamas and go to the living room. The user would record a physical solid image 144 of the door of the own room, the door of the bath room or the like into the recording unit 140.
h) Dinner Event
In the living room, the user would take a dinner. Similarly as in the a) event, the user would record a physical solid image 144 of the clock, the table or the like into the recording unit 140. It is to be noted that, if the image of any of them is recorded already, then the user need not record a physical solid image 144 of them newly.
i) Relax Event
The user would enjoy a television program lying down on the sofa to refresh the user itself. The user would record a physical solid image 144 of the sofa or the television set into the recording unit 140.
j) Sleep Event
The user would go to the bed room, dismount the information processing apparatus 10 and go to bed. The user would record a physical solid image 144 of the bed, the alarm clock or the like into the recording unit 140.
It is to be noted that, although the example described describes nothing of the sleep event from 0:00 to 7:00, this sleep event may be taken into consideration before the wake-up event.
Referring to
If the user carries out a double click gesture on the display region of the menu item “event registration” of the menu screen image 34, then the input acceptance unit 100 sends double click operation information and the associated AR display region coordinate values to the display processing section 120. If an event displaying portion 124 detects that the menu item “event registration” is selected, then it displays candidates for an event name which can be used for creation of a schedule table on the display apparatus 12.
If the user carries out a click gesture on the region in which an event name is displayed, then the input acceptance unit 100 sends click operation information together with the AR display coordinate values of the region to an event determination portion 116, and the event determination portion 116 retains the selected event name. If the user selects all event names to be used for creation of a schedule table, then the user would carry out a click gesture on the display region of “registration.” Then, the input acceptance unit 100 sends click operation information together with the AR display coordinate values of the display region of “registration” to the event determination portion 116, and the event determination portion 116 records all of the selected event names as events 146 into the recording unit 140. Here, “sleep,” “wake-up,” “going-to-work,” “work,” “break,” “return-home,” “taking-bath,” “dinner,” and “relax” are recorded as the events 146.
If the user carries out a double click gesture on the display region of the menu item “schedule creation” of the menu screen image 34, then the input acceptance unit 100 sends double click operation information and the AR display region coordinate values of the display region to the display processing section 120. The display processing section 120 receives the double click operation information and the AR display region coordinate values and provides a user interface for creating a schedule table.
In the schedule table creation screen image, items “event inputting” and “physical solid image inputting” at the upper stage are provided in order to specify targets to be inputted to the table region 42. In particular, “event inputting” is an item for inputting an event name to the table region 42, and “physical solid image inputting” is an item for inputting a physical solid image to the table region 42. In an initial screen image for schedule table creation, the item “event inputting” is set default, and the event displaying portion 124 reads out the events 146 recorded in the recording unit 140 and displays the recorded event names above the table region 42.
In the schedule table creation work described above, the user records physical solid images 144 in various environments into the recording unit 140 and then allocates events to the table region 42, whereafter the user allocates the physical solid images 144 to the events. In a modification, for example, after events are allocated to the table region 42, physical solid images 144 specified in various environments may be allocated directly to the events.
In this modification, the user would first allocate events to the table region 42 in accordance with the procedure described hereinabove with reference to
While the foregoing is directed to the examples in which a schedule table is created manually, alto is it possible for the image processing section 110 and the schedule registration section 150 to cooperate with each other to automatically create a schedule table. Also in this automatic creation, it is a prerequisite for the user to associate events and time information with each other in advance through the procedure described hereinabove with reference to
The real physical solid specification portion 112 specifies a real physical solid image included in a picked up image at start time associated with an event. At this time, the real physical solid specification portion 112 preferably specifies a real physical solid image having a characteristic factor with respect to a surrounding environment in the picked up image. The real physical solid specification portion 112 records the specified real physical solid image into the recording unit 140. The real physical solid specification portion 112 specifies and records the real physical solid image for each event into the recording unit 140. The real physical solid specification portion 112 carries out the recording process over a plurality of days. It is to be noted that, if the schedule table for which association between events and time information is performed is for weekdays, then the real physical solid specification portion 112 carries out the recording process over a plurality of weekdays, but if the schedule table is for a holiday, then the real physical solid specification portion 112 carries out the recording process over a plurality of holidays.
After sample images for a plurality of days are recorded into the recording unit 140 in this manner, the real physical solid specification portion 112 extracts real physical solid images picked up frequently for each event. For example, if the number of times by which an image of the clock is picked up in the wake-up event is great, then an image of the clock is extracted as a real physical solid image. By detecting a real physical solid image which is picked up by the greatest number of times at the start time of each event in this manner, the likelihood in that the real physical solid image is included in the starting condition of the event can be raised. In this manner, the real physical solid specification portion 112 specifies a real physical solid image, and the physical solid image determination portion 114 records the specified real physical solid image as the physical solid image 144 into the recording unit 140. It is to be noted that the physical solid image determination portion 114 may record a mark image corresponding to the specified real physical solid image as the physical solid image 144 into the recording unit 140 as described hereinabove. Since the image processing section 110 operates in such a manner as described above, a schedule table which associates events and physical solid images with each other is created while the user need not record a physical solid image manually into the recording unit 140.
It is to be noted that, while, in the example described above, the user sets an event manually or automatically, for example, the information processing apparatus 10 may prepare event models, which associate events for individual occupations and time information with each other, in advance such that the user selects a model in accordance with the own occupation so that an event is set automatically.
If the user carries out a double click gesture on the display region of the menu item “schedule editing” of the menu screen image 34, then the input acceptance unit 100 sends double click operation information and the AR display region coordinate values of the display region to the display processing section 120. The display processing section 120 receives the double click operation information and the AR display region coordinate values and provides a user interface for editing the schedule table. In this user interface, a schedule table editing screen image in which the schedule table created already is displayed in the table region 42 is presented to the user, and the user can edit the schedule table by carrying out such a work as described above.
If the user utilizes the user interface to set a schedule table as described above, then the information processing apparatus 10 creates an event list on the basis of the schedule table. The information processing apparatus 10 has an application processing function of using the event list to determine a situation of the user at present and starting processing of the application. In the following, as an example of the application processing function, a case is described in which a sound reproduction application reproduces a content in response to a situation of the user and outputs sound.
Although the components mentioned are implemented, where they are configured as hardware components, from a CPU and a memory of an arbitrary computer, a program stored in the memory, a storage and so forth,
The event list creation section 200 creates an event list which associates time information and physical solid image information with events on the basis of the schedule information registered in the recording unit 140 by the schedule registration section 150, and records the event list into the recording unit 140. More particularly, the event list creation section 200 derives time information for an event list from the time information set in the schedule table 148 to create an event list.
Here, differences between a schedule table and an event list are described. The schedule table represents behavior (event) schedules of the user for one day, and start time and end time set for each event indicate a time zone within which the event is to be carried out. Therefore, in the schedule table, time zones set for events do not overlap with each other in time, and at the same time with or later than end time of a preceding event, start time of a succeeding event is set.
Meanwhile, the event list is created in order to determine a situation of the user on the basis of the schedule table, and in order to raise the determination accuracy of the user situation, the time zone of each event set in the schedule table is set so as to be expanded. In the schedule table 148 depicted in
Therefore, the event list is created such that the time zone of each event in the schedule table 148 is expanded so that a behavior in a time zone displaced from that of the schedule table 148 can be grasped to make it possible to determine the situation of the user with high accuracy.
The event list creation section 200 creates an event list 152 on the basis of the schedule information registered in the schedule table 148 and records the event list 152 into the recording unit 140. As described hereinabove, the event list creation section 200 sets a time zone at least including a time zone of each event in the registered schedule information to the time information to be associated with the event.
It is to be noted that, as the event order number, the event list creation section 200 sets, for each event, a generation order number of the event in an associated relationship with the event. The event order number is set same as the arrangement order number of an event in the schedule table 148, and a number is applied in a descending order to each event beginning with the wake-up event to which “1” is applied as depicted in
The event list creation section 200 changes a time zone defined by starting time and ending time in the schedule information to a time zone defined by starting possible time and ending possible time. As described hereinabove, the time zone of each event in the event list 152 is expanded including at least the time zone of the event in the schedule table 148.
The event list creation section 200 determines starting possible time of each event in accordance with the event order number. Here, the starting possible time of a succeeding event, namely, an event to which a higher event order number is set, is set later than the starting possible time of the preceding event, namely, of the event to which the lower event order number is set. Although the starting possible time of the succeeding event may be same as the starting possible time of the preceding event, it must not be earlier than the starting possible time of the preceding event. Similarly, the ending possible time of the succeeding event is set later than the ending possible time of the preceding event. Although the ending possible time of the succeeding event may be same as the ending possible time of the preceding event, it must not be earlier than the ending possible time of the preceding event. The event list creation section 200 creates an event list in this manner. It is to be noted that the created event list may be modifiable manually by the user.
The mounted state determination section 210 determines a mounted state of the information processing apparatus 10 in accordance with measurement information from the motion sensor 50. Where the user has the information processing apparatus 10 mounted thereon, the motion sensor 50 provides measurement information representing that the information processing apparatus 10 is operative to the mounted state determination section 210 in an interlocked relationship with a motion of the user. On the other hand, if the user removes the information processing apparatus 10 and places the information processing apparatus 10, for example, on a table, then the motion sensor 50 provides measurement information representing that the information processing apparatus 10 is inoperative to the mounted state determination section 210. For example, if the measurement information from the motion sensor 50 does not vary for a predetermined period of time (for example, for three minutes), then the mounted state determination section 210 determines that the information processing apparatus 10 is removed, but if the measurement information varies within the predetermined time period, then the mounted state determination section 210 determines that the information processing apparatus 10 is mounted.
In the following, the application processing function of the information processing apparatus 10 is described. This application processing function is executed in a state in which the event list 152 which associates time information and physical solid image information with events is created by the event list creation section 200 and recorded in the recording unit 140. It is to be noted that, while the event list creation section 200 creates an event list 152 on the basis of the schedule table 148 created by the information processing apparatus 10, for example, the schedule table 148 may be created by a different apparatus and the event list creation section 200 may create an event list on the basis of the schedule table 148 created by the different apparatus.
If the user turns on the information processing apparatus 10 and mounts the information processing apparatus 10 on itself, then the image pickup unit 14 picks up an image of the real space periodically and provides picked up images to the control section 230. The control section 230 acquires the picked up images and refers to the event list 152 to determine whether or not an event starting condition is satisfied. Then, if it is determined that an event starting condition is satisfied, then the control section 230 instructs the application execution section 220 to perform processing of the application. In response to the instruction, the application execution section 220 starts processing of the application associated with the event with regard to which the starting condition is satisfied.
In order to implement the application processing function, the control section 230 includes a candidate extraction portion 232, an image processing portion 234, a condition determination portion 236, and a starting instruction portion 238. Here, the candidate extraction portion 232 acquires current time information from the insider or the outside and specifies an event between the starting possible time and the ending possible time of which the current time information is included. For example, if the current time information is 7:30, then the candidate extraction portion 232 extracts the wake-up event of the event order number 1, the going-to-work event of the event order number 2, and the work event of the event order number 3. The extracted events are event candidates which can satisfy the starting condition at the point of time of 7:30.
The image processing portion 234 determines whether an image associated with an event extracted by the candidate extraction portion 232 is included in a picked up image picked up by the image pickup unit 14. This corresponds to a process by the image processing portion 234 of determining whether an image corresponding to the physical solid image information is included in a picked up image picked up within a time zone specified by the time information in the event list 152. In particular, it is determined whether or not an image corresponding to the physical solid image information associated with each of the wake-up event, going-to-work event, and work event is included in the picked up image at the point of time of 7:30. It is to be noted that the determination process of whether or not an image corresponding to the physical solid image information is included in a picked up image is carried out by deriving a coincidence degree between the characteristic amount of the image corresponding to the physical solid image information and the characteristic amount of an image included in the picked up image. If the characteristic amounts coincide fully with each other, then the images are identical with each other, and if the coincidence degree is very high, then the images are similar to each other. If the image processing portion 234 determines that an image corresponding to the physical solid image information is included in the picked up image, then the condition determination portion 236 decides that an event starting condition is satisfied.
By recording physical solid image information of a plurality of physical solid images in an associated relationship with an event in the event list 152, the possibility that a starting condition of the event may be satisfied can be enhanced. For example, where physical solid image information of the clock and the table is recorded in an associated relationship with the wake-up event, if the image processing portion 234 determines that an image of the clock or the table is included in a picked up image, then the condition determination portion 236 can determine that a starting condition of the wake-up event is satisfied. By associating physical solid image information of a plurality of physical solid images with an event in this manner, it is possible to accurately determine satisfaction of a starting condition of the event. If the condition determination portion 236 determines that a starting condition of the wake-up event is satisfied, then the starting instruction portion 238 instructs the application execution section 220 to carry out processing of the application and notifies the application execution section 220 of the event name for which the starting condition is satisfied. The application execution section 220 starts processing of the application corresponding to the event whose starting condition is satisfied.
Although the content table 154 may be created by the user associating contents with events, in the example depicted in
It is to be noted that, as depicted in
It is to be noted that, while the image processing portion 234 determines whether an image associated with an event extracted by the candidate extraction portion 232 is included in a picked up image, it may exclude an image which is determined as being included in a picked up image from a subsequent determination target. This can reduce the load by the image determination process. Further, it may exclude entire images included in the event whose starting condition is satisfied from a determination target. This eliminates the necessity to carry out a determination process of a different image included in the started event and can reduce the load by the image determination process.
Further, the condition determination portion 236 may determine satisfaction of an event starting condition in accordance with the generation order number registered in the event list. For example, referring to
It is to be noted that the application execution section 220 continuously executes processing of an application until it is notified from the starting instruction portion 238 that a starting condition for a different event is satisfied. If the application execution section 220 is notified that a starting condition for a different event is satisfied, then the application execution section 220 refers to the content table 154 to start reproduction of a content associated with the different event. It is to be noted that, if the application execution section 220 is not notified during execution of a content associated with a certain event but before the ending possible time of the event that a starting condition for a different event is satisfied, then the application execution section 220 ends the reproduction process of the content at the ending possible time.
Further, the application execution section 220 may continue or end processing of an application in response to a mounted state of the information processing apparatus 10. If the mounted state determination section 210 determines that the user does not have the information processing apparatus 10 mounted thereon, then the mounted state determination section 210 may notify the application execution section 220 of this, and the application execution section 220 may stop the processing of the application. Where the information processing apparatus 10 is a wearable computer, since the user can listen to reproduced sound from the earphone 16 in a state in which the user has the information processing apparatus 10 mounted thereon, if the output sound cannot be heard, then there is no necessity for the application execution section 220 to output and reproduce sound. Therefore, it is preferable that, if the information processing apparatus 10 is not mounted on the user, the application execution section 220 stops processing of the application.
The present invention has been described in connection with the embodiment thereof. It can be recognized by those skilled in the art that the embodiment is illustrative and various modifications can be made to the combinations of the components and the processes of the embodiment and that also such modifications are included within the scope of the present invention.
While it is described in the description of the embodiment that the event list creation section 200 creates an event list which associates an event generation order number, time information, and physical solid image information of one or more physical solid images with an event, the event list creation section 200 may create an event list which further associates GPS information with the items mentioned. The GPS information represents position information of a position at which physical solid image information is acquired. Therefore, the candidate extraction portion 232 can easily extract a candidate event with regard to which GPS information recorded in the event list and GPS information of the position at present coincide with each other.
Further, while it is indicated in the embodiment that a content of music or the like is reproduced for the user, mark image information of a company or a merchandise may be registered in the event list. Similarly as in the embodiment, also in such an event list as just described, time information and mark image information are registered in an associated relationship with an event. The control section 230 determines whether an image corresponding to mark image information is included in a picked up image picked up within a time zone prescribed in the event list. Then, if the image is included in the picked up image, then the control section 230 determines that the event starting condition is satisfied, and instructs the application execution section 220 to carry out processing of the application. In response to the instruction, the application execution section 220 may cause the display apparatus 12 to display advertisement information or the like associated with the event.
Further, while it is described in the foregoing description of the embodiment that the application execution section 220 refers to the content table 154 to reproduce a content corresponding to an event, it may be devised to apply countermeasures for preventing, upon such reproduction, the same content from being always selected with regard to the event. The application execution section 220 may have a content searching function such that, for example, an object name included in a picked up image and, for example, a genre and an object name of the content table 154 are used as search keys to search for and acquire a content from the content server 4.
Further, while the event list creation section 200 creates an event list which associates physical solid image information with an event, for example, sound information may be associated with an event. For example, sound generated at a particular place or in a condition for event detection may be acquired from the microphone 20, and the acquired sound information may be utilized for a determination process of a starting condition of an event in association with an event.
It is to be noted that, while functional blocks of the information processing apparatus 10 are depicted in
In particular, the terminal apparatus 2 may include the input acceptance unit 100, the image processing section 110 and the schedule registration section 150 of the processing unit 80, and the recording unit 140 depicted in
1 . . . Information processing system, 10 . . . Information processing apparatus, 12 . . . Display apparatus, 14 . . . Image pickup unit, 16 . . . Earphone, 18 . . . Housing, 20 . . . Microphone, 50 . . . Motion sensor, 52 . . . GPS sensor, 80 . . . Processing unit, 100 . . . Input acceptance unit, 110 . . . Image processing section, 112 . . . Real physical solid specification portion, 114 . . . Physical solid image determination portion, 116 . . . Event determination portion, 120 . . . Display processing section, 122 . . . Menu displaying portion, 124 . . . Event displaying portion, 126 . . . Physical solid image displaying portion, 128 . . . Time axis displaying portion, 130 . . . Locus displaying portion, 140 . . . Recording unit, 150 . . . Schedule registration section, 200 . . . Event list creation section, 210 . . . Mounted state determination section, 220 . . . Application execution section, 230 . . . Control section, 232 . . . Candidate extraction portion, 234 . . . Image processing portion, 236 . . . Condition determination portion, 238 . . . Starting instruction portion.
The present invention can be utilized in the information processing field for processing a picked up image.
Filing Document | Filing Date | Country | Kind |
---|---|---|---|
PCT/JP2013/002990 | 5/9/2013 | WO | 00 |