The present invention generally relates to computer systems and, more particularly, to interactive computer systems.
The concept of ubiquitous computing is growing in popularity. A number of technologies have been developed in connection with the ubiquitous computing field. For example, U.S. Pat. No. 6,431,711 to Pinhanez discloses a multiple-surface display projector with interactive input capability. In the Pinhanez patent, techniques are disclosed wherein an image is projected onto a surface in a room and is distorted before projection so that a projected version of the image will not be distorted. The surface can be planar or non-planar. The projected image can be displayed at multiple locations along a surface or multiple surfaces. Thus, the projected image can move from one location on a surface to another location on that surface or another surface, while the projected image remains undistorted through the move.
Techniques using the Pinhanez invention allow interaction between people and a projector. Interactive input, such as from mice, can be used with versions of the Pinhanez invention. Versions of the Pinhanez invention can determine if an object is near an interactive item, such as a hyperlink, on the projected image. In such case, the interactive item can be activated. Thus, a person can interact with a projected image.
While the techniques of the Pinhanez patent provide a substantial improvement on the prior art, there may be potential limitations regarding the physical space that can effectively be displayed upon, and where interactions can be sensed. Further, there may be limitations regarding flexibility of display on multiple surfaces.
Principles of the present invention provide techniques for interacting with a subject in an environment.
In one aspect, an exemplary apparatus for interacting with a subject in an environment, according to an exemplary embodiment of the present invention, includes at least one display device, at least one sensing device, and at least one display controller. The at least one display device can be configured to display images on a number of surfaces in the environment. The at least one sensing device can be configured to sense interactions between the subject and the environment. At least some of these interactions can be dispersed through the environment, and remote from images displayed by the display device. The at least one display controller can be coupled to the display device and the sensing device, and can be configured to modify image display of the display device responsive to at least one of the interactions sensed by the sensing device.
The modification of the image display of the display device can include modification of one or more of image location, image orientation, and image content. At least some of the images can include interactive items, and the display controller can be configured to activate a given one of the interactive items when an object is proximate to the interactive item. The proximate object can be, e.g., one or more of a body of the subject, a hand of the subject, a finger of the subject, an obstruction manipulated by the subject, and a laser-pointer-illuminated region that is illuminated by action of the subject. The interactions that the sensing device is configured to sense can include proximity of an object to the surfaces and/or performance of a predefined motion by the subject proximate to one of the surfaces. The predefined motion can be one or more of touching and pointing.
Where desired, the exemplary apparatus can also include at least one sensor processor that is coupled to the sensing device and the display controller. The sensor processor can be configured to receive a data stream from the sensing device and to output a set of object parameters O and a set of interaction parameters K to the display controller. The set of object parameters can include one or more of a number of objects, a location of each of the objects, and a label of each of the objects. The set of interaction parameters can include one or more of a number of interactions detected, a location of each of the interactions, and a type of each of the interactions.
In another aspect, an exemplary apparatus for interacting with a subject in an environment, according to another exemplary embodiment of the invention, can include at least one projector, at least one physical display, at least one sensing device, and at least one display controller. The at least one projector can be configured to project images on a number of surfaces in the environment. The at least one physical display can be located in the environment and configured to display images to the subject. The at least one sensing device can be configured to sense interactions between the subject and the environment. The at least one display controller can be coupled to the projector, the physical display, and the sensing device, and can be configured to modify image display of the projector and the physical display responsive to one or more of the interactions sensed by the sensing device.
In still another aspect of the invention, an exemplary apparatus for interacting with a subject in an environment, according to still another exemplary embodiment of the invention, can include at least one display device, at least one camera, at least one sensor, and at least one display controller. The at least one display device can be configured to display images on a number of surfaces in the environment. The at least one camera can be configured to sense interactions between the subject and at least some of the images. The at least one sensor can be configured to sense interactions between the subject and the environment that are remote from the images. The at least one display controller can be coupled to the display device, the camera, and the sensor. The at least one display controller can be configured to modify image display of the display device, responsive to the interactions sensed by the camera and the interactions sensed by the sensor.
In yet another aspect of the invention, an exemplary apparatus for interacting with a subject in an environment, according to still another exemplary embodiment of the invention, can include at least one video projector, at least one video projector mounting, at least one video camera, and at least one controller coupled to the video projector, the video projector mounting, and the video camera. The at least one projector can be configured to display images on a number of surfaces in the environment. The at least one video projector mounting can be coupled to the video projector and configured to move the video projector in at least two rotational degrees of freedom. The at least one video camera can be configured to sense interactions between the subject and the environment. The at least one controller can include a memory and at least one processor, coupled to the memory, that is operative to modify image display by the video projector responsive to one or more of the interactions sensed by the video camera.
In still a further aspect, the present invention can include methods, which can be computer-implemented, for interacting with a subject in an environment, which implement the actions described with regard to the various exemplary apparatuses. By way of example and not limitation, one exemplary method can include the steps of displaying images on a number of surfaces in an environment, sensing interactions between a subject and the environment, and modifying image display responsive to one or more of the interactions sensed by the sensing device. At least some of the interactions can include interactions that are dispersed through the environment and are remote from the images.
One or more exemplary method steps and/or apparatus elements can be implemented using a computer system and/or can be embodied in a computer usable medium having computer usable program code thereon.
These and other objects, features and advantages of the present invention will become apparent from the following detailed description of illustrative embodiments thereof, which is to be read in connection with the accompanying drawings.
Apparatus 100 can also include one or more sensing devices 112 that are configured to sense interactions between the subject 102 and the environment 104. In one or more aspects of the invention, the interactions that are sensed can include interactions that are dispersed through the environment 104 and are remote from the images 108 displayed by the display device. The sensing devices are designated as S1, S2, . . . Sn.
Apparatus 100 can further include one or more display controllers such as display controller 114. Controller 114 can be coupled to the display device or devices 106 and the sensing device or devices 112, and can be configured to modify image display by the display device or devices 106 in response to one or more of the interactions sensed by the sensing device or devices 112. Such modification of the image display can include modification of one or more of image location, image orientation, and image content. It should be noted that display controller 114 can be implemented as a single unit 114 as shown, or may be realized as a collection of individual display controllers for each display device 106, or as some combination of the foregoing.
One or more of the images 108 can include an interactive item or items 116 (analogous to hyperlinks), and the display controller 114 can be configured to activate an interactive item such as item 116 when an object is close to the interactive item. Such an object could include, by way of example and not limitation, one or more of the body of subject 102, a hand of subject 102, a finger of subject 102, an obstruction manipulated by subject 102, or even an illuminated region. In the example depicted in
In view of the foregoing discussion, it will be appreciated that some of the exemplary types of interactions that can be sensed by the sensing devices 112 include proximity of an object to one of the surfaces such as surfaces 110, or the performance of a predefined motion by the subject 102 proximate to one of the surfaces 110. Such predefined motion could include, for example, touching and/or pointing.
The sensing devices 112 can sense and/or monitor various parts of the environment 104. The sensing devices 112 can be of different types. For example, some of the sensing devices can be optical cameras, others could be radio frequency identification tags, others could be microphones, and still others could be pressure sensors. Each of the sensing devices 112 can be configured to monitor different parts of the environment 104. By way of example and not limitation, in an environment such as a retail store, different cameras could be viewing different aisles in the store. Radio frequency identification tags may be present on products on the shelves in the store. Pressure sensors may be present in certain areas, such as the entrance to an aisle. Microphones might be present at selected locations on walls, for example, with an “ask for help here” sign located next to them. Again, the foregoing is intended as an illustrative example, not intended to be limiting, regarding the manner in which different types of sensing devices can sense and monitor different activities taking place in the environment 104. As noted, in one or more exemplary embodiments of the invention, one or more of the sensing devices can be configured to sense interactions (between the subject and the environment) that are dispersed through the environment 104 and remote from the images 108 displayed by one or more of the display devices 106. Again by way of example and not limitation, such dispersed interactions could be sensed by the aforementioned microphones, pressure sensors, or radio frequency identification tags, or by optical cameras focused on areas other than the images 108.
It should be noted that elements such as the one or more display devices 106, one or more sensing devices 112, and one or more sensor processors 122, to be discussed more fully below, are shown in block diagram form in
Another type of display could be a video projector, such as shown at 106B, which projects on a wall, or an area of the floor, etc., such as on surfaces 110. One or more of the projectors, such as 106B, could be movable, or the images therefrom could be movable using a redirection device such as a pan/tilt mirror as in the aforementioned U.S. Pat. No. 6,431,711 Pinhanez. Another possible type of redirection device is a motorized pan/tilt unit as available from Directed Perception Inc. under item number PTU-D46-17 or PTU-D46-70. Accordingly, appropriate redirection can be employed with such movable projectors, or projectors having redirection devices, such that images can be displayed on different surfaces within the environment 104.
It will be appreciated that one or more exemplary embodiments of the present invention enable sensing of the subject 102 (broadly understood to also include objects adjacent to, associated with, or under control of the subject) and interactions of the subject with the environment 104. Further, the location, orientation or content of one or more images can be changed in response to the sensed interactions, using one or more of the display devices 106. As noted above, apparatus 100 can further include one or more sensor processors 122. Such processors 122 can be coupled to the sensing device or devices 112 and to the display controller 114, and can be configured to receive a data stream from the sensing device or devices 112. One or more sensor processors 122 can be employed; in the exemplary embodiment depicted in
The set of object parameters could include, e.g., the number of objects, the locations of each object, and the label of each object. In a retail environment the objects could be human beings, specific products, shopping carts, and the like. The set of interaction parameters could include, for example, the number of interactions detected, the location of each interaction, and the type of each interaction. Again by way of example and not limitation, in a retail environment, examples of interactions include a person walking into an aisle, a shopper picking up an object, a user touching a display, or a shopping cart traversing the aisle. As noted, when reference is made to interacting with a subject in an environment, this should be broadly understood to include interactions with any of the foregoing objects present in the environment that are under the control of or otherwise associated with or adjacent the subject (for example, by virtue of being accessible to the subject in the environment).
The outputs of the sensor processors 122 can be received by display controller 114, along with a number of data streams V1, V2, . . . Vn. At each instance of time “t” the display controller(s) can also have one or more inputs designated generally as 124. Such inputs 124 can correspond to the status of the display devices 106 at the previous instance of time “t−1.” The parameters 124 could include, by way of example and not limitation, an image Ij displayed by the display device Dj and a set of parameters Pj that define one or more of the position, size and orientation of the pertinent displays. Based on the outputs of the sensor processors 122 and the display device parameters 124 at the previous instance of time, the display device(s) can determine new display device parameters I1t, I2t, . . . Irt and/or P1t, P2t, . . . Prt. Such parameters thus could include the image I to be currently displayed by each display device 106, and the set of position parameters P for that device. Thus, the display controller 114 can be configured to determine one or more of content, location and orientation of the images displayed in the environment 104, based on sensed objects and/or interactions in the environment. For example, in a retail environment, the entry of a person into an aisle may result in a welcome message displayed on the floor of the aisle in an orientation suitable for viewing by that person. In another example, when a person picks up a product from a shelf, the system may use a set of consecutive displays, both static displays such as 106A, and a movable projection display such as 106B, to guide the person to a related product that is on sale.
It will be appreciated that the exemplary apparatus 100 depicted in
The at least one display controller 114 can be coupled to the projector, physical display, and sensing device and can be configured to modify image display of the projector and the physical display 106A, responsive to one or more interactions sensed by the sensing device. The display controller 114 can be adapted to distort an undistorted image into a distorted image for projection by the projector 106B. The apparatus can further include a projector mounting, of a type to be described below, to which the projector 106B is mounted. The projector mounting could be configured to perform one or more of rotating and translating the projector 106B to project the distorted image on one or more of the plurality of surfaces 110. The mounting could, if desired, be a combined mounting wherein the aforementioned camera and projector 106B are collocated for substantially coupled motion. The image could be distorted such that when projected on one or more of surfaces 110, it appears in substantially undistorted form.
Display controller 114 could include a video adapter, and could be configured to set parameters of a correction surface in the video adapter and to place the undistorted image as a texture on the correction surface. The display controller could also include, for example, focus and/or zoom control, position control, and the like. Further, the display controller could be configured to set parameters of a transformation matrix to distort the undistorted image into a distorted image. As discussed below, projector 106B could be configured for both rotational and translational movement with respect to environment 104. The aforementioned correction surface could have parameters including three rotational coordinates, two translational coordinates, a lens parameter, and a scale parameter. One or more of the sensing devices 112 could be a camera configured to sense interactions. One or more of the sensor processors 122 could be image processors configured to receive a video data stream from an associated camera, and to output a set of object parameters O and a set of interaction parameters K to the display controller 114. The object parameters and the interaction parameters could be of the type described above. As noted, one of the sensors 112 can be a camera that is configured to receive a camera image. Such image can be formed by a reflection from one of the surfaces 110 of one of the images projected by the projector 106B. The display controller 114 can be configured to communicate zoom and focus information to the projector 106B. The aforementioned image processor or image processors can be configured, for example, to perform method or processing steps to be discussed with respect to
The exemplary apparatus of
Actions described above as being performed by various elements depicted in
In one or more exemplary embodiments of the present invention, the surfaces on which the images are displayed can be divided into one or more display zones. As shown at decision block 208, a determination can be made whether any of the display zones are adjacent a given one of the interactions that has been sensed. If it is determined that at least one of the display zones is adjacent to the given one of the interactions, as indicated by the “Y” branch emanating from decision block 208, the adjacent display zone can be selected for display, as at step 218, of an updated image that is modified in accordance with the given one of the interactions.
The images can be displayed, for example, by display devices in the display zones. A determination can be made, as shown at block 220, whether one or more of the display devices in an adjacent display zone is movable. If such is the case, as shown at block 222, the given display device can be moved to a new location (as required). Regardless of whether or not the device is movable, the updated image can be displayed as at block 224. If none of the display zones is adjacent the given one of the interactions, as indicated at the “N” branch emanating from decision block 208, a determination can be made, as at block 210, whether some other kind of notification to the subject is possible. If such alternative notification is possible, as indicated at the “Y” branch of decision block 210, the notification can be sent as indicated at block 212. Such notification could be, for example, by means of a loudspeaker or a flashing light. If no other notification is possible, and/or if an appropriate display has been made, the given interaction can be removed from a list of interactions, as shown at step 214. As indicated at block 216, one can continue to process, or if one had completed all desired processing, one could stop (the terminology “one” is intended to encompass performance of the steps by a computer).
A decision block 308 determines whether any of the display zones is within an acceptable range, for example, in terms of parameters such as location, orientation and size, from the selected interaction. If such is not the case, a further decision step 310 determines whether it is necessary and/or possible to provide an alternative notification (as discussed above) in response to the sensed interaction; for example, such alternative notification could be a voice message or a beep. If the decision in block 310 is affirmative, the appropriate notification can be selected and sent via processing step 312. If, conversely, the decision in 310 is negative, no display device need be activated, and the current interaction can be removed from the list of interactions as in step 314. The updated list of interactions can then be processed in step 306, continuing in a looping fashion until all detected interactions are processed. Note that if the notification is sent in block 312, the interaction can also be removed from the list, and the looping can continue, as just described.
If the decision in step 308 is affirmative, and there are one or more display zones available within an acceptable range for the sensed interaction, at step 316, a display zone and display device can be selected. At step 318, an image that is to be displayed can be obtained. Such image may come, for example, from a separate application such as an advertising application, or may be created within step 318. The content of the image obtained in step 318 can then be modified in step 320, based on the sensed interaction. Further, in decision block 322, a determination can be made whether the selected display device is static or movable. If not moveable, one can proceed directly to block 324, and the modified image can be sent to the selected display device.
If it is determined in step 322 that the selected display device is movable, at block 326, a determination of the appropriate device location and orientation parameters for the moveable device corresponding to the selected display zone can be made. In step 328, the selected display device can be moved to its new location and orientation, for example, by sending appropriate control signals. At step 330, appropriate parameters to distort the undistorted image obtained in step 320 can be determined, such that when displayed by the moveable display device, a substantially undistorted image appears on the selected surface and within the selected display zone. Techniques for distorting the image in a moveable projection system are known, for example, from the aforementioned U.S. Pat. No. 6,431,711 to Pinhanez. Appropriate image distortion can be performed as indicated at step 332, and the properly distorted image can be sent to the movable display device as in the aforementioned step 324 (distortion may not be needed for displays other than projectors). The currently selected interaction can then be removed from the list of interactions as at step 314, and the updated list of interactions can be processed in step 306, again continuing with a loop processing all desired detected interactions. Accordingly, it will be appreciated that
Display controller 414 can be coupled to the video projector 406, video projector mounting 426, and video camera 412. Controller 414 can include a memory and at least one processor. The processor can be coupled to the memory, and can be operative to modify image display by the video projector 406 responsive to one or more of the interactions sensed by the video camera 412. Processor 414 can also be operative to make determinations regarding display zones, as described above with respect to
In one preferred form of apparatus 400, a single video camera 412 and a single video projector 406 are employed. The processing capability can be provided, for example, via a computer system 448 of a type to be discussed in connection with
Image processor 422 could process the sequence of images V coming from the camera 412, and could output a set of parameters including the object parameters O and interactions K as described above. Interactions sensed by camera 412 could include, for example, presence of a person in an aisle, the act of picking up a product, a hand touching an image displayed by the projector 406, and the like. The parameters and the image sequences could be processed by the processor 414. Processor 414 can also receive the parameters 424 including the image I displayed by the projector 406, and the parameters P of the video projector mounting 426, corresponding to the previous time instance t−1. The parameters P can include, for example, pan value, tilt value, pan and tilt speeds, current position and orientation, focus and zoom of the projector 406, as well as image distortion parameters. Based on the sensed interaction K, the controller 414 can determine a new image to be displayed by the projector 406 at the current time instance t, as well as new parameters for the video projector mounting 426. Accordingly, apparatus 400 represents one possible form of a steerable interactive system.
A display zone can indicate an area in which an image can be displayed via the movable video projector. In step 508, the currently available display device parameters can be employed in the determination. In decision block 510, a check can be made whether any of the display zones is within an acceptable range, in terms of parameters such as location, orientation, and size, from the selected interaction. If such is not the case, at additional decision block 512, a check can be made whether other notification is necessary and/or possible in response to the sensed interaction. As noted above, such alternate notification could be, for example, a voice message or a beep. If the decision in block 512 is affirmative, the appropriate notification can be selected and can be sent as per step 514. In the event that the decision in block 512 is negative, no display device needs to be activated, and the current interaction can be removed from a list of interactions as at step 516. The updated list of interactions can then be processed in step 508, thus forming a loop for processing of detected interactions. Note that if the notification is sent in block 514, the interaction can also be removed from the list, and the looping can continue, as just described.
If the decision in block 510 is affirmative, and there are one or more display zones available within an acceptable range for the sensed interaction, at step 518, one can select a display zone from among those within the acceptable range. In step 520, the image to be displayed can be obtained. As noted above, such an image may come from a separate application such as an advertising application, or may be created within step 520. The content of the image obtained in step 520 can be modified in step 522, based on the sensed interaction.
In step 524, a determination can be made regarding appropriate redirection parameters corresponding to the selected display zone. In step 526, the projector can then be moved to its new location and orientation by sending appropriate control signals to the video projector mounting, which can also be referred to as a redirection device. In step 528, image distortion parameters can be estimated as needed to distort the undistorted image obtained in step 522, such that when displayed by the moveable video projector, a substantially undistorted image will appear on the selected surface within the selected display zone. Image distortion has been discussed elsewhere herein, and can be performed as indicated in step 530, with the distorted image sent to the moveable display device in step 532. The currently selected interaction can then be eliminated from the current interaction list in step 516, and the updated list processed in step 508, thus continuing the loop for processing of the detected interaction(s). Thus,
In one or more aspects or embodiments of apparatuses and methods in accordance with the present invention, a computer or other components of embodiments of the present invention may not be readily accessible for a human to control, drive, and/or set up or calibrate. Accordingly, techniques for remote set-up, calibration, and/or control of systems and methods according to the present invention may be desirable.
The local computer 748 can be connected, through a network connection 750 such as an Ethernet or wireless connection, to a remote computer 752. Remote control software such as the PCAnywhere software from Symantec, Inc. can be employed to enable replication and control of the displays and the interfaces on the local computer 748 at the remote computer 752. An appropriate user interface 754 can be provided on the local computer to enable user interaction with the system for set-up, viewing, control, and the like. Interface 754 can communicate with the display controller 714 to achieve appropriate display and control functions as required or desired for user interaction.
The replication of the entire desktop interface of the local computer 748 on the remote computer 752 can enable access to the user interface by a remote user 756 through the replicated interface 758. Accordingly, remote user 756 can perform functions such as set-up and/or control of the local system depicted in the left-hand portion of
During calibration of the local system, location, orientation, and size of one or more display zones in the environment may be defined by the user. Accordingly, the user should be able to view the environment while projecting images and manipulating them through the user interface 758. Camera views of the local environment may not permit proper calibration, as the cameras may present distorted views of the environment due to camera geometry. Thus, during calibration, a viewer might need to be in the target (local) environment even if a remote computer were being used to set up the system. Accordingly, it is desirable to be able to effectively calibrate the system from a remote location.
As shown in step 802, a three-dimensional model of the target environment can be obtained. It should be noted that such a three-dimensional model could be a partial model; for example, it might include only position, size, and orientation of one or more essential display surfaces in the environment. Indeed, the three-dimensional model might include a model of only a single two-dimensional planar surface, however, it would still be a three dimensional model because spatial orientation data of the two-dimensional planar surface would typically also be included. Stated in another way, the three-dimensional model might not include every book on a given shelf within the environment. Image and parameter data of the target environment can be obtained as suggested by the dotted line from block 828 under “Target Environment” to block 804 under “Remote Environment.” The image and parameter data can be mapped onto the three dimensional model of the target environment, for example, in simulation step 804. As shown within block 805, a number of steps can be employed in the process, culminating in the development of calibration data of the target environment based on the mapping.
The aforementioned three-dimensional model of the target environment can be maintained on the computer in the remote environment. In one or more embodiments, the model can be built beforehand based on measurements made in the target environment either manually or through automated or semi-automated techniques. In the simulation step 804, the remote computer can simulate a view of the target environment based on the aforementioned three dimensional model as well as the aforementioned projector and camera images received from block 828 under “Target Environment.” method. In the simulation, the current projector and camera images can be mapped onto the surfaces of the three-dimensional model of the target environment. In step 806, the simulated environment can be displayed to a user through an interactive interface. The user can be permitted to view, update, and/or manipulate current positions of the camera and projector, or other sensing and display devices, and to correct one or more of location, orientation, size, and distortion of images through the simulated environment. As the user manipulates the three-dimensional model, the user is able to view the environment substantially without distortions that might be present if mere camera views of the target environment were employed.
In decision block 808, a determination can be made whether the user needs to further update the environment. If such is the case, such updates can be performed in block 810, and updated parameters can be sent to step 804, resulting in a new simulation and display to the user in block 806. If no further update is required in block 808, new parameters for the redirection device, focus and zoom parameters of the projector and/or camera, image distortion parameters for the projected image in the target environment, and the like can be computed anew in step 812, based on the last value selected by the user in the simulation environment. In step 814, such new parameters can be transmitted back to the system computer in the target environment.
In view of the foregoing discussion, it will be appreciated that the development of calibration data of the target environment based on mapping can include one or more of updating a portion of the image and parameter data to obtain updated image and parameter data, mapping the updated image and parameter data onto the three-dimensional model, and configuring the updated image and parameter data for transmission to the interactive system associated with the target environment. Such transmitted data can be employed as calibration data. The updating and mapping steps can be repeated until it is determined, during the mapping, that the updated image and parameter data is satisfactory. The results of the repeated steps can then be configured for transmission at block 814. The aforementioned image and parameter data can include one or more of focus data, zoom data, pan and tilt data, image data from a camera in the remote environment, and image data from a projector in the remote environment. The image and parameter data could also include displacement data, where the projector or other display device in the target environment is capable of translational as well as rotational motion. As noted, the three-dimensional model of the target environment could be built beforehand, and accordingly, one or more embodiments of the method can include an additional step of building the aforementioned three-dimensional model of the target environment.
In view of the foregoing discussion, it will be appreciated that in one or more other aspects of the invention, the developing step could include displaying results of the mapping step to a user, obtaining user input defining updates to at least a portion of the image and parameter data, so as to obtain updated image and parameter data, repeating the displaying step and the step of obtaining user input until the user determines that the updated image and parameter data is satisfactory, and configuring the updated image and parameter data for transmission to the interactive system associated with the target environment, again, for use as calibration data.
It should be noted that one or more of the aforementioned steps might be performed by an external system, apparatus, process, or method. For example, a software module implementing techniques of the present invention might simply format data for display to a user, and/or format parameters for transmission to the target computer, and such display and/or transmission, for example, could be performed by one or more systems, apparatuses, processes, or methods external to routines implementing the present invention.
Attention should now be given to the right-hand portion of
New parameters, such as the aforementioned calibration data, can be read from the remote environment in step 830. In step 832, new distortion parameters can by applied to the image to be displayed. In step 834, the updated image can be sent to the projector for display. In 836, the focus and zoom parameters of the projector and camera can be adjusted and/or updated. In step 838, the redirection device can be driven to a new location based on the received parameters. Accordingly, it will be appreciated that one or more of the previous steps essentially set forth, as indicated by dotted block 840, a possible way or way(s) to perform a step of updating display and sensor operation of the interactive system based on the calibration system. The updated parameters can then again be read by steps 816-828, enabling the remote user to view the changes through the simulation environment. Thus, the exemplary method steps depicted in
In view of the foregoing discussion, it will be appreciated that an additional method step of obtaining the aforementioned precalibration data can be included. Furthermore, it will be appreciated that in block 818, more generally, the parameters that are estimated could constitute spatial parameters associated with the camera and/or projector, and such spatial parameters could be estimated based on the aforementioned precalibration data and/or the redirection device parameters. Further, such spatial parameters could form at least part of the aforementioned image and parameter data. The image and parameter data could also include, for example, camera and projector focus and zoom parameters, projector image data, and camera image data. The calibration data could include image distortion data, and the updating step or operation could include applying the image distortion data to the projector image data, and projecting an updated image based on the projector image data. The calibration data could also include focus data, zoom data, and pan and tilt data, and the updating step could also include driving a redirection device based at least in part on the pan and tilt data. Furthermore, an additional repeating step could be provided, wherein the step of obtaining image and parameter data, the configuring step, the step of obtaining calibration data, and the updating step are repeated until calibration of the interactive system associated with the target environment is determined to be satisfactory.
As discussed above with respect to
The channel defined in the bracket can also be configured and dimensioned for passage of electrical cables in addition to (or, as discussed further below, instead of) the cooling medium. The visual unit mounting portion 904 can include a pan-tilt head portion 914 that is secured to the second region 910 of the connecting bracket 906, and can also include a visual unit receiving portion 916 that is mounted to the pan-tilt head portion 914 for pan and tilt motion.
As shown in
The aforementioned cooling medium can be, for example, ambient air. The visual unit mounting portion 904 and the second region 910 of the connecting bracket 906 can be configured to be mounted substantially above the processing unit receiving portion 902 and the first region 908 of the connecting bracket such that heat generated by the processing unit within processing unit receiving portion 902 will cause the channel of the connecting bracket 906 to function as a chimney, inducing a flow of ambient air into the at least one aperture of the processing unit receiving portion 902, through the channel 906, and out of the at least one aperture 912 that is provided in the visual unit mounting portion 904, or the second region 910 of the connecting bracket 906 (as shown in
The processing unit receiving portion 902 can be formed with sides 928, on the cabinet portion 922, for example. At least a portion of the two legs 920 can extend along the sides 928 to enhance structural rigidity of the processing unit receiving portion. The processing unit receiving portion 902, the visual unit mounting portion 904, and the connecting bracket 908 can be formed with a predetermined number of cable access points, as will be discussed below, and cable access can be substantially limited to these cable access points, for example, to reduce the chance of damage and/or tampering.
The aforementioned operation without tools and/or fasteners can be facilitated by slots 930 formed in a projecting lip 932 on portion 922. Such projecting lip 932 and slots 930 may be formed at both upper and lower areas of portion 922 as shown in
As seen in
It should be noted that the channel in bracket 906 does not always need to be dimensioned to function as a channel for flow of a cooling medium. If desired, it could be formed as a cable raceway, and its functionality could be limited to that of a cable raceway. Alternatively, it could function only as a “chimney.” However, it is believed preferable that it be configured and dimensioned to serve both functions. It will be further appreciated that the apparatus 900 is useful for other applications, for example, it could be employed to mount a plasma television with an associated integrated notebook computer.
Attention should now be given to
As best seen in
For purposes of illustrative convenience, only a single pan-tilt unit cable 947 is depicted in
Attention should now be given to
A bulb cover 974 can be provided to allow ready access to the bulb of projector 962. A rear cover/cable holder 976 can also be provided to enclose the rear portion of the projector enclosure 956, and to provide routing for cables and the like. The projector enclosure 956 can, if desired, be formed with a lower portion 978 and an upper cover portion 980. Further, the lens housing 970 can be formed, if desired, from upper half 982 and lower half 984.
It will be appreciated that various configurations or assemblies fall within the scope of the present invention. For example, the elements described herein can be provided as an assembly or as a kit of parts. Furthermore, they can be provided together with a projector 962 and a video camera 966, or without such items, which could be furnished by an end user or otherwise. Similarly, computer 924 can be included or could be provided otherwise. In one aspect, elements can be provided to form a visual apparatus for collocation with a processing unit. The visual apparatus could include the processing unit receiving portion 902 and a visual unit formed by projector 962 and camera 966 (as will be discussed more fully in connection with
Attention should now be given to
In one or more applications, interference between the optical systems of the projector 962 and the camera 966 may be significant. This may especially be so under varying lighting conditions. Such interference might result in artifacts in the image sensed by the camera 966 and, in turn, this might confound the ability to detect interactions when processing the image. For example, in bright lighting conditions, the shutter speed of the camera 966 is normally automatically increased to reduce exposure time. A very short exposure time can result in the camera sensing small changes in color of the projected image that occur over such short periods in, for example, DLP projectors such as the aforementioned XJ-450. Such sensing may cause the sensed image to artificially show sudden changes in color, which may in turn may confound the processing algorithm for detecting objects and their interactions through analysis of the sensed imaged as discussed hereinabove. In one or more exemplary embodiments of the present invention, an optical filter 988 can be provided for the lens 968 of camera 966. Such filter can reduce the intensity of light seen by the camera 966, so as to reduce or eliminate interference between the optics of the projector 962 and the camera 966. Such use of an optical filter may be applicable to camera-projector systems in general, and may enhance the effectiveness of such systems in sensing objects and interactions.
As noted above, one or more elements described herein can be provided as an assembly or as a kit of parts. By way of example and not limitation, one such kit of parts could be provided for collocating a processing unit and a visual unit. The kit of parts could include a processing unit receiving portion of the kind described above and a visual unit mounting portion of the kind described above. A connecting bracket of the kind described above could also be provided and it could include a channel configured and dimensioned for passage of cables and/or flow of a cooling medium. The parts can be designed such that, upon assembly, the channel would be in communication with the processing unit receiving portion, and also with the visual unit mounting portion and/or the second region of the connecting bracket.
It will be appreciated that the processing unit receiving portion 902 could be configured and dimensioned as desired to receive any type of processing unit. In the exemplary embodiments depicted in the figures, the processing unit receiving portion 902 is formed to receive a computer having dimensions and a form factor consistent with standard laptop computers. It will also be appreciated that the exemplary embodiment depicted in
The details of the exemplary embodiment depicted in
In view of the discussions herein, it will be appreciated that one or more exemplary embodiments of the present invention can provide one or more of the features and advantages discussed below. Interactions with the environment can be detected by one or more devices that can be dispersed throughout the environment and need not be limited to a single fixed camera tied to a projector. Further, a number of display devices can be provided, and a plurality of display zones can be provided for as well, including a physical display or displays or a moveable projector display or displays, or any desired combination thereof. Interactions to be sensed are not limited to interactions with a projected image; indeed, sensed interactions can be any place throughout the environment and are not necessarily limited to the region of a projected (or other) image. An appropriate display or displays in an appropriate display zone or zones can be activated or modified in response to actions sensed throughout the environment. Such modification can include, but is not limited to, position, orientation, content and type of display, and may also include steering and other activation activities. Coverage can be provided, if desired, throughout the subject environment.
A variety of techniques, utilizing dedicated hardware, general purpose processors, firmware, software, or a combination of the foregoing may be employed to implement the present invention. At present, it is believed that the preferred implementation will make substantial use of software running on a general purpose computer. With reference to
Accordingly, computer software including instructions or code for performing the methodologies of the invention, as described herein, may be stored in one or more of the associated memory devices (e.g., ROM, fixed or removable memory) and, when ready to be utilized, loaded in part or in whole (e.g., into RAM) and executed by a CPU. Such software could include, but is not limited to, firmware, resident software, microcode, and the like.
Furthermore, the invention can take the form of a computer program product accessible from a computer-usable or computer-readable medium (e.g., media 2018) providing program code for use by or in connection with a computer or any instruction execution system. For the purposes of this description, a computer usable or computer readable medium can be any apparatus for use by or in connection with the instruction execution system, apparatus, or device.
The medium can be an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system (or apparatus or device) or a propagation medium. Examples of a computer-readable medium include a semiconductor or solid-state memory (e.g. memory 2004), magnetic tape, a removable computer diskette (e.g. media 2018), a random access memory (RAM), a read-only memory (ROM), a rigid magnetic disk and an optical disk. Current examples of optical disks include compact disk-read only memory (CD-ROM), compact disk-read/write (CD-R/W) and DVD.
A data processing system suitable for storing and/or executing program code will include at least one processor 2002 coupled directly or indirectly to memory elements 2004 through a system bus 2010. The memory elements can include local memory employed during actual execution of the program code, bulk storage, and cache memories which provide temporary storage of at least some program code in order to reduce the number of times code must be retrieved from bulk storage during execution.
Input/output or I/O devices (including but not limited to keyboards 2008, displays 2006, pointing devices, and the like) can be coupled to the system either directly (such as via bus 2010) or through intervening I/O controllers (omitted for clarity).
Network adapters such as network interface 2014 may also be coupled to the system to enable the data processing system to become coupled to other data processing systems or remote printers or storage devices through intervening private or public networks. Modems, cable modem and Ethernet cards are just a few of the currently available types of network adapters.
In any case, it should be understood that the components illustrated herein may be implemented in various forms of hardware, software, or combinations thereof, e.g., application specific integrated circuit(s) (ASICS), functional circuitry, one or more appropriately programmed general purpose digital computers with associated memory, and the like. Given the teachings of the invention provided herein, one of ordinary skill in the related art will be able to contemplate other implementations of the components of the invention.
Although illustrative embodiments of the present invention have been described herein with reference to the accompanying drawings, it is to be understood that the invention is not limited to those precise embodiments, and that various other changes and modifications may be made by one skilled in the art without departing from the scope or spirit of the invention.
This application claims the benefit of U.S. Provisional Application No. 60/644,395, filed Jan. 15, 2005, incorporated by reference herein.
Number | Date | Country | |
---|---|---|---|
60644395 | Jan 2005 | US |