Other objects and features of the present invention will become apparent from the following detailed descriptions considered in conjunction with the accompanying drawing. It is to be understood, however, that the drawings are designed solely for the purposes of illustration and not as a definition of the limits of the invention.
In the drawings, like numbers refer to like objects throughout. The pointing device described is held and operated by a user, not shown in the drawings. The user can communicate by means of a suitable user interface, also not shown in the drawings, with the device for interacting with the pointing device.
The received image data 3 are analysed in the image analysis unit 6 of the interacting device 13, where they are compared to other images or templates retrieved from an internal memory 20 or an external source 21, 22 by an accessing unit 19. Ideally, the accessing unit 19 has a number of interfaces allowing access to external data, for example the user might provide pre-defined templates stored on a memory medium 21 such as floppy disk, CD or DVD, or the accessing unit 19 might retrieve suitable template information from an external network such as the internet 22. The templates may also be configured by the user, for example in a training session in which the user specifies the correlation between specific areas on a template with particular items or functions.
The user in this case may be trying to locate an item, so that the image analysis unit 6 compares the image data 3 with the templates to determine whether the item sought is within the target area or not, and directs a control signal generator 8 to generate appropriate control signals 9, which are transmitted by a sending unit 11 of the interacting device 13 in a wireless manner to a communication interface 7 of the pointing device 1.
A laser light source 12, incorporated in the pointing device 1, emits a beam of laser light L in a direction not necessarily parallel to the direction of pointing D. The actual direction of the beam of laser light L is controlled by a directing arrangement 4 which applies the received control signals 9 to adjust the direction of pointing of the laser light source 12. The light point is directed in such a way that the user is eventually guided to the item being sought. In the figure, the directing arrangement 4 applies the control signals 9 to alter the position of the laser light source 12 accordingly, by means of, for example, a miniature motor. The beam of laser light L is thus aimed in the desired direction. In an alternative realisation, the directing arrangement 4 may comprise a number of small mirrors, whose position can be altered, and arranged in such a way that the mirrors deflect the beam of laser light L in the required direction. It is also feasible that a combination of miniature motor and mirrors might be used to control the direction of the beam of laser light L.
In another scenario, the pointing device 1 is being used to select an item, for example when training the interacting device to recognise and locate items. In this case, image data 3 is generated by aiming the pointing device at the item to be recognised, and is sent to the image analysis unit 6 to be analysed and processed in some way before being stored in a suitable format in the internal or external memories 20, 21.
In another application, the interacting device 13 features an interface 24 for communicating with an external device 25 such as a television, VCR, or any type of device with which a dialog might be initiated. Here, the interacting device 13 informs the external device 25 in some way of the user's actions. For example, the image analysis unit 6 determines, with the aid of templates for the options of this device 25, the area in the template at which the user is pointing, and sends this information to the external device 25, which interprets the information and send appropriate signals to the interacting device, where they are converted to control signals 9 for the directing arrangement 4 of the pointing device 1. In this way, the pointing device 1 together with the interacting device 13 can be used to assist the user in controlling or communicating with external devices 25.
If the user wants to locate an object, for example the item M4, he might request an interacting device (not shown in the drawing) to assist him in locating it. Images 16 of the target area T are transmitted at intervals to the interacting system, where they are analysed to determine the area at which the pointing device 1 is aimed, and whether this area contains the item M4 being sought. As long as this item M4 cannot be detected in the image 16 of the target area T, the light source 12 of the pointing device 1 is directed by means of control signals so that the ensuing light point PL is moved in such a way as to indicate to the user the direction in which he must aim the pointing device 1 so that the item M4 can ultimately be detected in the image 16 of the target area T, at which stage the light point PL is positioned over the desired item M4 to show the user where it is. In the event that the system cannot locate an item because the item is missing or because the system is unable to understand the user's wishes, the light point PL might behave in a predefined manner e.g. by being turned on and off in a particular sequence, or by describing a predefined pattern. This would be of use, when, for example, the interacting device is unable to communicate with user by means of speech.
Should the user wish to select one of the items M1, M2, M3, M4 visible in the visual presentation VP, for example when training the interacting device to remember the locations of objects or to recognise their appearance, the user can aim the pointing device 1 at the visual presentation VP so that the object in question is indicated by the light point PL. When the pointing device is being used in such a training mode, the light point PL can maintain a fixed position relative to the centre of the target area A, given by PT. The light point PL might be directed at a fixed position at a point removed from the centre point PT or it might coincide with the centre point PT.
With the aid of the light point PL, the user can select one of the items M1, M2, M3, M4 shown in the visual presentation VP. A camera in the pointing device generates an image of the target area T centred around an image centre point PT. The light point PL also appears in the target area image. In this example, the light point PL appears at a very small distance away from the image centre point PT, so that the user can use the light point PL to accurately point out items to the interacting device, in this case the item M3. The user then describes the object M3 for the interacting device, for example by saying “This book is ‘Middlemarch’ by George Eliot”, so that the interacting device performs any necessary image processing before storing the information describing the item M3 to memory.
The interacting device 13 is in this example might be incorporated in some kind of home dialog system, allowing the user to communicate with it by means of spoken commands. For example, the user has asked the interacting device 13 a question, such as “Where is my Dire Straits CD ‘Money for Nothing’?”. The user aims the pointing device 1 in the general direction of the shelves on which his CD collection is kept, and allows the interacting device 13, in conjunction with the pointing device 1, to show him where the requested CD is kept. The interacting device 13, which has been trained in a previous training session to remember the locations of all the CDs in the collection, now sends control signals to the directing arrangement of the pointing device 1 so that the light point PL is directed at the requested CD. If the requested CD is located within the target area T, the light point comes to rest on this CD, or might be caused to describe a tight circle over the CD. However, if the CD is outside of the target area T, the control signals issued by the interacting device 13 cause the light point PL to repeatedly move against the appropriate edge of the target area T, so that the user will realise that he must move the pointing device 1 in the indicated direction until the target area T includes the requested CD.
In this embodiment, the pointing device 1 also features a button 15. The button 15 can be pressed by the user, for example to confirm that he has made a selection and to record the image of the target area.
Alternatively or additionally, such a button 15 might be used to activate or deactivate displaying of a dynamic visual presentation VP′ on, for example, a television screen, so that items or options are only displayed on the screen when actually required by the user. Alternatively, the function of the button 15 or a different button on the pointing device 1 might be to activate or deactivate the light source 12 incorporated in the pointing device 1, to activate or deactivate the pointing device 1 itself, or to switch between “locate” and “select” modes of operation. The pointing device 1 might be activated by means of a motion sensor incorporated in the pointing device 1, so that the laser light source is activated when the user takes hold of the pointing device 1, and that the pointing device starts to send images of the target area to the interacting device as soon as it the pointing device is taken up or moved.
The pointing device 1 draws its power from one or more batteries, not shown in the figure. Depending on the consumption of the pointing device 1, it may be necessary to provide a cradle into which the pointing device 1 can be placed when not in use, to recharge the batteries.
The user will not always aim the pointing device at right angles to the visual presentation—it is more likely that the pointing device will be aimed at a more or less oblique angle to the visual presentation, since it is easier to wave the pointing device that it is to change one's own position. This is illustrated in
Regardless of the angle of the pointing device 1 with respect to the visual presentation VP′, the target area image 16 is always centred around an target point PT. The laser point PL also appears in the target area image 16, and may be a distance removed from the target point PT, or might coincide with the target point PT. The image processing unit of the dialog system compares the target area image 16 with pre-defined templates to determine the item being pointed at the user, or to determine the location of the target point relative to the location of an item which the user is trying to locate.
To this end, the point of intersection PT of the longitudinal axis of the pointing device 1 with the visual presentation VP′ is located. The point in the template corresponding to the point of intersection PT can then be located. Computer vision algorithms using edge- and corner detection methods are applied to locate points in the target area image [(xa, ya), (xb, yb), (xc, yc)] which correspond to points in the template [(xa′, ya′), (xb′, yb′), (xc′, yc′)] of the visual presentation VP′.
Each point can be expressed as a vector e.g. the point (xa, ya) can be expressed as {right arrow over (v)}a. As a next step, a transformation function Tλ is developed to map the target area image to the template:
where the vector {right arrow over (v)}i represents the coordinate pair (xi, yi) in the target area image, and the vector {right arrow over (v)}i′ represents the corresponding coordinate pair (x′i, y′i) in the template. The parameter set λ, comprising parameters for rotation and translation of the image yielding the most cost-effective solution to the function, can be applied to determine the position and orientation of the pointing device 1 with respect to the visual presentation VP. The computer vision algorithms make use of the fact that the camera 2 within the pointing device 1 is fixed and “looking” in the direction of the pointing gesture. The next step is to calculate the point of intersection of the longitudinal axis of the pointing device 1 in the direction of pointing D with the plane of the visual presentation VP. This point may be taken to be the centre of the target area image PT. Once the coordinates of the point of intersection have been calculated, it is a simple matter to locate this point in the template of the visual presentation VP.
Although the present invention has been disclosed in the form of preferred embodiments and variations thereon, it will be understood that numerous additional modifications and variations could be made thereto without departing from the scope of the invention. The pointing device can serve as the universal user interface device in the home or for navigation through business presentations. Outside of the home, it can be used in any environment where the user can be guided by means of the light point. In short, it can be beneficial wherever the user can express an intention by pointing, or wherever something can be actively pointed out to the user. Its small form factor and its convenient and intuitive usage can elevate such a simple pointing device to a powerful universal remote control or teaching tool. As an alternative to the pen shape, the pointing device could for example also be a personal digital assistant (PDA) with a built-in camera, or a mobile phone with a built-in camera. The pointing device might be combined with other traditional remote control features, e.g. with additional buttons for performing dedicated functions, or with other input modalities such as voice control.
For the sake of clarity, it is also to be understood that the use of “a” or “an” throughout this application does not exclude a plurality, and “comprising” does not exclude other steps or elements. A “unit” may comprise a number of blocks or devices, unless explicitly described as a single entity.
Number | Date | Country | Kind |
---|---|---|---|
04103527.0 | Jul 2004 | EP | regional |
Filing Document | Filing Date | Country | Kind | 371c Date |
---|---|---|---|---|
PCT/IB05/52353 | 7/15/2005 | WO | 00 | 1/18/2007 |