The invention can be more fully understood by reading the subsequent detailed description and examples with references made to the accompanying drawings, wherein:
Several exemplary embodiments of the invention are described with reference to
The invention discloses a space positioning and directing input system and processing method therefor.
An embodiment of a space positioning and directing input system is a system comprising a human machine interface, and a space positioning and directing input device. The space positioning and directing input system receives light (the light from an active light source or passively reflected light) emitted by a positioned object (i.e. a space positioning and directing input device) using a sensor, processes detected data by the sensor using an operation device to reversely calculate three-dimensional (3D) or two-dimensional (2D) projection coordinates, and generates other information, such as velocity, acceleration, depression operations, and so forth, according to the calculated coordinates or detection information, such as from moving a positioned object or depressing a preset button.
The following illustrates embodiments of a space positioning and directing input system and processing method therefor.
An embodiment of employed components are first described and symbolized respectively.
Imaging display devices (symbolized by d1, d2 . . . ) indicate computer monitors, personal digital assistants (PDA), cellular phones, TV monitors, and so forth.
Image sensors (symbolized by s1, s2 . . . ) indicate image input devices comprising charge coupled devices (CCD) sensors, complementary metal-oxide semiconductor (CMOS), and so forth.
Imaging pictures (symbolized by i1, i2 . . . ) indicate pictures detected by image sensors (s1, s2 . . . ).
Imaging positions of objects on a monitor (symbolized by p1, p2 . . . ) indicate shape centers of gravity, geometric centers of gravity, or 2D coordinates of a point representing an object that is imaged on an imaging display device.
Light sources (symbolized by l1, l2 . . . ) indicate visible light, infrared (IR) rays, ultraviolet (UV) rays, and so forth.
Positioned objects (symbolized by o1, o2 . . . ) indicate space positioning and directing input devices. It is noted that a positioned object represents a space positioning and directing input device, which will not be further explained.
Reflection devices (symbolized by r1, r2 . . . ) indicate reflective structures and special shapes or textures composed of reflective structures.
Imaging display device d1 (a monitor, for example) comprises image sensors s1 and s2 and an operations device c1. Positioned object o1 (a joystick of a TV game, for example) comprises light source l1. Light source l1 of positioned object o1 first emits light while image sensors s1 and s2 receives the light to generate two imaging pictures i1 and i2, as shown in
As described, 3D space coordinates of imaging positions at different time points corresponding to light source l1 can be calculated and other imaging information (such as velocity, acceleration, depression operations, and so forth) can thus be generated by depression operation recognition.
Imaging display device d1 (a monitor, for example) comprises light source l1, image sensors s1 and s2, and operation device c1. Positioned object o1 (a joystick of a TV game, for example) comprises reflection device r1. Light source l1 first emits light, reflected by reflection device r1, while image sensors s1 and s2 receives the reflective light to generate imaging pictures i1 and i2, as shown in
Imaging display device d1 (a monitor, for example) comprises image sensors s1 and s2 and operation device c1. Positioned object o1 (a joystick of a TV game, for example) comprises light source l1. Positioned object o2 (a joystick of a TV game, for example) comprises light source l2. Light sources l1 and l2 emits light while image sensors s1 and s2 receive the light to generate imaging pictures i1 and i2, as shown in
Next, operation device c1 corresponds imaging positions p1(i1), p2(i1) and p1(i2), and p2(i2) to imaging positions p1(l1), p2(l1), p1(l2), and p2(l2), respectively using a correspondence method. When imaging pictures i1 or i2 corresponding to light source l1 and l2 overlaps, operation device cl calculates imaging positions p1(l1), p2(l1), p1(l2), and p2(l2) using an Epipolar method and calculates 3D space coordinates of an imaging position at a time point corresponding to light sources l1 and l2 using a triangulation method.
As described, 3D space coordinates of imaging positions at different time points corresponding to light sources l1 and l2 can be calculated and other imaging information (such as velocity, acceleration, depression operations, and so forth) can thus be generated by depression operation recognition or a labeling method.
It is noted that 3D space coordinates of imaging positions at different time points corresponding to light sources (l1, l2, . . . , ln) for multiple positioned objects (o1, o2, . . . , on) can also be obtained using the described method, thereby obtaining other imaging information, such as velocity, acceleration, depression operations, and so forth.
Imaging display device d1 (a monitor, for example) comprises light source l1, image sensors s1 and s2, and operation device c1. Positioned object o1 (a joystick of a TV game) comprises a reflection device r1 while positioned object o2 (a joystick of a TV game) comprises a reflection device r2. Light source l1 emits light while image sensors s1 and s2 receive the reflective light from reflection devices r1 and r2 to generate imaging pictures i1 and i2, as shown in
As described, 3D space coordinates of imaging positions at different time points corresponding to reflection devices (r1, r2, . . . , rn) for multiple positioned objects (o1, o2, . . . , on) can also be obtained using the described method, thereby obtaining other imaging information, such as velocity, acceleration, depression operations, and so forth.
Imaging display device d1 (a monitor, for example) comprises image sensor s1 and operation device c1. Positioned object o1 (a light pen, for example) comprises light source l1. Light source l1 emits light while image sensor s1 receives the light to generate imaging picture i1, as shown in
2D coordinates of imaging positions at different time points corresponding to light source l1 can be calculated and other imaging information (such as velocity, acceleration, depression operations, and so forth) can thus be generated by depression operation recognition.
Imaging display device d1 (a monitor, for example) comprises light source l1, image sensor s1, and operation device c1. Positioned object o1 (a light pen, for example) comprises reflection device r1. Light source l1 emits light while image sensor s1 receives the reflective light from reflection device r1 to generate imaging picture i1, as shown in
Imaging display device d1 (a monitor, for example) comprises image sensors s1 and operation device c1. Positioned object o1 (a light pen, for example) comprises light source l1. Positioned object o2 (a light pen, for example) comprises light source l2. Light sources l1 and l2 emits light while image sensors s1 and s2 receive the light to generate imaging picture i1, as shown in
2D coordinates of imaging positions p1 and p2 at different time points can be calculated using a labeling method and other imaging information (such as velocity, acceleration, depression operations, and so forth) can thus be generated by depression operation recognition.
It is noted that 2D coordinates of imaging positions at different time points corresponding to light sources (l1, l2, . . . , ln) for multiple positioned objects (o1, o2, . . . , on) can also be obtained using the described method, thereby obtaining other imaging information, such as velocity, acceleration, depression operations, and so forth.
Imaging display device d1 (a monitor, for example) comprises light source l1, image sensor s1, and operation device c1. Positioned object o1 (a light pen, for example) comprises reflection device r1. Positioned object o2 (a light pen, for example) comprises reflection device r2. Light source l1 emits light while image sensor s1 receives the reflective light from reflection device r1 to generate imaging picture i1, as shown in
It is noted that 2D coordinates of imaging positions at different time points corresponding to reflective light from reflection devices (r1, r2, . . . , m) for multiple positioned objects (o1, o2, . . . , on) can also be obtained using the described method, thereby obtaining other imaging information, such as velocity, acceleration, depression operations, and so forth.
Imaging display device d1 (a PDA, for example) comprises image sensor s1 and operation device c1. Positioned object o1 (a light pen, for example) comprises light source l1. Light source l1 installed on a plane emits light while image sensor s1 receives the light to generate imaging picture i1, as shown in
Next, operation device c1 generates a converted imaging position pp1 displayed on display device d1 using a 2D grid positioning method and other imaging information (such as velocity, acceleration, depression operations, and so forth) can thus be generated by depression operation recognition based on the time variation of imaging position pp1.
Imaging display device d1 (a PDA, for example) comprises light source l1, image sensor s1, and operation device c1. Positioned object o1 (a light pen, for example) comprises reflection device r1. Light source l1 emits light while image sensor s1 receives the reflective light from reflection device r1 installed on a plane to generate imaging picture i1, as shown in
Imaging display device d1 (a PDA, for example) comprises image sensor s1 and operation device c1. Positioned object o1 (a light pen, for example) comprises light source l1. Positioned object o2 (a light pen, for example) comprises light source l2. Light sources l1 and 12 installed on a plane emit light while image sensor s1 receives the light to generate imaging picture i1, as shown in
Next, operation device c1 generates converted imaging positions pp1 and pp2 displayed on display device d1 using a 2D grid positioning method. 2D coordinates of imaging positions pp1 and pp2 at different time points can be calculated using a labeling method and other imaging information (such as velocity, acceleration, depression operations, and so forth) can thus be generated by depression operation recognition.
It is noted that imaging positions (p1, p2, . . . , pn) on imaging picture i1 corresponding to light sources (l1, l2, . . . , ln) for multiple positioned objects (o1, o2, . . . , on) can also be calculated using the described method, and 2D coordinates of converted imaging positions (pp1, pp2, . . . , ppn) displayed on display device d1 can be further calculated using a 2D grid positioning method.
Imaging display device d1 (a PDA, for example) comprises light source l1, image sensor s1, and operation device c1. Positioned object o1 (a light pen, for example) comprises reflection device r1. Positioned object o2 (a light pen, for example) comprises reflection device r2. Light source l1 emits light while image sensor s1 receives the reflective light from reflection devices r1 and r2 to generate imaging picture i1, as shown in
Next, the process of calculating imaging positions p1 and p2 on imaging picture i1 corresponding to the reflective light from reflection devices r1 and r2 and obtaining converted imaging positions pp1 and pp2 displayed on display device d1 using a 2D grid positioning method is identical to that described in the eleventh embodiment, as such will not be further described.
It is noted that imaging positions (p1, p2, . . . , pn) on imaging picture i1 corresponding to reflection devices (r1, r2, . . . , m) for multiple positioned objects (o1, o2, . . . , on) can also be calculated using the described method, and 2D coordinates of converted imaging positions (pp1, pp2, . . . , ppn) displayed on display device d1 can be further calculated using a 2D grid positioning method.
The following describes the object extraction method, the labeling method, the correspondence method, the triangulation method, the 2D grid positioning method, and the depression operation recognition.
The object extraction provides a thresholding method, also named object and background segmentation method. With respect to IR ray of invisible light, for example, only the object itself (representing an active light source or comprising a light reflection portion) of the input image shines while other areas of the input image represent the background and show a black color. A traced object and the background of such input image can be separated, comprising the following.
An input image is first divided to pixels belonging to an object and that belonging to the background according to a predetermined fixed threshold. The process can be accurately implemented when the threshold is calculated using an Otsu method. Next, pixels belonging to the traced object are connected to form an object using connected component labeling (CCL).
With respect to a nature light source, for example, the traced object can be represented using a specified color or pattern to be discriminated from the background. Suppose the background is a white wall and the traced object shows the red color, the traced object and background can be easily discriminated based on the color. The position and scope of the traced object is located using CCL.
With respect to the correspondence method, objects pictured using two sensors are extracted and correspondence between the objects of the two image frames is obtained. In this embodiment, the correspondence between the objects is obtained according to shapes (as shown in
The triangulation method positions target objects in the space using two camera systems. Internal and external parameters for a camera system define K1, R1, t1, K2, R2, and t2 respectively. Ri and ti represent a rotation matrix and a translation vector, respectively.
where s is a skew value, (u0,v0) is an optical center, f is a focal length, and α is an aspect ratio.
A point X in the space is projected in the two camera systems, generating points x1 and x2, projection relationship thereof is described as:
where plane coordinates are represented in homogenous coordinates.
Further,
space coordinates are also represented in homogenous coordinates.
As the internal parameters of a camera are known, space coordinates of a point can be calculated according to pictured projection points, i.e. x1{circle around (x)}K1[R1|t1]X=0 and x2{circle around (x)}K2[R2|t2]X=0.
With respect to the 2D grid positioning method, as shown in
Transforming a coordinate position x from grid 2 to grid 3 represents a plane transformation, represented as a formula in the following, in which H represents a 3×3 matrix:
where x is a point of grid 2 and x′ is the corresponding point of x of grid 3.
The matrix is spread as:
The described formulas are further transposed as:
When plane transformation is implemented with wi=1, wi=1 and H33=1, 8 unknown elements are generated. When 4 corresponding points are provided, for instance A2, B2, D2, and E2 in
With respect to the depression operation recognition, when fast moving, disappearance, rotation, shape variation, color variation, violent gray level variation, texture variation, object number variation, or the combinations occur to detect objects imaged on imaging pictures i1 and i2, a button depression operation is thus activated. Other kinds of variations can also act as different button depressions or equivalent behaviors, such as the up-down and left-right movement of a joystick, for example.
Referring to
The space positioning and directing input device can be, but is not limited to, a joystick of a TV game or a light pen. Additionally, when only one image sensor is installed on the display device, an imaging position corresponding to the light source of the space positioning and directing input device and 2D coordinates of the imaging position corresponding to the light source are calculated using the operation device, as shown in
Referring to
The space positioning and directing input device can be, but is not limited to, a joystick of a TV game or a light pen. Additionally, when only one image sensor is installed on the display device, an imaging position corresponding to the reflection device and 2D coordinates of the imaging position corresponding to the reflection device are calculated using the operation device, as shown in
Referring to
The first and second space positioning and directing input device can be, but are not limited to, joysticks of a TV game or light pens. Additionally, when only one image sensor is installed on the display device, imaging positions corresponding to the first and second light sources of the first and second space positioning and directing input devices and 2D coordinates of the imaging positions corresponding to the first and second light sources are calculated using the operation device, as shown in
Referring to
The first and second space positioning and directing input device can be, but are not limited to, joysticks of a TV game or light pens. Additionally, when only one image sensor is installed on the display device, imaging positions corresponding to the first and second reflection devices and 2D coordinates of the imaging positions corresponding to the first and second reflection devices are calculated using the operation device, as shown in
Referring to
When a first space positioning and directing input device and a second space positioning and directing input device are installed on the display device. Light is emitted using the first light source and the second light source of the first and second space positioning and directing input devices installed on the plane. The light is received to generate an imaging picture using the image sensor. A first imaging position and a second imaging position corresponding to the first and second light source are calculated according to imaging information of the imaging picture using the operation device. A third imaging position and a fourth imaging position on the display device are calculated according to the first and second imaging positions, as shown in
Referring to
When a first space positioning and directing input device and a second space positioning and directing input device are installed on the display device, light emitted by the light source is reflected using a first reflection device of a first space positioning and directing input device and a second reflection device of a second space positioning and directing input device installed on the display device to generate a first reflective light and a second reflective light. The first and second reflective light is received to generate an imaging picture using the image sensor. A first imaging position and a second imaging position corresponding to the first and second reflection source according to imaging information of the imaging picture and a third imaging position and a fourth imaging position on the display device according to the first and second imaging positions are calculated using the operation device, as shown in
It is noted that an embodiment of a space positioning and directing input system and processing method employ at least one space positioning and directing input device. However, two or more space positioning and directing input devices can also be employed. Additionally, while at least one sensor is applied to implement the invention, two or more may also be applied. The detailed process thereof has been described.
Methods and systems of the present disclosure, or certain aspects or portions of embodiments thereof, may take the form of a program code (i.e., instructions) embodied in media, such as floppy diskettes, CD-ROMS, hard drives, firmware, or any other machine-readable storage medium, wherein, when the program code is loaded into and executed by a machine, such as a computer, the machine becomes an apparatus for practicing embodiments of the disclosure. The methods and apparatus of the present disclosure may also be embodied in the form of a program code transmitted over some transmission medium, such as electrical wiring or cabling, through fiber optics, or via any other form of transmission, wherein, when the program code is received and loaded into and executed by a machine, such as a computer, the machine becomes an apparatus for practicing the embodiment of the disclosure. When implemented on a general-purpose processor, the program code combines with the processor to provide a unique apparatus that operates analogously to specific logic circuits.
While the invention has been described by way of example and in terms of the preferred embodiments, it is to be understood that the invention is not limited to the disclosed embodiments. To the contrary, it is intended to cover various modifications and similar arrangements (as would be apparent to those skilled in the art). Therefore, the scope of the appended claims should be accorded the broadest interpretation so as to encompass all such modifications and similar arrangements.
Number | Date | Country | |
---|---|---|---|
60832601 | Jul 2006 | US |