1. Field of the Invention
The invention relates generally to equipment and techniques for enabling a person to perceive graphical information using the sense of touch such as, for example, assisting visually impaired persons.
2. Description of the Related Art
Conveying information to visually impaired persons through the sense of touch is well established. For example, there are currently four commonly-utilized types of Braille text which employ unique patterns of dots that are sensed by passing one's fingertips over them and which are arranged to designate numbers, letters, and grammatical symbols. Although Braille is useful for conveying such alphanumeric information to visually impaired persons, it is not well adapted for situations in which graphical information displayable as a two-dimensional image is to be conveyed. As used herein, the term “graphical information” refers to, for example, pictures, paintings, maps, floor plans, layouts, schematic diagrams, sketches, and drawings. In many cases, it is critically important to be able to perceive the graphical information. For example, floor plans are typically utilized to show the layout of fire exits in office buildings and hotels. Visually impaired persons would benefit from having a portable, hand-held device for learning the locations of these exits, yet currently existing techniques provide no practical mechanism by which the visually impaired are able to perceive graphical information in such a way.
Many thoughts, concepts, and ideas that are easily conveyed in graphical form are relatively difficult and time-consuming to convey using linguistic forms of communication. To provide enhanced communication of graphical information, electronic techniques have been developed which enable users to communicate using two-dimensional images of hand-drawn sketches shown on a display screen of a wireless handheld device. One example of such a graphical communication technique is a service known as Scribble™ which is disclosed in U.S. provisional patent application Ser. No. 60/434,991 and International application no. PCT/EP03/13845, the content of both of which is hereby incorporated by reference. Scribble™ users communicate by updating a shared virtual page using a mobile phone with a stylus interface, or by using an Internet-connected computer with a mouse or a stylus based pointer. Although Scribble™ is well adapted to the needs of hearing and speaking impaired users by providing an expressive visual mode of communication without the use of audio, it does not allow visually impaired users to share in this genre of visually-based communication.
Since graphical information is by definition visually based, visually impaired persons cannot fully perceive such information unless it is converted into a form suitable for perception by a non-impaired sense, such as audio. The expression “visually impaired” as used herein applies not only to persons with a permanent disability but also to persons who, for example, occasionally encounter conditions which temporarily affect their ability to see normally. Converting graphical information into audio signals is problematic. While it may be possible to utilize sonic imaging techniques in order to visualize simple shapes, these techniques do not enable enjoyment or appreciation of more complex two-dimensional drawings. Moreover, they prevent the user from simultaneously participating in voice communication. Written text can be presented by providing spoken transcription of written communication processed by handwriting recognition software, but no existing software package is equipped to provide a complete spoken description of a drawing, for example.
Converting graphical information into tactile form is advantageous relative to audio-based techniques. A tactile interface allows visually impaired persons to “feel” their way through a drawing without compromising or monopolizing other human senses. An illustrative example of a tactile interface is Phantom™, available from SensAble Technologies, Inc., of Woburn, Massachusetts. Phantom™ provides a force-feedback mechanism which guides visually impaired persons as they attempt to trace lines in a drawing. The motorized interface leads users through a drawing according to the order in which the lines were drawn. This approach suffers from various drawbacks. First of all, Phantom™ is a fixed, nonportable device that must be hard-wired as a wired peripheral to a desktop computer, with corresponding weight and power requirements. Given the complexity of the device, it is expensive and, therefore, outside the domain of a mass market device.
Another haptic interface device is IBM's Tactile Trackpoint™ which provides tactile feedback to a user as a cursor is positioned. Nestled between the G, H, and B keys on many IBM notebook computers is a small projection shaped like the eraser end of a conventional pencil. This projection, termed a “TrackPoint™ pointing device”, is often referred to as a “little red button.” TrackPoint is available on the IBM ThinkPad™ line of computers, available from the IBM Corp. of Armonk, N.Y. Similar devices are now available from other manufacturers, such as NEC and Dell. More recently, IBM has improved the TrackPoint™ device through the addition of new features and modifications. For example, IBM's Tactile TrackPoint™ has an actuator beneath the TrackPoint device which can stimulate a user's finger. As the position of the cursor is changed, this actuator bumps up against the user's finger to provide tactile feedback indicative of the meaning of the cursor's present location. With Tactile Trackpoint™, a user can “feel” icons, title bars, menu items, and window borders. Such tactile feedback allows more accurate movement and selection of menu items by complementing visual feedback with a sense of touch. Unfortunately, this approach is not well suited to visually impaired persons because it is dependent upon visual observation of a changing cursor position. Moreover, displacement of the Trackpoint™ device is monitored within a tightly-confined two-dimensional space. This provides the visually impaired person with a spatial displacement which is inadequate for perceiving many graphical images.
One aspect of the present invention is directed to a technique for providing a haptic feedback for enabling a visually impaired person to mentally visualize an image that includes at least one line. Graphical information data corresponding to the image is stored. A portable, handheld device including position sensing means is manually movable by the visually impaired person through an adjacent space, for providing a position signal representative of a current position within the space to where the device is moved by the person. The stored graphical information data is compared with the position signal to generate a haptic output activation signal when the current position of the device is determined to overlap an image line. The position signal in the space is relatable to the stored graphical information. A tactile feedback mechanism in the portable device is responsive to the haptic output activation signal for providing to the person a tactile indication to create a mental image that the current position of the device is at a line of the image.
Another aspect of the present invention is directed to a technique for providing a haptic feedback for enabling a visually impaired person to mentally visualize an image that includes at least one line. The image is stored electronically. A portable, handheld device including position sensing means for providing a position signal when the device is moved by the person. A monitoring means monitors the position signal as the device is moved by the person, and a tactile feedback mechanism in the device is responsive to the monitoring means for providing to the person a tactile indication when a current position of the device corresponds to a line of the image.
Another aspect of the present invention is directed to a technique for providing a haptic feedback for enabling a visually impaired person to navigate a document composed of alphanumeric characters at respective character locations. The graphical information data corresponding to the document is stored electronically. A portable, handheld device including position sensing means is manually movable by the visually impaired person through an adjacent space, for providing a position signal representative of a current position within the space to where the device is moved by the person. A leading feedback determining means is responsive to the position signal for generating a leading feedback control signal, and a leading feedback mechanism in the device is responsive to the leading feedback control signal for providing a directional tactile indication to the visually impaired person for moving the device toward a particular one of the character locations. A tactile pad is adapted to be in contact with a finger of the person and it includes means for selectively forming a Braille counterpart of the alphanumeric character at the particular one of the character locations when the device reaches a position corresponding to the particular one of the character locations.
Another aspect of the present invention is directed to a technique for providing a haptic feedback for assisting a visually impaired person to comprehend an image that includes a plurality of line elements drawn in a particular time ordered sequence. Graphical information data corresponding to the image is stored such that the time ordered sequence in which the line elements were drawn is determinable. A portable, handheld device including position sensing means is manually movable by the visually impaired person through an adjacent space, for providing a position signal representative of a current position within the space to where the device is moved by the person. A leading feedback determining means is responsive to the position signal for generating a leading feedback control signal. A leading feedback mechanism in the device is responsive to the leading feedback control signal for providing a directional tactile indication to the visually impaired person for moving the device from the current position toward a particular image element based on the time ordered sequence in which the line elements were drawn.
Another aspect of the present invention is directed to a technique for enabling a visually impaired person to electronically draw an addition to an existing electronically reproducible image that includes at least one line. Graphical information data corresponding to the image is stored. A portable, handheld device including position sensing means is manually movable by the visually impaired person through an adjacent space, for providing a position signal representative of a current position within said space to where the device is moved by the person. A means responsive to the position signal controls a tactile feedback mechanism in the portable device for providing to the person a tactile indication to create a mental visualization of the image. A drawing control means on the device is actuatable by the person to designate any subsequently occurring position signal as line information representative of an image addition drawn by the person.
Another aspect of the present invention is directed to a technique for providing a haptic feedback for enabling a visually impaired person to mentally visualize an image that includes at least one line. A portable, handheld sensing device is adapted to be manually moved by the person through a three-dimensional virtual reality environment related to a two-dimensional image that includes one or more lines. A projection mechanism is provided for projecting the position information into the two-dimensional image. A tactile feedback mechanism is responsive to the position sensing mechanism sensing a position in the three-dimensional virtual reality environment that is projected into the one or more lines, for providing a tactile indication that the portable sensing device is passing over a line.
Since a visually impaired person cannot perceive an image by using the sense of sight, the present invention enables such person to mentally visualize the image by “feeling” the image. More specifically, a portable sensing device 40 is provided to enable the visually impaired person to “feel” where the lines are of the graphical information, displayable as a two-dimensional image, as the device 40 is manually moved relative to the image. When device 40 reaches a position that is located over a line which is part of the image, a haptic output is generated by a tactile actuation mechanism within device 40 so as to be felt by the visually impaired person as an indication of that condition. The visually impaired person visualizes in his mind's eye a blank “sheet” of paper. By manually positioning device 40 within an adjacent space within the person's reach that corresponds to that “sheet”, a certain association is established by the person in his mind's eye between the physical position of device 40 within such space and a respective visualized point on the “sheet”. Also, by manually moving device 40 within such space, the person creates a corresponding, visualized path on the “sheet”.
For illustrative purposes to facilitate the explanation of the present invention,
Point 32, as well as all other points of the graphical information shown in two-dimensional image plane 20, are defined by a set of coordinates (x,y,z) where z=0. For every point along the path of device 40 which does not intersect any line, such as point 56, the tactile actuation mechanism is not activated.
As a practical matter, it is inconvenient to require the visually impaired person to navigate a drawing while confining motion of device 40 to a particular two-dimensional image plane 20. A visually impaired person would only be able to do so if device 40 were to be moved along a flat surface. However, such a flat surface may not always be readily available. It is preferable to allow the visually impaired person to freely move device 40 through a three-dimensional space within reach. Therefore, it is necessary to convert the motion of device 40 in a three dimensional environment so that the positions are projected into the two-dimensional plane in which the graphical information is imaged.
Illustratively, portable sensing device 40 can provide x,y,z coordinates for each discreet position reached as device 40 is moved by the visually impaired person through environment 30. Thus, if such motion moves from plane 21 to plane 22, let us presume that one point where the image is crossed is at point 32A in plane 21 having the coordinates xa, zb, zd. The next crossing point is shown as 38 in plane 22 and having the coordinates xg,yh, ze. In this particular example, the crossing points 32A and 38 in planes 21 and 22 are then projected into two-dimensional image plane 20 simply by setting z=0. Thus, point 32A when projected from environment 30 into plane 20 has the coordinates xa,yb, whereas point 38 when projected from environment 30 into plane 20 (see point 38 in
The graphical information data displayable as images are stored in a database as, for example, pixel-by-pixel locations with their x, y coordinates specified in two-dimensional image plane 20. The tactile actuation mechanism of portable sensing device 40 is activated when the portable sensing device intersects, or overlaps, any point in three-dimensional virtual environment 30 that is projected into the one or more lines 52, 54, thereby communicating to the visually impaired person that portable sensing device 40 is located over a line. This is determined by comparing the current position of device 40, as projected into image plane 20, with the locations stored in the database. When the comparison results in a match, a crossing point has been identified and the haptic output is generated.
From an intuitive standpoint, one way to understand how portable sensing device 40 allows a visually impaired user to be able to “feel” two-dimensional drawings in three-dimensional virtual environment 30 is to regard the two dimensional image in plane 20 as having, in essence, “walls” projecting outwardly from lines 52, 54 in the drawing. As the user waves portable sensing device 40 in three-dimensional virtual environment 30, the user is essentially “scanning” for the drawing. When the user passes portable sensing device 40 past a “wall” projecting from the drawing into three-dimensional virtual environment 30, the device 40 will present a haptic texture for the width of the drawn surface. Portable sensing device 40 creates a virtual environment that may be perceived by the user as a virtual brush, canvas, and easel. The intent is to mimic as closely as possible the experience that an artist would experience when drawing on a canvas placed vertically on an easel. This is important. Although conceivably the image could be placed horizontally with the “walls” projecting upwardly, the particular type of device 40 described in detail below as using a gyroscopic mouse operates with the image being vertical and the “walls” projecting laterally therefrom. This is because such device provides x and y outputs, but not a z output. Portable sensing device 40 is preferably designed to work in a wide variety of operating environments, including wired, wireless, fixed, and mobile.
File server 117 is programmed with three-dimensional to two-dimensional projection software 46. File server 117 uploads the three-dimensional position information from portable sensing device 40 and projects it, as described above, into the two-dimensional image plane. An illustrative uploading mechanism utilizes a first wireless/optical communications port 125 associated with file server 117 and a second wireless/optical communications port 48 associated with portable sensing device 40. File server 117 uploads position information transmitted by the first wireless/optical communications port 48 of portable sensing device 40 and received at the second wireless/optical communications port 115. Although the configuration of
Pursuant to one illustrative embodiment of the invention, portable sensing device 40 is implemented using a 3-D gyroscopic mouse (see details provided below) equipped with a tactile actuation mechanism 10 in the form of a vibrating motor. Position sensing unit 44 may be implemented using gyroscopes, accelerometers, and/or distance sensors to perform three-dimensional position sensing in three-dimensional virtual environment 30. Accordingly, position sensing unit 44 may provide three-dimensional position information directly, or may provide derivatives (such as angular velocity) that are integrated in software to obtain three-dimensional position information. Position is determined with reference to the body of a user who may, but need not, be a visually impaired person. The user is assumed to be holding portable sensing device 40 in his or her hand.
One example of a suitable 3-D gyroscopic mouse is known as the Gyrations Ultra Cordless Optical Mouse™, available from Gyration, Inc, of Saratoga, Calif. Further details regarding this device are disclosed in U.S. Pat. No. 5,440,326, U.S. Pat. No. 5,898,421, and U.S. Pat. No. 5,825,350. For implementing position sensing unit 44, the Gyrations Gyroscopic Mouse™ utilizes an assembly within the mouse housing that includes a dual axis gyroscope. This assembly is available from Gyration, Inc., of Saratoga, Calif. as the MicroGyro™. The gyroscopes in the MicroGyro 100™ assembly are monitored and controlled by a software driver to detect position changes in a three-dimensional virtual reality space in front of the user. More specifically, the gyroscopes use the Coriolis effect to sense rotation and angular changes about a vertical axis and a horizontal axis, thus permitting efficient mapping of three-dimensional position information to a two-dimensional image or document. The gyroscopes respond to swiveling or rotation in a horizontal plane (a plane perpendicular to a gravitational source) by outputting a value for x (refer to
In operation, a user produces swiveling or rotation in a horizontal plane around the vertical axis by holding portable sensing device 40 (
The x and y values obtained from portable sensing device 40 are compared by file server 117 against the x and y coordinates of points or pixels stored in database 101. If there is a match between an (x,y) coordinate transmitted by portable sensing device 40 and an (x,y) coordinate representing a point or pixel of a line in a two-dimensional image array, file server 117 generates a haptic output activation signal which is received by portable sensing device 40. In response to the haptic output activation signal, tactile actuation mechanism 10 of portable sensing device 40 generates an output that is perceived by the sense of touch, such as vibration, level of heat, or a mild electric shock. Optionally, the level of heat, frequency of vibration, or shock amplitude can be used to convey additional information to the user, such as a color.
Alternative implementations of position sensing unit 44 may be employed. Advantageously, two of these alternative implementations are easily embedded into a cellular or wireless telephone handset to leverage existing RF circuitry that is already present in the handset. A first alternative implementation uses distance between the handset antenna and the human body to sense position. This provides a z output. In particular, antenna signal strength may be used as an indicator of position. Assume that a user's feet are standing in a fixed position while the user is holding a telephone handset in his hand. If the user moves his arm, body or hands, the handset will measure a change in received signal strength as a result of the user's new body position. The handset can be programmed such that, when it is in a signal reception mode, it will input signal strength changes into a scrolling feature, thus providing a one-degree of freedom input to the handset. This idea may be employed in a more sophisticated way by distinguishing between signal strength changes caused by a user's body capacitively loading the antenna and signal strength changes caused by propagational variables. Similarly, the handset may be programmed with a new function that measures antenna capacitance. Changes in antenna capacitance will be presumed to relate to changes in body position, whereas changes in signal strength may be the result of changes in body position as well as changes in propagation. Changes in antenna standing wave ratio (SWR) can also be measured to determine changes in a user's body position.
A second alternative implementation of position sensing unit 44 is to utilize a diversity antenna system on an existing cellular or wireless telephone handset. The handset is equipped with two or more antennas to measure changes in signal quality and/or strength which may be the result of changes in the position of a user's body. Relative changes in signal strength between two or more antennas can be used to predict changes that have occurred in the environment. If a handset has been moved or rotated relative to the user's body, the signal strength and signal-to-noise ratio of signals captured by each of the antennas will change. By positioning two antennas at different locations on the handset, the antennas will be equipped to measure one, two, or possibly three dimensions of motion. Received signal strength at both antennas will be less influenced by body proximity as the phone is moved further and further away from the body. By measuring capacitive loading and/or standing wave ratio of each antenna, these diversity antennas can be used to measure position, distance and rotation of the handset relative to the user. This is the basis of an input device that could be used for drawing or selecting functions on a cellular or wireless telephone handset.
Portable sensing device 40 includes a processing mechanism 42, illustratively implemented using a microcontroller or microprocessor, that is equipped to accept inputs from the position sensing unit 44. Depending upon the specific characteristics of the device used to implement position sensing unit 44, this unit may generate an analog output signal, whereupon an analog to digital converter (not shown, and can be a part of processing mechanism 42) is employed to convert generated analog position information into digital form. Processing mechanism 42 controls tactile actuation mechanism 10. Depending upon the specific characteristics of the device used to implement tactile actuation mechanism 10, this device may accept digital drive signals directly from processing mechanism 42, or processing mechanism 42 may drive this device through one or more digital to analog converters (not shown). Tactile actuation mechanism 10 is preferably placed such that it is relatively far away from position sensing unit 44, or such that it is mechanically isolated from position sensing unit 44, in order to avoid generating spurious position signals.
Tactile actuation mechanism 10 is an output mechanism, controlled by processing mechanism 42, that indicates to the user a change in texture or content at the point in three-dimensional virtual environment 30 (
Processing mechanism 42 may, but need not, accept input from an optional pushbutton 41. Pushbutton 41 is preferably positioned for index finger activation, and provides a mechanism by which software-defined functionality is activated using file server 117 and/or processing mechanism 42. In addition to pushbutton 41, portable sensing device 40 may provide other optional interfaces including one or more switches, a keypad, and one or more indicator LEDs for indicating items such as power-on, charging, and communication link status to file server 117. An optional power switch allows the unit to be switched on and off. Portable sensing device 40 may, but need not, be powered by a power supply including a rechargeable battery, charging circuitry, and protective circuitry. If a rechargeable battery is used, processing mechanism 42 may be programmed to monitor the battery and control any necessary charging functions.
Pursuant to a further embodiment of the invention, an optional leading feedback mechanism 51 provides tactile feedback to the visually impaired person so as to enable navigation of one or more lines in the order in which these lines were drawn, and so as to enable navigation of one or more characters in sequence to form a word or a sentence. Not being able to know the order in which a two-dimensional image was made, and not being able to ascertain the order in which the image should be interpreted, can be significant limitations which create unnecessary ambiguities for a visually impaired person that are not present for those who are able to view the image directly. For example, with reference to
Optional leading feedback mechanism 51 (
Pursuant to the foregoing illustrative implementation of optional leading feedback mechanism 51, a tracking ball is motor-driven so as to permit a user to feel the direction in which the ball spins. This arrangement is somewhat similar to what is used in the type of computer mouse which has a hand-operated tracking ball, but instead of including rotational sensors to monitor the hand-operated rotation of the tracking ball about two mutually orthogonal axes, leading feedback mechanism 51 utilizes two motors to spin the tracking ball about two mutually orthogonal axes. The role of the tracking ball has been changed from that of a passive sensor to that of an active output device to direct the user in a particular direction. The tracking ball is preferably mounted within portable sensing device 40 such that a portion of the surface of the ball is exposed through the housing of portable sensing device 40. In this manner, the pad of one finger, such as the middle finger, of the user can contact the ball. The housing of portable sensing device 40 is preferably fabricated in such a way that a surface of the tracking ball is beneath the middle finger when the device is held in the user's hand.
The housing of portable sensing device 40 should be shaped such that the device can be held securely by the palm and fingers while, at the same time, not requiring the palm or the fingers to contact pushbutton 41 or the tracking ball.
Preferably, the tracking ball of optional leading feedback mechanism 51 should be exposed through the housing of portable sensing device 40 such that the tracking ball presents a surface of approximately 5-10 millimeters in diameter, or roughly the size of the meaty portion of an average adult human finger. The surface material and profile of the tracking ball is preferably selected to provide relatively high traction on the user's fingertip, while being resistant to contamination from skin products such as oils produced by the skin as well as skin lotions and dirt.
In cases where optional leading feedback mechanism 51 is employed, tactile actuation mechanism 10 may (but need not) be combined with leading feedback mechanism 51. For example, tactile actuation mechanism 10 may be implemented by vibrating the tracking ball, or by vibrating a separate part of the housing of portable sensing device 40 that contacts a finger or at least a portion of the palm of the user's hand.
Operation of portable sensing device 40 typically involves a user holding the device in his/her hand and moving it around in three-dimensional virtual environment 30 (
Next, a three-dimensional virtual environment is generated based on the two-dimensional image array (
Position information is received from the portable sensing device 40 at block 406. At block 408, the received position information is mapped into the two-dimensional image array. A test is performed at block 410 to ascertain whether the mapped position intersects, i.e. is located on, or sufficiently near to, a line in the two-dimensional image array. As explained above, this involves comparing the x and y values obtained from portable sensing device 40 against the x and y coordinates of points or pixels stored in database 101. If there is a match between an (x,y) coordinate transmitted by portable sensing device 40 and an (x,y) coordinate representing a point or pixel of a line in a two-dimensional image array, file server 117 generates a haptic output activation signal which is received by portable sensing device 40 (block 412), and the program loops back to block 406. The negative branch from block 410 loops back to block 406.
In order to enable generation of a leading feedback signal, it is necessary to store the graphical information data representing the image in a time ordered manner. Each image element is stored in the order it was drawn. This makes it possible for step 602 to retrieve each element in its proper place in this time ordered sequence in order to provide the leading feedback signal. As used herein, the term “image element” applies to points, lines, line segments and any other portion of an image that lends itself to being stored, retrieved and processed as a unit in connection with implementing the present invention.
The negative branch from block 610 leads to block 616 where a leading vector feedback signal is generated by calculating a vector from the mapped position to at least one of: (a) a nearest projected line, and (b) a successive portion of the most recently overlapped line, determined with reference to the order in which the line was drawn. At block 618, a leading vector feedback signal is transmitted to the portable sensing device. The program then loops back to block 606.
At block 711, a test is performed to ascertain whether the portable sensing device 40 receives a leading vector feedback signal from the file server 117. If so, the leading feedback mechanism 51 is activated to provide feedback representing a vector leading to at least one of: (a) a successive portion of the currently overlapped line, determined with reference to the order in which the line was drawn, (b) a nearest projected line, and (c) a successive portion of the most recently overlapped line, determined with reference to the order in which the line was drawn (block 713). The program then loops back to block 703. The negative branch from block 711 leads to block 703.
The invention can be employed to enable a person to perceive graphical information using the sense of touch such as, for example, when that person's eyesight is otherwise engaged or when sight is not possible, as when it is too dark to see. One advantageous application of the invention is to enable a visually impaired person to perceive graphical information in the form of an electronic text document that contains any of alphabetic, numeric and punctuation characters, by using the sense of touch. As recognized above, Braille has long been available to communicate alphanumeric characters to visually impaired persons. However, with the increased proliferation of text in electronic form, it would be highly advantageous to readily communicate the text in electronic form to the visually impaired person without the cumbersome intermediate task of statically printing it to Braille. The present invention can achieve that by providing a dynamic, mobile Braille-like interface, as follows.
The text document is readily converted electronically into a file in which each character is associated with the x, y coordinates in a two-dimensional image plane that are associated with it. This is stored in database 101 just as any other image of graphic information, as disclosed above. Position information in three-dimensional virtual environment 30 is gathered by portable sensing device 40 (
Tactile pad 63 is preferably stationary and is coupled to sensor 117.
Pursuant to a further embodiment of the invention, leading feedback mechanism 51 provides tactile feedback to the visually impaired person so as to enable navigation within the text document to proceed through a plurality of characters in sequence to form one or more words. Illustratively, the leading feedback mechanism 51 can be used which, as described above, is incorporated into portable sensing device 40 in the form of a motor driven tracking ball that is rotated in a direction so as to guide the visually impaired person toward one or more characters in three-dimensional virtual environment 30. If portable sensing device 40 intersects a point in three-dimensional virtual environment 30 that is projected into a character which, for example, is the first letter “W” of the word “WIND”, the leading feedback mechanism is rotated in a direction to lead the user toward the immediately successive character “I” in the sequence of characters forming the word (or it can be the next word in a sentence). Alternatively, the other above-described leading feedback mechanisms can also be used.
Pursuant to a further embodiment of the invention, server 117 can be provided with at least one of handwriting recognition software and optical character recognition software for interpreting two-dimensional image files generated by scanning a printed or handwritten document. Such interpretations may be conveyed to visually impaired persons using an optional electronic text-to-speech synthesizer or tactile pad 63. This approach is useful, for example, when the two-dimensional image file is a map that includes roads marked with alphanumeric labels.
Pursuant to a further embodiment of the invention, pushbutton 41 of portable sensing device 40 is arranged with a for selecting one of a plurality of operational modes including read mode and drawing mode. In read mode, portable sensing device 40 provides tactile output as described previously in connection with at least one of lines and characters. However, the two-dimensional image 20 (
Pursuant to a further embodiment of the invention, portable sensing device 40 is equipped with an orientation sensor for determining the orientation of a reference plane of portable sensing device 40 relative to an orientation of a reference plane in three-dimensional virtual environment 30. The determined orientation is used to calculate a calibration factor for application to a sensed position, so as to enable a user to hold portable sensing device 40 in any of a plurality of rotational orientations within the three-dimensional virtual environment.
Pursuant to a further embodiment of the invention, portable sensing device 40 is used as a control mechanism for navigating menu systems that are laid out virtually in front of the user, so the user can manipulate systems without a visual interaction. This feature is advantageous for use in low light situations or, for example, situations in which the user's visual sense is otherwise occupied with another task.
Pursuant to a further embodiment of the invention, tactile actuation mechanism 10 (
Although various embodiments of the present invention have been described in detail above, various modifications thereof will readily occur to anyone with ordinary skill in the art. For example, instead of the rotating ball used in leading feedback mechanism 51, a mechanically actuated tilting paddle can be used which tilts in a direction so as to guide the visually impaired individual. Also, the position of the rotating ball can be under the thumb rather than the middle finger.
Thus, while there have been shown and described features of the invention as applied to particular embodiments thereof, it will be understood that various omissions and substitutions and changes in the form and details of the devices illustrated, and in their operation, may be made by those skilled in the art without departing from the spirit of the invention. For example, it is expressly intended that all combinations of those elements and/or method steps which perform substantially the same function in substantially the same way to achieve the same results are within the scope of the invention. Moreover, it should be recognized that structures and/or elements and/or method steps shown and/or described herein in connection with any disclosed form or embodiment of the invention may be incorporated in any other disclosed or described or suggested form or embodiment as a general matter of design choice. It is the intention, therefore, to be limited only as indicated by the scope of the claims appended hereto.