Recent publications have proposed combining the functionality of computer mice into cordless and wired telephone handsets in an effort to reduce workplace clutter. In these proposals, the combined computer mouse and telephone handset communicates with a computer that is connected by a fixed wire line connection to a public switched telephone network (PSTN). In a first mode of operation, the combined computer mouse and telephone handset operates as a standard wired or cordless telephone. In a second mode of operation, the combined computer mouse and telephone handset operates as a standard computer mouse capable of entering commands into the computer.
Many different data input schemes have been developed for handheld devices. For example, some mobile telephone apparatus, such as a combined mobile telephone and personal digital assistant, include a keypad that contains a minimal number of keys and a touch screen display that displays an output and receive a touch input. Software may be used to display icons on the touch screen that represent buttons or keys. A user may activate a desired function by touching the touch screen display region displaying the button corresponding to the desired function. The limited number of keys on the keypad and the small size of the touch screen make the entry of data and commands into such mobile telephones difficult. In an effort to overcome this difficulty, a mobile telephone that includes a detachable input apparatus has been proposed. The detachable input apparatus includes a coordinate information generator (i.e., a trackball) that generates coordinate information based on movement of the input apparatus on a flat surface. The coordinate information is used to control the position of a cursor on a screen of the mobile telephone.
What is needed are systems and methods of producing display control signals for a handheld device display and a remote display.
In one aspect, the invention features a handheld device that includes a display subsystem and a pointing device subsystem. The pointing device subsystem produces display control signals in response to user input. The pointing device subsystem has a handheld device mode of operation in which the pointing device subsystem selectively communicates the display control signals to the display subsystem. The pointing device subsystem also has an input device mode of operation in which the pointing device subsystem selectively communicates the display control signals to a peripheral device communication channel.
In another aspect, the invention features a method of producing display control signals in accordance with which display control signals are produced in response to user input. In a handheld device mode of operation, the display control signals are selectively communicated to a display subsystem of a handheld device. In an input device mode of operation, the display control signals are selectively communicated to a peripheral device communication channel.
Other features and advantages of the invention will become apparent from the following description, including the drawings and the claims.
In the following description, like reference numbers are used to identify like elements. Furthermore, the drawings are intended to illustrate major features of exemplary embodiments in a diagrammatic manner. The drawings are not intended to depict every feature of actual embodiments nor relative dimensions of the depicted elements, and are not drawn to scale. Elements shown with dashed lines are optional elements in the illustrated embodiments incorporating such elements.
The embodiments that are described in detail below provide systems and methods of producing display control signals for a handheld device display and a remote display.
The pointing device subsystem 14 produces display control signals in response to user input (
The pointing device subsystem 14 has a handheld device mode operation in which it operates as a handheld device (e.g., a cellular telephone, a cordless telephone, a pager, a personal digital assistant (PDA), a digital audio player, a digital camera, and a digital video game console), and a pointing device mode of operation in which it operates as a pointing device for a remote system. In general, the mode of operation of the pointing device subsystem 14 may be set manually by a user (e.g., by setting a mode control switch) or it may be set automatically by the pointing device subsystem 14 (e.g., by detecting a condition or state of the handheld device 10).
In the handheld device mode of operation, the pointing device subsystem 14 selectively communicates the display control signals 20 to the display subsystem 12 (
In the input device mode of operation, the pointing device subsystem 14 selectively communicates the display control signals 24 to a peripheral device communication channel 26 (
The display control signals 24 are transmitted over the peripheral device communication channel 26 to a remote display-based system 30. In general, the remote system 30 may be any type of display-based appliance that receives user input, including a general-purpose computer system, a special-purpose computer system, and a video game system. In the embodiment shown in
The telephone subsystem 42 communicates the telephony signals 46 over a telephony communications channel 50 (
The display screen 62 and the microphone 68 are exposed through an inner face of a top part 72 of the housing 62. The keypad 66 and the speaker 70 are exposed through an inner face of a bottom part 74 of the housing 62. The top and bottom parts 72, 74 of the housing 62 are connected together by a hinged portion 76, which allows the top and bottom parts 72, 74 to pivot between an open state and a closed state. In the open state shown in
In the illustrated embodiment, the light source 90 is implemented by a light emitting diode 132 and an optical element 134 that collimates the light 136 that is produced by the light emitting diode 132 into a collimated beam 138. An optical element 137 (e.g., a mirror) deflects the collimated beam 138 through the optical port 88 that is formed in an exterior wall 142 of the bottom part 74 of the housing 62. In the illustrated example, the optical port 88 is positioned adjacent to a surface 140 (e.g., a desktop surface). A portion of the deflected beam that reflects from the surface 140 is focused by an optical element 144 onto the active areas of the optical sensor 92.
In some embodiments, the optical sensor 92 corresponds to an optical navigation sensor module that includes an imager 146 and a movement detector 148. The imager 146 may be any form of imaging device that is capable of capturing one-dimensional or two-dimensional images of the surface 140. The imager 146 includes at least one image sensor. Exemplary image sensors include one-dimensional and two-dimensional CMOS (Complimentary Metal-Oxide Semiconductor) image sensors and CCD (Charge-Coupled Device) image sensors. The imager 146 captures images at a rate (e.g., 1500 pictures or frames per second) that is fast enough so that sequential pictures of the surface 140 overlap. The imager 146 may include one or more optical elements that focus the light that reflects from the surface 140 onto the one or more image sensors.
The movement detector 148 may be part of the processing system 126 or it may be part of the optical sensor 92 as shown in
In some implementations, the movement detector 148 correlates features identified in successive images to provide information relating to the position of the surface 140 relative to the imager 146. In general, any type of correlation method may be used to track the positions of features across successive images. In some embodiments, a sum of squared differences correlation method is used to find the locations of identical features in successive images in order to determine the displacements of the features across the images. In some of these embodiments, the displacements are summed or integrated over a number of images. The resulting integration values may be scaled to compensate for any image scaling by the optics associated with the imager 146. The movement detector 148 translates the displacement information into two-dimensional relative motion vectors 150 (e.g., X and Y motion vectors) that describe the relative movement of the handheld device 60 across the surface 140. Additional details relating to the image processing and correlating methods that are performed by the movement detector 148 can be found in U.S. Pat. Nos. 5,578,813, 5,644,139, 5,703,353, 5,729,008, 5,769,384, 5,825,044, 5,900,625, 6,005,681, 6,037,643, 6,049,338, 6,249,360, 6,259,826, 6,233,368, and 6,927,758. In some embodiments, the imager 146 and the movement detector 148 may be implemented by an optical mouse navigation sensor module (e.g., the ADNS-2051 optical mouse navigation sensor available from Agilent Technologies, Inc. of Palo Alto, Calif., U.S.A.).
The processing system 126 produces the display control signals 20, 24 from the two-dimensional motion vectors 150 that are generated by the movement detector 148.
The output mode detector 128 produces an output mode control signal 152 that controls the output path over which the multiplexer 130 transmits the display control signals 20, 24 (i.e., over a path that leads to the display subsystem 12 or over a path that leads to the peripheral device communication channel 26). In some embodiments, the output mode control signal 152 has a handheld device state and a pointing device state. In response to the receipt of the output mode control signal 152 in the handheld device state, the multiplexer 130 transmits the display control signals 20 to the display subsystem 12 without transmitting the display control signals 24 to the peripheral device communication channel 26. In response to the receipt of the output mode control signal 152 in the pointing device state, the multiplexer 130 transmits the display control signals 24 to the peripheral device communication channel 26 without transmitting the display control signals 20 to the display subsystem 12.
In some implementations, the handheld device state of the output mode control signal corresponds to a first (e.g., HIGH) logic state, and the pointing device state of the output mode control signal corresponds to a second (e.g., LOW) logic state. In some embodiments, the output mode detector 128 establishes the state of the output mode control signal 152 based on a switch that is manually set by a user. In other embodiments, the output mode detector 128 establishes the state of the output mode control signal 152 based on a determination of whether the handheld device 60 is in the open state (
The directional polarity difference between the first and second input mode states accommodates the user's different expectations regarding the movement of the pointer (38, 156) in response to movement of the handheld device 170 relative to a fixed surface as compared to movement of a surface, such as a finger, relative to the handheld device 170. For example, when the user moves the handheld device 170 in an upward direction across a fixed surface, the user expects the pointer to move in an upward direction in the display screen. Similarly, when the user moves a finger in an upward direction across the optical port 88, the user expects the pointer to move in an upward direction in the display screen. However, the movement detector 148 sees the upward movement of the handheld device across the fixed surface as an upward movement, whereas the movement detector 148 sees the upward movement of the finger across the optical port 88 as downward movement. The processing system 126 corrects this discrepancy by setting the directional polarity of the display control signals based on the input mode control signal 176.
In some embodiments, the input mode detector 172 includes a photodetector, which is positioned to receive light through a second optical port 180 in the exterior wall 142 of the bottom part 74 of the housing 62. The second optical port 180 preferably is positioned at a location on the outer face of the bottom part 74 of the housing 62 that is displaced from the first optical port 88 and typically is not covered by the user's hand during normal use of the handheld device in the second input mode. In these embodiments, the input mode detector 172 sets the input mode state of the input mode control signal 176 based on the intensity of light that is detected through the second optical port 180 in relation to an empirically determined threshold. If the detected light intensity is below the threshold, the input mode detector 172 sets the input mode control signal 176 in the first input mode state (under the assumption that the detected light intensity will be low when the bottom part 74 of the handheld device is positioned adjacent to the surface 140). Conversely, if the detected light intensity is equal to or greater than the threshold, the input mode detector 172 sets the input mode control signal 176 in the second input mode state (under the assumption that the detected light intensity will be high when the bottom part 74 of the handheld device is removed from the surface 140).
The rechargeable power supply 194 may be any type of battery or other electrical power store that can be recharged by an external power source through the communications port 192. The rechargeable power supply 194 supplies power to the energizable components of the handheld device 190, including the display subsystem 12 and the pointing device subsystem 14.
The power controller 196 controls the supplying of power from the rechargeable power supply 194. In general, the power controller 196 may implement any one of a wide variety of different power management algorithms. In some embodiments, the power controller 196 is implemented by the separate or shared processing resources of the pointing device subsystem 14 and the other subsystems 16, instead of a separate component as shown in
Other embodiments are within the scope of the claims.
Number | Name | Date | Kind |
---|---|---|---|
5682529 | Hendry et al. | Oct 1997 | A |
5926166 | Khederzadeh et al. | Jul 1999 | A |
5935226 | Klein | Aug 1999 | A |
5943625 | Yeom et al. | Aug 1999 | A |
5991883 | Atkinson | Nov 1999 | A |
6038670 | Oh | Mar 2000 | A |
6057540 | Gordon et al. | May 2000 | A |
6145029 | Deschepper et al. | Nov 2000 | A |
6285354 | Revis | Sep 2001 | B1 |
6304244 | Hawkins et al. | Oct 2001 | B1 |
6438622 | Haghighi et al. | Aug 2002 | B1 |
6462763 | Mostyn | Oct 2002 | B1 |
6654826 | Cho et al. | Nov 2003 | B1 |
6665173 | Brandenberg et al. | Dec 2003 | B2 |
6864891 | Myers | Mar 2005 | B2 |
6909906 | Miyashita | Jun 2005 | B1 |
7102626 | Denny, III | Sep 2006 | B2 |
7209813 | Namaky | Apr 2007 | B2 |
7248231 | Hurley et al. | Jul 2007 | B2 |
7249257 | Brundage et al. | Jul 2007 | B2 |
7450110 | Shahoian et al. | Nov 2008 | B2 |
7508418 | Renkis | Mar 2009 | B2 |
7576730 | Anwar | Aug 2009 | B2 |
20020011990 | Anwar | Jan 2002 | A1 |
20020030668 | Hoshino et al. | Mar 2002 | A1 |
20020073247 | Baweja et al. | Jun 2002 | A1 |
20020147910 | Brundage et al. | Oct 2002 | A1 |
20030014212 | Ralston et al. | Jan 2003 | A1 |
20030034959 | Davis et al. | Feb 2003 | A1 |
20030067470 | Main et al. | Apr 2003 | A1 |
20030137469 | Hansen et al. | Jul 2003 | A1 |
20030206394 | Ossia | Nov 2003 | A1 |
20040069117 | Akins | Apr 2004 | A1 |
20040090424 | Hurley et al. | May 2004 | A1 |
20040135741 | Tomisawa et al. | Jul 2004 | A1 |
20040189609 | Estes et al. | Sep 2004 | A1 |
20040263472 | Tachikawa | Dec 2004 | A1 |
20050117130 | Bohn et al. | Jun 2005 | A1 |
20050275633 | Varanda | Dec 2005 | A1 |
20060063539 | Beyer | Mar 2006 | A1 |
20060075934 | Ram | Apr 2006 | A1 |
20060092866 | Kim et al. | May 2006 | A1 |
20060152489 | Sweetser et al. | Jul 2006 | A1 |
20060244733 | Geaghan | Nov 2006 | A1 |
20060250377 | Zadesky et al. | Nov 2006 | A1 |
20070026371 | Wood | Feb 2007 | A1 |
20070135057 | Nallapureddy et al. | Jun 2007 | A1 |
20070146347 | Rosenberg | Jun 2007 | A1 |
20070162858 | Hurley et al. | Jul 2007 | A1 |
20080291855 | Bata et al. | Nov 2008 | A1 |
20090266892 | Windauer et al. | Oct 2009 | A1 |
Number | Date | Country |
---|---|---|
WO 9737293 | Oct 1997 | WO |
Number | Date | Country | |
---|---|---|---|
20070120824 A1 | May 2007 | US |