The field of the present invention is user interfaces for electronic devices and, more particularly, to touch screen user interfaces.
Touch screens provide user interfaces through which a user enters input to a computing device by touching a screen at a selected location, with a stylus or with his finger.
Conventional touch screens are limited as to the types of user inputs that they can recognize. For example, conventional touch screens are unable to distinguish between a soft tap and a hard press. In some prior art embodiments users initially select an item on the screen, and then subsequently activate the selected item. However, because prior art touch screens do not distinguish between degrees of pressure, the user is required to remove his finger or stylus from the screen and activate his selection with a second tap. It would be advantageous to produce touch screens that distinguish between varying amounts of pressure applied to the screen so that a user can select an item by touching its location on the screen, and then activate the item by applying additional pressure to the touch location without having to first remove his finger or stylus from the screen.
In conventional touch screens the keys are often small relative to the touch area, especially in handheld devices. The keys are also often situated close together. This can make it difficult to determine which key is being pressed by the user. It would be advantageous to clearly indicate to the user which key the user has selected, and furthermore, to allow the user to accept or reject the selected key without first removing his finger or stylus from the screen in order to perform a second tap.
Current user interfaces are basic and often require navigation through a series of menus in order to perform a desired operation. The present invention presents a user interface that is versatile in providing the user with many options, while requiring only few selections to activate a desired function. To further enhance user experience, certain functions are performed automatically without requiring the user to enter a selection.
Aspects of the present invention relate to user interfaces designed for use with a touch screen. The present invention relates to computer readable media storing computer programs with computer program code, which, when read by a computer unit, allows the computer to present a user interface for the computer unit.
In accordance with embodiments of the present invention, the computer unit features a touch sensitive display area. According to preferred embodiments an initial display configuration presents a plurality of gadgets on the display. These gadgets are small areas of the screen that indicate which functions each gadget will perform when activated by the user. When a gadget is activated, typically by the user touching the area of the display on which the gadget is displayed, the gadget increases in size and provides the user with icons and information about the gadget's functions. Significantly, the gadget does not cover the entire display area. Thus, when a gadget, or even a plurality of gadgets, is active, the primary display of all available gadgets is still accessible. This primary display can be compared to a desktop in computer operating system user interfaces. However, this primary display in the user interface of the present invention is not the same as a desktop where active windows can cover icons on the desktop. In the present invention, gadgets are arranged in a manner that open gadgets do not cover other gadgets. Rather, when an open gadget expands in size, other gadgets are shifted to make room for the expanded gadget. This allows the user to scroll the primary display or desktop to view any gadget. In the context of the present invention, this primary display area that includes both open and closed gadgets is called the home window. The user scrolls the home window to view gadgets that are shifted beyond the viewable area of the display. An expanded, or activated, gadget has an expanded window, but often it is not large enough to display everything contained in that gadget window. To view contents of the gadget not displayed in the visible portion of the gadget window, the user scrolls the gadget window. Thus, two different scroll operations are provided: scrolling the home window and scrolling a gadget window. According to one embodiment, scrolling is executed by gliding a finger or stylus along the touch screen to shift the active display area of the home window or of the gadget. The scrolling affects the home window if the finger glide began in an area of the screen that does not belong to an active gadget; the scroll affects an active gadget window if the finger glide began inside that active gadget window.
Various embodiments of the invention support several methods of scrolling a window. According to one embodiment, scrolling is done when the user touches the display inside the window area, for example at an edge of the window, or on an icon, such as an arrow or scrollbar, indicating a scroll operation. According to another embodiment, scrolling is done by the user touching the window with a finger or stylus and then gliding the finger or stylus along the touch sensitive screen in a direction indicating the desired direction of the scroll. When the content of the home display is larger in two dimensions than the actual display screen, this operation is like panning an image or map. When it is larger along only one axis (e.g., only vertically), the scrolling only scrolls in the one axis even when the glide is not orthogonal along the one axis.
Another aspect of the present invention relates to computer readable media storing a computer program with computer program code, which, when read by a mobile handheld computer unit, allows the computer to present a user interface for the mobile handheld computer unit. The user interface features a touch sensitive area in which representations of a plurality of keys are displayed, and each key is mapped to a corresponding location in the touch sensitive area at which the representation of the key is displayed. A key in this context includes, inter alia, alphabetic keys such as in a QWERTY keypad, numeric keys and also icons representing programs or functions. A key is selected, but not activated, when an object touches the corresponding location. This intermediate status of being selected but not activated facilitates the user to subsequently activate a desired key and avoid activating a neighboring key that the user selected but does not wish to activate. A selected key is activated when the object touching it applies additional pressure to the key location.
According to preferred embodiments of the invention, when a key is selected, the user interface generates a secondary representation of the key, such as a callout balloon containing the key representation. The callout balloon is placed away from the key location (being touched) so that the user can easily view which key is selected without lifting his finger. According to another embodiment, an audio representation of the selected key is generated so the user hears which key was selected.
According to still further features in preferred embodiments of the invention, the user touches the screen (with a finger or stylus) at a first location, for example selecting a first key. The user then glides his finger or stylus over the screen to additional locations. At each additional location a new key is selected and the previously selected key is deselected. The user can activate any selected key by applying additional pressure to the screen. The user does not have to remove the object from the screen to glide and select additional keys even after activating a first key.
Additional touch pressure is detected in various ways according to several embodiments. According to one embodiment, the touch sensitive area is a light-based touch screen operable to detect different levels of touch pressure. For example, light-based touch screens typically include a calculating unit operable to identify the size, shape and contours of an area being touched based on a pattern of obstructed light. See applicant's co-pending U.S. patent application Ser. No. 10/494,055, titled ON A SUBSTRATE FORMED OR RESTING DISPLAY ARRANGEMENT, the contents of which are incorporated herein by reference. When a finger or flexible object is used as a touch object, as additional pressure is applied to the touch surface, the contact area of the finger or object touching the screen increases. Thus, additional pressure is detected when an increase in the contours of the covered touch area is detected.
Alternatively, or in combination with the above, the touch sensitive area features both a touch screen operable to identify a touch location on the screen and a pressure sensor operable to detect pressure applied to the screen but not sensitive to the location of the object applying the pressure.
Other aspects of the present invention relate to convenient arrangement and function of icons to perform popular functions within a user interface. Thus, a camera gadget features a multimedia messaging service (MMS) button facilitating sending an active photo in an MMS message; a keylock gadget locks the computer and displays an instrumental keypad for entering a musical code to unlock the computer; a reporting gadget displays information for a first period of time and is then automatically deactivated. Several reporting gadgets are provided, including a gadget that displays the time of day; a gadget displays a weather forecast; a gadget that displays stock market information.
According to still further features in preferred embodiments the reporting gadget continues to display its information for a second period of time if the gadget is touched during the first period of time. I.e., automatic deactivation after the first period of time is canceled.
The present invention will be more fully understood and appreciated from the following detailed description, taken in conjunction with the drawings in which:
Touch screen keypads offer great flexibility in keypad interface design—flexibility that cannot be achieved with electromechanical keypads. Custom design keypads can be generated on a touch screen, where the markings on each key and the functions that each key provides are optimized for a designated application. Moreover, touch screen keypads can change modes, from one pad of keys and associated functions to a different pad of keys and associated functions. Custom keypads are of particular advantage for multi-lingual applications.
A general description of touch screen keypad interfaces, in accordance with embodiments of the present invention, and several examples thereof, are described in detail hereinbelow.
Embodiments of the present invention relate to improved keypads for inputting Chinese characters using XT9 stroke input, and using Chinese Pinyin. XT9 stroke input builds Chinese characters using six basic strokes, and offers selection of possible characters and phrases based on a set of strokes that have been input. Chinese Pinyin uses Latin characters that transliterate a sound or a syllable, in combination with a digit that represents an intonation or inflection. E.g., Ma in a rising tone is m-a-1, and Ma in a descending tone is m-a-2.
Reference is now made to
In accordance with an embodiment of the present invention, keypad 100 is generated and displayed on a touch screen. Keypad 100 has fewer than the standard 12 keys in a touch pad, allowing more room on screen for displaying characters.
Further in accordance with an embodiment of the present invention, the keys of keypad 100 are customized so that they contain only relevant information. For example, a prior art keypad displays a digit, 3-4 characters, and a basic Chinese stroke, all inside one key, even though in XT9 stroke mode the basic Chinese stroke is the only useful one. The custom keys of the present invention display only the basic Chinese strokes, or the strokes and numbers, but no characters.
There are two types of key presses supported by keypad 100—regular and long. A regular key press adds the stroke shown on the key to the series of strokes 101-106 already pressed. As strokes are successively entered, a numbered array 121 of Chinese characters or phrases is dynamically displayed along the top of the keypad. These characters or phrases are predicted based on the key presses already entered. In order to select one of the numbered elements of array 121, the user performs a long key press on that number. Alternatively, the user may keep entering strokes until only one option remains.
Often, however, more predicted characters or phrases exist than can be displayed along the top of the keypad. The N2 phone, manufactured by Neonode of Stockholm, Sweden, has a joystick button at the bottom of the phone. Twitching the joystick up/down displays different sets of predicted characters or phrases. When the intended character or phrase is displayed and associated with a given digit, a long press on that digit serves to select the intended character or phrase.
Reference is now made to
In accordance with an embodiment of the present invention, keypad 200 uses accent characters, rather than digits, to convey an intended intonation or inflection. Further in accordance with an embodiment of the present invention, keypad 200 displays only information relevant for Pinyin input on each key; no Chinese basic strokes are shown.
There are two types of Pinyin input. A user enters a Latin transliteration of an intended word using the Latin keypad input (12 keys). For each key, several letters are possible. The list of predicted Latin syllables based on the current sequence of keypad presses is displayed. Twitching the joystick right or left selects the desired combination. Also, a series of predicted Chinese characters or phrases is displayed and selected by a long press on a respective digit. Twitching the joystick up/down displays other predicted Chinese characters or phrases. Entering a space after a series of letters indicates the end of a previous character or phrase.
In accordance with an embodiment of the present invention, the user is able to combine stroke and Pinyin input, and compose a sequence of at least two Chinese characters using XT9 stroke input for at least one character and Pinyin input for at least one other character. The user switches between XT9 stroke input mode and Pinyin input mode by performing a sweeping motion in relation to the touch screen, such as, inter alia, sweeping a finger across the top of the touch screen. The series of at least two Chinese characters may be a text message, a name, a data entry, or any other such input.
Further in accordance with an embodiment of the present invention, the user is able to compose a series of at least one Chinese character and at least one non-Chinese term, wherein the non-Chinese term includes at least one Latin character, digit, emoticon, punctuation mark, another non-Chinese symbol, or any combination thereof. The series is composed by switching input modes for each alphabet or Chinese input or digit input by sweeping across the top of the touch screen. For example, the user may input at least one Chinese character using either Pinyin or stroke input, or a combination thereof. The user may then perform a sweeping motion in relation to the touch screen to change the input mode to English. For example, the user may sweep a finger across the top of the touch screen to change the input mode to English. In this mode, the keypad presents Latin characters. The user then proceeds to input Latin characters using the Latin keypad displayed on the touch screen. Alternatively, the user may repeat a series of sweeping motions; e.g., the user sweeps a finger across the top of the touch screen, repeatedly, changing the input mode with each sweeping motion, until a digit keypad is displayed and digit input mode is active. The user may then proceed to enter at least one digit, adding the at least one digit to the series of Chinese characters already contained in the message. It will thus be appreciated that the user may switch between different input modes while composing a single message, a command, a name, a data entry or another such input, including at least two different types of characters, in an easy and simple and convenient manner.
Further in accordance with an embodiment of the present invention, a keypad displaying emoticons is displayed. In this mode, the user may select an emoticon to be entered into the text of a message, or such other input.
Yet further in accordance with an embodiment of the present invention, drawings, including inter alia, emoticons, are constructed in a similar manner to XT9 stroke input. In this mode, the user interface displays the basic building blocks for the drawing, such as a curve, a semicolon, a circle, and other symbols. As the user taps multiple symbols, possible drawings or emoticons that can be formed using the selected elements are displayed, and the user may either select the desired complete drawing or emoticon from the displayed list, or may continue entering additional building blocks until only one option remains. This mode of input is convenient as the number and size of the keys presented is optimized for the number of available building blocks, and each key only displays information relevant for the active input mode.
Embodiments of the present invention provide methods and systems for enabling multiple input modes, whereby the screen display in each input mode is optimized for that mode. Optimizations include (i) configuring the number of keys displayed, (ii) configuring the size, position and shape of the keys in relation to the screen, (iii) configuring the size, position and shape of the area of the display showing text already entered, (iv) configuring the size, position and shape of the area of the display showing possible completions for the current character, phrase or symbol, and (v) displaying only at least one character, symbol, digit or other figure that is relevant to the active input mode on each key.
Embodiments of the present invention also provide methods and systems for enabling multiple input modes and switching between the input modes by performing a sweeping motion in relation to the screen. These methods and systems are easier and more convenient than using a menu interface to switch input modes. Additionally, these methods do not use up screen space to provide a switching key, to switch between input modes, and, as such, screen space may be used for information related to the current input mode and the current text. These methods and systems also enable switching between many input modes, by repeating the sweeping motions to advance through the input modes. In the N2 phone manufactured by Neonode AB, this mode of switching active keypads is used with standard 12-key keypads, for example, inter alia, switching between numeric, Swedish and English keypads. However, switching the layout of the keys is not used in the Neonode N2 phone. Moreover, other, prior art methods of switching between input modes require more than one action to advance through a list of input modes.
Reference is now made to FIGS. 3 and 9-45, which include several exemplary touch screen interfaces. Shown in
According to embodiments of the present invention, selection and subsequent activation of an on-screen element, such as, inter alia, a gadget, a letter, a key, a button or an icon, is implemented by two primary features. The first feature, referred to as pressure sensing, provides the touch screen with one or more pressure sensors operative to determine contact with the screen by an external object such as a finger. In one embodiment, the user places a finger or a stylus over the desired on-screen element to select it, as illustrated in
The second feature, referred to as callout balloons, indicates a selected on-screen element to the user by displaying a graphic, such as by (i) enlarging the on-screen element, (ii) displaying a callout balloon with a copy of the on-screen element inside, or (iii) a combination thereof. The callout balloon is illustrated in
In accordance with an embodiment of the present invention, touch screens for phones captioned “Power ON/Power OFF” are shown in
Additionally in accordance with an embodiment of the present invention, touch screens for phones captioned “Key lock” are shown in
Further in accordance with an embodiment of the present invention, a touch screen for a phone captioned “Key lock high security” is shown in FIG. 15. When the home key is activated, such as by touching the home key, a keypad is displayed and a user is prompted to enter a security code. Subsequently, the phone cannot be unlocked unless the security code is entered. Optionally, additional security is implemented by use of fingerprint identification, wherein the phone cannot be unlocked unless a fingerprint is authenticated. Optionally, additional security is implemented by requiring the user to glide a finger over the screen in a designated gesture, in order to unlock the phone. The phone cannot be unlocked unless the touch screen detects the designated gesture. Examples of designated gestures include inter alia, an “S” shaped gesture, or a gesture wherein multiple corners of the screen are touched concurrently or in sequence. Optionally, additional security is implemented by displaying a musical keypad or keyboard and the user must play a particular tune of at least one note on the keypad or keyboard in order to unlock the phone.
Yet further in accordance with an embodiment of the present invention, a touch screen for a phone captioned “Home/return from application” is shown in
Moreover in accordance with an embodiment of the present invention, a touch screen for a phone captioned “Scroll alternatives” is shown in
In
In
In
Additionally in accordance with an embodiment of the present invention, a touch screen for a phone captioned “Home screen” is shown in
After the user scrolls to reveal additional gadgets and a specific time limit has elapsed, e.g., 3 seconds, the screen automatically returns to a default display, hiding recently exposed gadgets and revealing previously hidden gadgets. Such return to default display is of advantage for revealed gadgets that do not require extended viewing, such as, inter alia, a clock or a weather report. The return to default display obviates the need for the user to perform an action in order to return the screen to its initial display. Activating the home key after scrolling also returns the screen to its initial display, showing the originally displayed gadgets and hiding the recently revealed gadgets. Such return to initial display is graphically presented (i) by returning the original display completely, (ii) by fading in the original display over the current scrolled display, (iii) by gradually displacing the scrolled display with the original display, (iv) by graphically scrolling the display in reverse to the original display, or (v) by a combination of such presentations.
Further in accordance with an embodiment of the present invention, a touch screen for a phone captioned “Fav 5” is shown in
Further in accordance with an embodiment of the present invention, a touch screen for a phone captioned “History” is shown in
Yet further in accordance with an embodiment of the present invention, a touch screen for a phone captioned “Telephone” is shown in
Moreover in accordance with an embodiment of the present invention, a touch screen for a phone captioned “Active call” is shown in
Additionally in accordance with an embodiment of the present invention, a touch screen for a phone captioned “Volume” is shown in
Additionally in accordance with an embodiment of the present invention, a touch screen for a phone captioned “Mp3 player” is shown in FIGS. 33-34. Two phones are displayed under the “Mp3 player” caption.
Further in accordance with an embodiment of the present invention, a touch screen for a phone captioned “Operator ad” is shown in
Yet further in accordance with an embodiment of the present invention, a touch screen for a phone captioned “Video” is shown in
Yet further in accordance with an embodiment of the present invention, a touch screen for a phone captioned “Camera” is shown in
Yet further in accordance with an embodiment of the present invention, a touch screen for a phone captioned “Messaging SMS” is shown in
The effect of the scrolling activity is illustrated in
When a gadget is activated, for example by tapping on the screen position associated with the gadget, the display area of the selected gadget extends in one direction, for example down-screen to the bottom of the screen or beyond the bottom of the screen. This extension is shown in
In the foregoing specification, the invention has been described with reference to specific exemplary embodiments thereof. It will, however, be evident that various modifications and changes may be made to the specific exemplary embodiments without departing from the broader spirit and scope of the invention as set forth in the appended claims. Accordingly, the specification and drawings are to be regarded in an illustrative rather than a restrictive sense.
This application is a continuation-in-part of pending application U.S. Ser. No. 10/315,250, filed on Dec. 10, 2002, entitled USER INTERFACE FOR MOBILE HANDHELD COMPUTER UNIT. This application claims priority from provisional application No. 61/132,469, filed on Jun. 19, 2008, entitled IMPROVED KEYPAD FOR CHINESE CHARACTERS.
Number | Name | Date | Kind |
---|---|---|---|
4243879 | Carroll et al. | Jan 1981 | A |
4301447 | Funk et al. | Nov 1981 | A |
4518249 | Maruta et al. | May 1985 | A |
4550250 | Mueller et al. | Oct 1985 | A |
4703316 | Sherbeck | Oct 1987 | A |
4710760 | Kasday | Dec 1987 | A |
4782328 | Denlinger | Nov 1988 | A |
4790028 | Ramage | Dec 1988 | A |
4847606 | Beiswenger | Jul 1989 | A |
4880969 | Lawrie | Nov 1989 | A |
5036187 | Yoshida et al. | Jul 1991 | A |
5053758 | Cornett et al. | Oct 1991 | A |
5179369 | Person et al. | Jan 1993 | A |
5194863 | Barker et al. | Mar 1993 | A |
5283558 | Chan | Feb 1994 | A |
5414413 | Tamaru et al. | May 1995 | A |
5463725 | Henckel et al. | Oct 1995 | A |
5559727 | Deley et al. | Sep 1996 | A |
5577733 | Downing | Nov 1996 | A |
5579035 | Beiswenger | Nov 1996 | A |
5581681 | Tchao et al. | Dec 1996 | A |
5603053 | Gough et al. | Feb 1997 | A |
5612719 | Beernink et al. | Mar 1997 | A |
5618232 | Martin | Apr 1997 | A |
5748185 | Stephan et al. | May 1998 | A |
5785439 | Bowen | Jul 1998 | A |
5801703 | Bowden et al. | Sep 1998 | A |
5825352 | Bisset et al. | Oct 1998 | A |
5880743 | Moran et al. | Mar 1999 | A |
5886697 | Naughton et al. | Mar 1999 | A |
5889236 | Gillespie et al. | Mar 1999 | A |
5914709 | Graham et al. | Jun 1999 | A |
5936615 | Waters | Aug 1999 | A |
5943044 | Martinelli et al. | Aug 1999 | A |
5988645 | Downing | Nov 1999 | A |
6010061 | Howell | Jan 2000 | A |
6031989 | Cordell | Feb 2000 | A |
6088032 | Mackinlay | Jul 2000 | A |
6292179 | Lee | Sep 2001 | B1 |
6323846 | Westerman et al. | Nov 2001 | B1 |
6340979 | Beaton et al. | Jan 2002 | B1 |
6359632 | Eastty et al. | Mar 2002 | B1 |
6421042 | Omura et al. | Jul 2002 | B1 |
6429857 | Masters et al. | Aug 2002 | B1 |
6448985 | McNally | Sep 2002 | B1 |
6549217 | De Greef et al. | Apr 2003 | B1 |
6570557 | Westerman et al. | May 2003 | B1 |
6664983 | Ludoplh | Dec 2003 | B2 |
6677932 | Westerman | Jan 2004 | B1 |
6690365 | Hinckley et al. | Feb 2004 | B2 |
6690387 | Zimmerman et al. | Feb 2004 | B2 |
6707449 | Hinckley et al. | Mar 2004 | B2 |
6757002 | Oross et al. | Jun 2004 | B1 |
6765559 | Hayakawa | Jul 2004 | B2 |
6788292 | Nako et al. | Sep 2004 | B1 |
6803906 | Morrison et al. | Oct 2004 | B1 |
6857746 | Dyner | Feb 2005 | B2 |
6864882 | Newton | Mar 2005 | B2 |
6888536 | Westerman et al. | May 2005 | B2 |
6947032 | Morrison et al. | Sep 2005 | B2 |
6954197 | Morrison et al. | Oct 2005 | B2 |
6972401 | Akitt et al. | Dec 2005 | B2 |
7007239 | Hawkins et al. | Feb 2006 | B1 |
7030861 | Westerman et al. | Apr 2006 | B1 |
7155683 | Williams | Dec 2006 | B1 |
7176905 | Baharav et al. | Feb 2007 | B2 |
7184030 | McCharles et al. | Feb 2007 | B2 |
7232986 | Worthington et al. | Jun 2007 | B2 |
7283845 | De Bast | Oct 2007 | B2 |
7304638 | Murphy | Dec 2007 | B2 |
RE40153 | Westerman et al. | Mar 2008 | E |
7339580 | Westerman et al. | Mar 2008 | B2 |
7355594 | Barkan | Apr 2008 | B2 |
7372456 | McLintock | May 2008 | B2 |
7441196 | Gottfurcht et al. | Oct 2008 | B2 |
7464110 | Pyhälammi et al. | Dec 2008 | B2 |
7479949 | Jobs et al. | Jan 2009 | B2 |
8022941 | Smoot | Sep 2011 | B2 |
8120625 | Hinckley | Feb 2012 | B2 |
8127141 | HYpponen | Feb 2012 | B2 |
20010002694 | Nakazawa et al. | Jun 2001 | A1 |
20010022579 | Hirabayashi | Sep 2001 | A1 |
20010026268 | Ito | Oct 2001 | A1 |
20010028344 | Iwamoto et al. | Oct 2001 | A1 |
20010055006 | Sano et al. | Dec 2001 | A1 |
20020015064 | Robotham et al. | Feb 2002 | A1 |
20020109843 | Ehsani et al. | Aug 2002 | A1 |
20020173300 | Shtivelman et al. | Nov 2002 | A1 |
20020175900 | Armstrong | Nov 2002 | A1 |
20030002809 | Jian | Jan 2003 | A1 |
20030010043 | Ferragut, II | Jan 2003 | A1 |
20030030656 | Ang et al. | Feb 2003 | A1 |
20030095102 | Kraft et al. | May 2003 | A1 |
20030098803 | Gourgey et al. | May 2003 | A1 |
20030234346 | Kao | Dec 2003 | A1 |
20040046960 | Wagner et al. | Mar 2004 | A1 |
20040125143 | Deaton et al. | Jul 2004 | A1 |
20050091612 | Stabb et al. | Apr 2005 | A1 |
20050104860 | McCreary et al. | May 2005 | A1 |
20050122308 | Bell et al. | Jun 2005 | A1 |
20050165615 | Minar | Jul 2005 | A1 |
20050174473 | Morgan et al. | Aug 2005 | A1 |
20050198582 | Hennum et al. | Sep 2005 | A1 |
20050253818 | Nettamo | Nov 2005 | A1 |
20050271319 | Graham | Dec 2005 | A1 |
20060001654 | Smits | Jan 2006 | A1 |
20060048073 | Jarrett et al. | Mar 2006 | A1 |
20060161870 | Hotelling et al. | Jul 2006 | A1 |
20060161871 | Hotelling et al. | Jul 2006 | A1 |
20060229509 | Al-Ali et al. | Oct 2006 | A1 |
20070024598 | Miller et al. | Feb 2007 | A1 |
20070082707 | Flynt et al. | Apr 2007 | A1 |
20070084989 | Lange et al. | Apr 2007 | A1 |
20070146318 | Juh et al. | Jun 2007 | A1 |
20070150842 | Chaudhri et al. | Jun 2007 | A1 |
20070152984 | Ording et al. | Jul 2007 | A1 |
20080008472 | Dress et al. | Jan 2008 | A1 |
20080012850 | Keating, III | Jan 2008 | A1 |
20080013913 | Lieberman et al. | Jan 2008 | A1 |
20080055273 | Forstall | Mar 2008 | A1 |
20080068353 | Lieberman et al. | Mar 2008 | A1 |
20080086703 | Flynt et al. | Apr 2008 | A1 |
20080093542 | Lieberman et al. | Apr 2008 | A1 |
20080100593 | Skillman et al. | May 2008 | A1 |
20080111797 | Lee | May 2008 | A1 |
20080117176 | Ko et al. | May 2008 | A1 |
20080117183 | Yu et al. | May 2008 | A1 |
20080121442 | Boer et al. | May 2008 | A1 |
20080122792 | Izadi et al. | May 2008 | A1 |
20080122796 | Jobs et al. | May 2008 | A1 |
20080122803 | Izadi et al. | May 2008 | A1 |
20080136790 | Hio | Jun 2008 | A1 |
20080158174 | Land et al. | Jul 2008 | A1 |
20080168404 | Ording | Jul 2008 | A1 |
20080192025 | Jaeger et al. | Aug 2008 | A1 |
20080211779 | Pryor | Sep 2008 | A1 |
20080259053 | Newton | Oct 2008 | A1 |
20080297409 | Klassen et al. | Dec 2008 | A1 |
20080297487 | Hotelling et al. | Dec 2008 | A1 |
20090006418 | O'Malley | Jan 2009 | A1 |
20090027357 | Morrison | Jan 2009 | A1 |
20090031208 | Robinson | Jan 2009 | A1 |
20090058833 | Newton | Mar 2009 | A1 |
20090064055 | Chaudhri et al. | Mar 2009 | A1 |
20090066673 | Molne et al. | Mar 2009 | A1 |
20090096994 | Smits | Apr 2009 | A1 |
20090187840 | Moosavi | Jul 2009 | A1 |
20090192849 | Hughes et al. | Jul 2009 | A1 |
20090285383 | Tsuei | Nov 2009 | A1 |
20100031203 | Morris et al. | Feb 2010 | A1 |
20110239155 | Christie | Sep 2011 | A1 |
Number | Date | Country |
---|---|---|
0513694 | Nov 1992 | EP |
0330767 | Oct 1993 | EP |
0618528 | Oct 1994 | EP |
0703525 | Mar 1996 | EP |
8600446 | Jan 1986 | WO |
8600447 | Jan 1986 | WO |
0102949 | Nov 2001 | WO |
0295668 | Nov 2002 | WO |
03038592 | May 2003 | WO |
2006020304 | Feb 2006 | WO |
2009008786 | Jan 2009 | WO |
Entry |
---|
Southern Belle, Nancy Drew—Message in a Haunted Mansion—Walkthrough, Adventure Lantern, available at http://www.adventurelantern.com/walkthroughs/nancydrew3/nancydrew3.htm (posted Oct. 2006). |
Wagner, Richard, Professional iPhone and iPod touch Programming: Building Applications for Mobile Safari, Wiley Publishing, pp. ii and 113 (Jan. 2008). |
Adamb et al., Any advice on using DD to drag-pan an overflow:hidden div (a la Google maps) [Online forum thread], Sencha Inc., available at http://www.sencha.com/forum/showthread.php?35505 (first post Mar. 20, 2008). |
Keckeroo et al., Large image—small window [Online forum thread], Sencha Inc., available at http://www.sencha.com/forum/showthread.php?35505 (first post May 14, 2008). |
Hodges, S., Izadi, S., Butler, A., Rrustemi, A., and Buxton, B., ThinSight: Versatile Multi-Touch Sensing for Thin Form-Factor Displays, UIST'07, Oct. 7-10, 2007. |
Number | Date | Country | |
---|---|---|---|
20100017872 A1 | Jan 2010 | US |
Number | Date | Country | |
---|---|---|---|
61132469 | Jun 2008 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 10315250 | Dec 2002 | US |
Child | 12486033 | US |