The following relates generally to displaying a character using a gesture.
Many electronic devices, including mobile devices, include one or more touch-sensitive input mechanisms such as a touch-sensitive display or a touch pad to provide inputs to the electronic device. The user can provide an input to the touch-sensitive component using an object (e.g. a finger of a user or a stylus) to perform a gesture near or directly on the surface of the touch-sensitive component. For example, the gesture can include swiping the object upwards or downwards across a portion of the touch-sensitive display to scroll content displayed on the touch-sensitive display.
Embodiments will now be described by way of example only with reference to the appended drawings wherein:
a), 1(b), 1(c) and 1(d) show different stages of a graphical user interface (GUI) for using touch gestures to display a time.
a) and 8(b) is an illustration of different example seven-segment displays.
a), 16(b), 16(c), 16(d), 16(e), 16(f), 16(g) and 16(h) show different stages of a GUI for displaying the number ‘2’ using a multi-segment display.
a), 17(b), 17(c), 17(d), 17(e), 17(f), and 17(g) show different stages of a GUI for displaying the number ‘3’ using a multi-segment display.
a) is an illustration of a GUI showing multi-segment displays for the month, day and year as well as inputted touch gestures.
b) is an illustration of the GUI in
a) and 21(b) illustrate different touch gestures over a multi-segment display that are used to show the numeral ‘2’.
a), 23(b) and 23(c) show different stages of a GUI for displaying the time ‘1:30’.
a), 24(b) and 24(c) show different stages of a GUI for displaying the time ‘1:35’.
a), 25(b) and 25(c) show different stages of a GUI for displaying the time ‘1:15’.
a), 26(b) and 26(c) show different stages of a GUI for displaying the time ‘11:00’.
a), 27(b) and 27(c) show different stages of another GUI for displaying the time ‘1:15’.
a), 28(b) and 28(c) show different stages of another GUI for displaying the time ‘1:35’.
a), 29(b), 29(c) and 29(d) show different stages of yet another GUI for displaying the time ‘1:15’.
a) illustrates a GUI of two multi-segment displays that is receiving a touch gesture for the number ‘8’.
b) illustrates the GUI of
c) illustrates the GUI of
It will be appreciated that for simplicity and clarity of illustration, where considered appropriate, reference numerals may be repeated among the figures to indicate corresponding or analogous elements. In addition, numerous specific details are set forth in order to provide a thorough understanding of the example embodiments described herein. However, it will be understood by those of ordinary skill in the art that the example embodiments described herein may be practised without these specific details. In other instances, well-known methods, procedures and components have not been described in detail so as not to obscure the example embodiments described herein. Also, the description is not to be considered as limiting the scope of the example embodiments described herein.
Various software applications display numbers and letters on electronic devices. Numbers and alphabets are herein generally referred to as “characters”. Examples of such software applications include an alarm clock, a scheduling application for a calendar, an email application, a calculator, an address book, and a phone book. For electronic devices that have touch screens, a user may use a virtual keyboard that is displayed on the touch screen to enter in numbers and letters. In some situations it is not preferable to use a virtual keyboard because the virtual keyboard uses a significant amount of space on the touch screen.
To select a time or date, some applications use a rolling picker interface that allows a user to scroll between a series of different numbers. Such interfaces are used on electronic devices with touch screens. Rolling picker interfaces can sometimes be frustrating to use.
For example, a user may not necessarily think of a particular time in relation to the current time that is displayed on the rolling time picker. In other words, it is cognitively demanding to try to get to 11:45 PM when the rolling time picker is currently displaying 8:30 PM. More often than otherwise, a user may think of a meeting time in absolute value rather than in relative terms. A rolling time picker is better suited for selecting a relative time, and typically does not as easily facilitate selecting an absolute time value.
Rolling pickers for touch screens may also suffer from inaccuracy. A user may inadvertently allow the rolling interface to roll pass the desired number, and consequently, the user will need to provide another input to the rolling picker interface to arrive at the desired number. This can be time-consuming and inconvenient to the user.
It is also recognized that rolling picker interfaces display the preceding and proceeding values on a virtual wheel. As a result, the rolling picker interface typically consumes a significant amount of real estate on the display of the touch screen.
The proposed systems and methods described herein allows a user to trace their finger on top of a multi-segment display. Segments of the multi-segment display are activated or highlighted when the finger touches them. This allows loose tracing gestures to be used. When a combination of segments are touched, a character is displayed. In example embodiments, different combinations of segments can be touched to invoke the display of the same character.
The proposed systems and methods for entering in characters addresses the issues associated with rolling picker interfaces. In an example embodiment, a guide in the form of a multi-segment display is combined with handwritten character recognition algorithms. In another example embodiment, segments of a multi-segment display are activated when touched by the user to provide immediate feedback. In another example embodiment, character recognition can also make use of heuristics to decipher or predict, or both, what is being entered through a touch gesture.
a) to 1(d) show various example states of a time display which is sequentially modified using touch gestures.
Turning to
In the example shown in
Turning to
Turning to
Turning to
Examples of applicable mobile electronic devices may include, without limitation, cellular phones, smart-phones, tablet computers, wireless organizers, personal digital assistants, computers, laptops, handheld wireless communication devices, wirelessly enabled notebook computers, portable gaming devices, and the like. Such devices will hereinafter be commonly referred to as “mobile devices” 100 for the sake of clarity. It will however be appreciated that the principles described herein are also suitable to other electronic devices, e.g. “non-mobile” devices. For example, the principles herein are equally applicable to personal computers (PCs), tabletop computing devices, wall-mounted screens such as kiosks, or any other computing device.
The mobile device 100 may be a two-way communication device with advanced data communication capabilities including the capability to communicate with other mobile devices or computer systems through a network of transceiver stations. The mobile device may also have the capability to allow voice communication. Depending on the functionality provided by the mobile device, it may be referred to as a data messaging device, a two-way pager, a cellular telephone with data messaging capabilities, a wireless Internet appliance, or a data communication device (with or without telephony capabilities).
Referring to
To aid the reader in understanding an example configuration of a mobile device 100, reference will be made to
The main processor 302 also interacts with additional subsystems such as a Random Access Memory (RAM) 306, a flash memory 308, a touch-sensitive display 102, an auxiliary input/output (I/O) subsystem 312, a data port 314, a keyboard 316, a speaker 318, a microphone 320, a GPS receiver 321, short-range communications 322, a camera 323, a accelerometer 325 and other device subsystems 324. Some of the subsystems of the mobile device 100 perform communication-related functions, whereas other subsystems may provide “resident” or on-device functions. By way of example, the display 102 and the keyboard 316 may be used for both communication-related functions, such as entering a text message for transmission over the network 202, and device-resident functions such as a calculator or task list. In one example, the mobile device 100 can include a non touch-sensitive display in place of, or in addition to the touch-sensitive display 102.
The mobile device 100 can send and receive communication signals over the wireless network 202 after required network registration or activation procedures have been completed. Network access is associated with a subscriber or user of the mobile device 100. To identify a subscriber, the mobile device 100 may use a subscriber module component or “smart card” 326, such as a Subscriber Identity Module (SIM), a Removable User Identity Module (RUIM) and a Universal Subscriber Identity Module (USIM). In the example shown, a SIM/RUIM/USIM 326 is to be inserted into a SIM/RUIM/USIM interface 328 in order to communicate with a network. Once the SIM/RUIM/USIM 326 is inserted into the SIM/RUIM/USIM interface 328, it is coupled to the main processor 302.
The mobile device 100 is typically a battery-powered device and includes a battery interface 332 for receiving one or more rechargeable batteries 330. In at least some examples, the battery 330 can be a smart battery with an embedded microprocessor. The battery interface 332 is coupled to a regulator (not shown), which assists the battery 330 in providing power to the mobile device 100. Although current technology makes use of a battery, future technologies such as micro fuel cells may provide the power to the mobile device 100.
The mobile device 100 also includes an operating system 334 and software components 336 to 346 which are described in more detail below. The operating system 334 and the software components 336 to 346 that are executed by the main processor 302 are typically stored in a persistent store such as the flash memory 308, which may alternatively be a read-only memory (ROM) or similar storage element (not shown). Those skilled in the art will appreciate that portions of the operating system 334 and the software components 336 to 346, such as specific device applications, or parts thereof, may be temporarily loaded into a volatile store such as the RAM 306. Other software components can also be included, as is well known to those skilled in the art.
The subset of software applications 336 that control basic device operations, including data and voice communication applications, may be installed on the mobile device 100 during its manufacture. Software applications may include a message application 338, a device state module 340, a Personal Information Manager (PIM) 342, a connect module 344 and an IT policy module 346. A message application 338 can be any suitable software program that allows a user of the mobile device 100 to send and receive electronic messages, wherein messages are typically stored in the flash memory 308 of the mobile device 100. A device state module 340 provides persistence, i.e. the device state module 340 ensures that important device data is stored in persistent memory, such as the flash memory 308, so that the data is not lost when the mobile device 100 is turned off or loses power. A PIM 342 includes functionality for organizing and managing data items of interest to the user, such as, but not limited to, e-mail, contacts, calendar events, and voice mails, and may interact with the wireless network 202. A connect module 344 implements the communication protocols that are required for the mobile device 100 to communicate with the wireless infrastructure and any host system, such as an enterprise system, that the mobile device 100 is authorized to interface with. An IT policy module 346 receives IT policy data that encodes the IT policy, and may be responsible for organizing and securing rules such as the “Set Maximum Password Attempts” IT policy.
Other types of software applications or components 339 can also be installed on the mobile device 100. These software applications 339 can be pre-installed applications (i.e. other than message application 338) or third party applications, which are added after the manufacture of the mobile device 100. Examples of third party applications include games, calculators, utilities, etc.
The additional applications 339 can be loaded onto the mobile device 100 through at least one of the wireless network 202, the auxiliary I/O subsystem 312, the data port 314, the short-range communications subsystem 322, or any other suitable device subsystem 324.
The data port 314 can be any suitable port that enables data communication between the mobile device 100 and another computing device. The data port 314 can be a serial or a parallel port. In some instances, the data port 314 can be a USB port that includes data lines for data transfer and a supply line that can provide a charging current to charge the battery 330 of the mobile device 100.
For voice communications, received signals are output to the speaker 318, and signals for transmission are generated by the microphone 320. Although voice or audio signal output is accomplished primarily through the speaker 318, the display 102 can also be used to provide additional information such as the identity of a calling party, duration of a voice call, or other voice call related information.
The touch-sensitive display 102 may be any suitable touch-sensitive display, such as a capacitive, resistive, infrared, surface acoustic wave (SAW) touch-sensitive display, strain gauge, optical imaging, dispersive signal technology, acoustic pulse recognition, and so forth, as known in the art. In the presently described example, the touch-sensitive display 102 is a capacitive touch-sensitive display which includes a capacitive touch-sensitive overlay 364. The overlay 364 may be an assembly of multiple layers in a stack which may include, for example, a substrate, a ground shield layer, a barrier layer, one or more capacitive touch sensor layers separated by a substrate or other barrier, and a cover. The capacitive touch sensor layers may be any suitable material, such as patterned indium tin oxide (ITO).
The display 362 of the touch-sensitive display 102 may include a display area in which information may be displayed, and a non-display area extending around the periphery of the display area. Information is not displayed in the non-display area, which is utilized to accommodate, for example, electronic traces or electrical connections, adhesives or other sealants, and/or protective coatings around the edges of the display area.
One or more touches, also known as touch contacts or touch events, may be detected by the touch-sensitive display 102. The processor 302 may determine attributes of the touch, including a location of a touch. Touch location data may include an area of contact or a single point of contact, such as a point at or near a center of the area of contact, known as the centroid. A signal is provided to the controller 366 in response to detection of a touch. A touch may be detected from any suitable object, such as a finger, thumb, appendage, or other items, for example, a stylus, pen, or other pointer, depending on the nature of the touch-sensitive display 102. The location of the touch moves as the detected object moves during a touch. The controller 366 and/or the processor 302 may detect a touch by any suitable contact member on the touch-sensitive display 102. Similarly, multiple simultaneous touches, are detected.
One or more gestures are also detected by the touch-sensitive display 102. A gesture is a particular type of touch on a touch-sensitive display 102 that begins at an origin point and continues to an end point. A gesture may be identified by attributes of the gesture, including the origin point, the end point, the distance travelled, the duration, the velocity, and the direction, for example. A gesture may be long or short in distance and/or duration. Two points of the gesture may be utilized to determine a direction of the gesture.
An example of a gesture is a swipe (also known as a flick). A swipe has a single direction. The touch-sensitive overlay 364 may evaluate swipes with respect to the origin point at which contact is initially made with the touch-sensitive overlay 364 and the end point at which contact with the touch-sensitive overlay 364 ends rather than using each of location or point of contact over the duration of the gesture to resolve a direction.
Examples of swipes include a horizontal swipe, a vertical swipe, and a diagonal swipe. A horizontal swipe typically comprises an origin point towards the left or right side of the touch-sensitive overlay 364 to initialize the gesture, a horizontal movement of the detected object from the origin point to an end point towards the right or left side of the touch-sensitive overlay 364 while maintaining continuous contact with the touch-sensitive overlay 364, and a breaking of contact with the touch-sensitive overlay 364. Similarly, a vertical swipe typically comprises an origin point towards the top or bottom of the touch-sensitive overlay 364 to initialize the gesture, a horizontal movement of the detected object from the origin point to an end point towards the bottom or top of the touch-sensitive overlay 364 while maintaining continuous contact with the touch-sensitive overlay 364, and a breaking of contact with the touch-sensitive overlay 364.
Swipes can be of various lengths, can be initiated in various places on the touch-sensitive overlay 364, and need not span the full dimension of the touch-sensitive overlay 364. In addition, breaking contact of a swipe can be gradual in that contact with the touch-sensitive overlay 364 is gradually reduced while the swipe is still underway.
Meta-navigation gestures may also be detected by the touch-sensitive overlay 364. A meta-navigation gesture is a gesture that has an origin point that is outside the display area of the touch-sensitive overlay 364 and that moves to a position on the display area of the touch-sensitive display. Other attributes of the gesture may be detected and be utilized to detect the meta-navigation gesture. Meta-navigation gestures may also include multi-touch gestures in which gestures are simultaneous or overlap in time and at least one of the touches has an origin point that is outside the display area and moves to a position on the display area of the touch-sensitive overlay 364. Thus, two fingers may be utilized for meta-navigation gestures. Further, multi-touch meta-navigation gestures may be distinguished from single touch meta-navigation gestures and may provide additional or further functionality.
In some examples, an optional force sensor 370 or force sensors is disposed in any suitable location, for example, between the touch-sensitive display 102 and a back of the mobile device 100 to detect a force imparted by a touch on the touch-sensitive display 102. The force sensor 370 may be a force-sensitive resistor, strain gauge, piezoelectric or piezoresistive device, pressure sensor, or other suitable device. Force as utilized throughout the specification refers to force measurements, estimates, and/or calculations, such as pressure, deformation, stress, strain, force density, force-area relationships, thrust, torque, and other effects that include force or related quantities.
Force information related to a detected touch may be utilized to select information, such as information associated with a location of a touch. For example, a touch that does not meet a force threshold may highlight a selection option, whereas a touch that meets a force threshold may select or input that selection option. Selection options include, for example, displayed or virtual keys of a keyboard; selection boxes or windows, e.g., “cancel,” “delete,” or “unlock”; function buttons, such as play or stop on a music player; and so forth. Different magnitudes of force may be associated with different functions or input. For example, a lesser force may result in panning, and a higher force may result in zooming.
Referring to
The mobile device 100a shown in
The mobile device 100a in
It will be appreciated that for the mobile device 100, a wide range of one or more positioning or cursor/view positioning mechanisms such as a touch/track pad, a positioning wheel, a joystick button, a mouse, a touch-sensitive display, a set of arrow keys, a tablet, an accelerometer (for sensing orientation and/or movements of the mobile device 100 etc.), OLED, or other whether presently known or unknown may be employed. Similarly, any variation of keyboard 420a may be used. It will also be appreciated that the mobile devices 100 shown in
A front view of an example of the mobile device 100b is shown in
In the example of
For the purpose of the present example, the touch-sensitive overlay 364 extends to cover the display area 506 and the non-display area 508. Touches on the display area 506 may be detected and, for example, may be associated with displayed selectable features. Touches on the non-display area 508 may be detected, for example, to detect a meta-navigation gesture. Alternatively, meta-navigation gestures may be determined by both the non-display area 508 and the display area 506. The density of touch sensors may differ from the display area 506 to the non-display area 508. For example, the density of nodes in a mutual capacitive touch-sensitive display, or density of locations at which electrodes of one layer cross over electrodes of another layer, may differ between the display area 506 and the non-display area 508.
Gestures received on the touch-sensitive display 102 may be analyzed based on the attributes to discriminate between meta-navigation gestures and other touches, or non-meta navigation gestures. Meta-navigation gestures may be identified when the gesture crosses over a boundary near a periphery of the display 362, such as a boundary 510 between the display area 506 and the non-display area 508. In the example of
A buffer region 512 or band that extends around the boundary 510 between the display area 506 and the non-display area 508 may be utilized such that a meta-navigation gesture is identified when a touch has an origin point outside the boundary 510 and the buffer region 512 and crosses through the buffer region 512 and over the boundary 510 to a point inside the boundary 510. Although illustrated in
Gestures that have an origin point in the buffer region 512, for example, may be identified as non-meta navigation gestures. Optionally, data from such gestures may be utilized by an application as a non-meta navigation gesture. Alternatively, data from such gestures may be discarded such that touches that have an origin point on the buffer region 512 are not utilized as input at the mobile device 100.
The touch 538 begins at the origin point outside the boundary 510 and outside the buffer region 512. The path of the touch 538 crosses the buffer region 512 and the boundary 510 and is therefore identified as a meta-navigation gesture. Similarly, the touches 520, 530, 524, 522, 526, 540, 534 each have origin points outside the boundary 510 and the buffer region 512 and their paths cross the buffer region 512 and the boundary 510. Each of the touches 520, 530, 524, 522, 526, 540, 534 is therefore identified as a meta-navigation gesture. The touch 528, however, has an origin point that falls within the buffer region 512 and the touch 528 is therefore not identified as a meta-navigation gesture. The touch 536 begins at an origin point outside the boundary 510 and the buffer region 512. The path of the touch 536, however, does not cross the boundary 510 and is therefore not identified as a meta-navigation gesture. The touch 532 also has an origin point outside the boundary 510 and the buffer region 512 but is not a gesture and therefore does not cross the boundary 510 and is not identified as a meta-navigation gesture.
Referring to
An evaluate gesture module 701 is in communication with the touch-sensitive display 102. The evaluate gesture module 701 is also in communication with a multi-segment character module 702. The multi-segment character module 702 includes numerical characters 703 and alphabetic characters 704. Although Arabic numerals and the English alphabet is used as common examples, other characters specific to other languages or fields (e.g. music, science, math, etc.) can also benefit from the principles described herein. In other words, other characters can be used. The evaluate gesture module 701 evaluates touch gestures detected by the touch-sensitive display 102. The module 701 evaluates text gestures in the context of a multi-segment display. It should be appreciated that the touch-sensitive display 102 displays a multi-segment display, which provides the context of a touch gesture. The module 701 evaluates touch gestures based on a gesture library 705 as well as a mappings library 706 that includes mappings between user modifications and characters.
The multi-segment character module 702 can display numbers or alphabets, or other symbols which can be displayed on a multi-segment display. One or more applications 707 will use the multi-segment character module 702 to receive inputs to select numbers or alphabets, or characters in general. Examples of such applications that may use the multi-segment character module 702, include a clock and alarm application 708, a calendar application 709, an email application 710, and a calculator application 711. For example, a clock application will allow a user to use touch gestures over a multi-segment display to select the time. In another example, a calendar application will allow a user to use touch gestures over multi-segment display to input a date. In another example, a calculator application will allow a user to use touch gestures over multi-segment displays to enter in numbers.
It will be appreciated that any module, subsystem component exemplified herein that executes instructions or operations may include or otherwise have access to computer readable media such as storage media, computer storage media, or data storage devices (removable and/or non-removable) such as, for example, magnetic disks, optical disks, or tape. Computer storage media may include volatile and non-volatile, removable and non-removable media implemented in any method or technology for storage of information, such as computer readable instructions, data structures, program modules, or other data, except transitory propagating signals per se. Examples of computer storage media include RAM, ROM, EEPROM, flash memory or other memory technology, CD-ROM, digital versatile disks (DVD) or other optical storage, magnetic cassettes, magnetic tape, magnetic disk storage or other magnetic storage devices, or any other medium which can be used to store the desired information and which can be accessed by an application, module, or both. Any such computer storage media may be part of the mobile device 100 or accessible or connectable thereto. Any application or module herein described may be implemented using computer readable/executable instructions or operations that may be stored or otherwise held by such computer readable media.
Multi-segment displays herein include, for example, seven-segment displays, nine-segment displays, fourteen-segment displays and sixteen-segment displays, for example. Examples of seven-segment displays are shown in
Although the examples shown in the figures illustrate seven-segment displays, it can be appreciated that other multi-segment displays may be used to suit the application.
Turning to
Turning to
Turning to
Turning to
Turning to
Turning to
Turning to
Turning to
Turning to
Turning to
For setting the day, there is a first and a second multi-segment displays. The touch gesture 2003 is a vertical swipe down the middle of the second multi-segment display, and this results in displaying the numbers ‘03’ for the day.
For setting the year, the gestures 2004 and 2005 are used to activate segments representing a ‘1’ and a ‘3’.
Turning to
It can also be appreciated that the touch gestures can have different sequences to represent the same characters. For example, turning to
In an example embodiment, various touch gestures can be used to show the same character. A mapping of the touch gestures in relation to a multi-segment display can be used to identify the appropriate character.
Examples of rules 2201 for mapping touch gestures numbers for a seven segment display is shown in
In other example embodiments, characters can be recommended by highlighting certain segments of a multi-segment display. A recommended character or recommended characters can be based on commonly used times or dates.
Turning to
Turning to
a), 25(b) and 25(c) are similar to
a), 26(b) and 26(c) also show the different stages for setting a time value ‘11:00’. In
As another example embodiment, a suggested character for a multi-segment display can be shown above the multi-segment display.
For example, in
a), 28(b) and 28(c) show another example in which a number ‘0’ is shown above a multi-segment display as a suggested value. However, as shown in
a), 29(b), 29(c) and 29(d) show another example of selecting a suggested character to be displayed in place of a multi-segment display. After receiving an initial touch gesture on a first multi-segment display, as shown in
As an alternative example to
Turning to
Turning to
Turning to
The above principles for displaying a suggested character and selecting the same can be used for dates, telephone numbers, money values, names, postal or zip codes, etc. The suggested characters are determined based on the context of the application.
The suggested characters can also be take into account the habits of the user. For example, in the context of time, if a user consistently schedules meetings to start at 35th minute of an hour (e.g. 10:35 am, 1:30 pm, 2:35 pm, etc.), then the mobile device 100 will suggest the minute value ‘35’. For example, if the mobile device detects that a user is scheduling a meeting and has activated segments to display a ‘3’ for the first minute digit, then the mobile device will highlight segments for the second minute digit that correspond to the suggested number ‘5’. When a tap gesture is received on the second minute digit, the number ‘5’ is displayed, so that the minute value ‘35’ is set in the display. In other words, the suggested character is associated with the displayed character based on data previously provided by a user.
Other suggested characters for other contexts can also be learned by the mobile device according to the user's inputs or habits (e.g. repeated actions). For example, when inserting alphabets for a month, if the mobile device has detected that the user has entered in the characters ‘DEC’ before on previous occasions, the mobile device will highlight segments in the multi-segment displays to suggest the letters ‘D’, ‘E’ or ‘C’, or the combination of those letters.
In another example embodiment, if two or more multi-segment displays are shown, if one of the displays is used, the other display or the other displays may automatically disappear, or show a value of ‘0’.
Turning to
As per
Alternatively to
In other words, the user does not need to provide an input that the value for the first multi-segment display 3301 is null (e.g. as per
Turning to
In another example embodiment, when a user provides a touch gesture over a multi-segment display, the displayed character may not be correct or as the user intended. The mobile device then provides a control to allow a user to modify the displayed character. The modified character can be saved in relation to the touch gesture. In future use, when the same touch gesture is detected again, the modified character is displayed.
Turning to
The mobile device saves a mapping of the gesture 3502 in relation to the modified number ‘3’. Therefore, next time a similar gesture is detected, a ‘3’ will be displayed instead of a ‘2’. In another example embodiment, the occurrence of the gesture and the modified number ‘3’ is detected a predetermined number of times before the mapping is established, so that when the gesture is detected, the modified number is automatically displayed.
In another example, the user can tap below the number ‘2’ 3503 to display the modified number ‘3’. In other words, tapping above the displayed number shows either the preceding number or the proceeding number. Conversely, tapping below the displayed number shows either the proceeding number or the preceding number.
Turning to
Turning to
The suggested characters depend on the context of the application. In the example of numbers for the minute value, the character series consists of the sequence [1, 2, 3, . . . , 9, 0]. In the example of the first digit d1 in the days format d1d2, more commonly expressed as ‘dd’, the character series consists of the sequence [0, 1, 2, 3], since there will never be more than forty days in a month. In another example, if the characters are alphabets, the sequence of characters is [A, B, C, . . . , Z].
Turning to
Examples of implementing 3801 are provided in 3804, 3805 and 3806. In particular, at 3804, the mobile device detects a tap gesture above or below the displayed character to select a preceding or proceeding character in a character series. At 3805, the mobile device detects a slide or a swipe gesture on the displayed character, in an upward or downward direction. The mobile device then displays the preceding or proceeding character in a character series. At 3806, the mobile device displays a preceding or proceeding character in a character series, and detects a flick or swipe gesture that moves the preceding or proceeding character towards the displayed character. The mobile device then displays the preceding or proceeding character.
Turning to
The methods, user interfaces and systems described above allow a user to select characters using an interface that uses less real-estate on the mobile device's touch screen display. Furthermore, using the above approaches, a user can more quickly and more accurately input the desired character for display on using a touch screen.
Example embodiments and example aspects of the systems and methods are described below.
An example method for displaying a character on an electronic device is provided. The method includes: displaying a multi-segment display; detecting a touch gesture overlaid the multi-segment display; activating segments of the multi-segment display that correspond to the touch gesture; and displaying the character that corresponds to the activated segments.
In an aspect, the multi-segment display is initially displayed in a neutral mode. In another aspect, the method further includes highlighting the activated segments. In another aspect, the method further includes no longer displaying the multi-segment display when the character is displayed. In another aspect, the multi-segment display is initially displayed in a neutral mode and the method further includes: highlighting the activated segments; and returning the multi-segment display to the neutral mode when the character is displayed overlaid the multi-segment display. In another aspect, the character is displayed in a different font than the multi-segment display. In another aspect, the segments are activated by detecting that the touch gesture position coincides with the segments. In another aspect, the multi-segment display is used to select a time or a date. In another aspect, the multi-segment display is a seven-segment display. In another aspect, the method further includes receiving an input to change the displayed character to a modified character; displaying the modified character; and saving a mapping associating the touch gesture with the modified character. In another aspect, the input includes receiving another touch gesture to display the modified character, and the modified character is a preceding character or a proceeding character relative to the displayed character in a character series. In another aspect, the method further includes displaying the modified character when detecting the same touch gesture again. In another aspect, another multi-segment display is displayed, and the method further includes: highlighting segments on the other multi-segment display that correspond to a suggested character; detecting a selection input on the other multi-segment display; and displaying the suggested character. In another aspect, the selection input is a tap gesture.
An example method for displaying a character on an electronic device is provided. The method includes: displaying a first multi-segment display and a second multi-segment display; detecting a touch gesture overlaid one of the first and the second multi-segment displays; activating segments of the one of the first and the second multi-segment displays, the segments corresponding to the touch gesture; displaying a first character associated with the one of the first and the second multi-segment displays, the first character corresponding to the activated segments; and modifying the other one of the first and the second multi-segment displays.
In another aspect, the other one of the first and the second multi-segment displays is modified based on the first character. In another aspect, the other one of the first and the second multi-segment displays is modified to display a suggested character, the suggested character associated with the first character. In another aspect, the method further includes receiving an input to display a second character equivalent to the suggested character in association with the other one of the first and the second multi-segment displays. In another aspect, the suggested character is associated with the first character based on data previously provided by a user.
In another aspect, the one of the first and the second multi-segment displays is the first multi-segment display and the other one of the first and the second multi-segment displays is the second multi-segment display; the first character displayed is a numeral ‘3’; and the second multi-segment display is modified by highlighting segments to display a numeral ‘0’.
In another aspect, the one of the first and the second multi-segment displays is the first multi-segment display and the other one of the first and the second multi-segment displays is the second multi-segment display; the first character displayed is a numeral ‘1’; and the second multi-segment display is modified by highlighting segments to display a numeral ‘5’.
In another aspect, the one of the first and the second multi-segment displays is the second multi-segment display and the other one of the first and the second multi-segment displays is the first multi-segment display; and the first multi-segment display is modified by displaying a numeral ‘0’ or modified by no longer displaying the first multi-segment display.
An example electronic device is provided and it includes a processor, a touch-sensitive input mechanism and memory. The memory stores computer executable instructions for: displaying a multi-segment display; detecting a touch gesture overlaid the multi-segment display; activating segments of the multi-segment display that correspond to the touch gesture; and displaying the character that corresponds to the activated segments.
It will be appreciated that the examples and corresponding diagrams used herein are for illustrative purposes only. Different configurations and terminology can be used without departing from the principles expressed herein. For instance, components and modules can be added, deleted, modified, or arranged with differing connections without departing from these principles.
The steps or operations in the flow charts and diagrams described herein are just for example. There may be many variations to these steps or operations without departing from the spirit of the invention or inventions. For instance, the steps may be performed in a differing order, or steps may be added, deleted, or modified.
Although the above has been described with reference to certain specific examples, various modifications thereof will be apparent to those skilled in the art as outlined in the appended claims.