This disclosure relates generally to electronic devices and corresponding methods, and more particularly to physically deformable electronic devices.
Mobile electronic communication devices, such as smartphones, are used by billions of people. These users employ mobile communication devices for many different purposes including, but not limited to, voice communications and data communications for text messaging, Internet browsing, commerce such as banking, and social networking.
As the technology of these devices has advanced, so too has their feature set. For example, not too long ago all electronic devices had physical keypads. Today touch sensitive displays are more frequently seen as user interface devices. Today some devices are even equipped with voice recognition that allows a user to speak commands to a device instead of typing them.
As the technology evolves, conventional techniques for manipulating, controlling, and otherwise interacting with improved electronic devices is sometimes inefficient, tedious, and cumbersome. It would be advantageous to have improved control and usage modes for electronic devices that are able to adapt performance to a given environment, condition, or application.
The accompanying figures, where like reference numerals refer to identical or functionally similar elements throughout the separate views and which together with the detailed description below are incorporated in and form part of the specification, serve to further illustrate various embodiments and to explain various principles and advantages all in accordance with the present disclosure.
Skilled artisans will appreciate that elements in the figures are illustrated for simplicity and clarity and have not necessarily been drawn to scale. For example, the dimensions of some of the elements in the figures may be exaggerated relative to other elements to help to improve understanding of embodiments of the present disclosure.
Before describing in detail embodiments that are in accordance with the present disclosure, it should be observed that the embodiments reside primarily in combinations of method steps and apparatus components related to controlling, manipulating, or actuating content or user actuation targets presented on one portion of a deformed touch-sensitive display by interacting with another portion of the deformed touch-sensitive display. Any process descriptions or blocks in flow charts should be understood as representing modules, segments, or portions of code which include one or more executable instructions for implementing specific logical functions or steps in the process. Alternate implementations are included, and it will be clear that functions may be executed out of order from that shown or discussed, including substantially concurrently or in reverse order, depending on the functionality involved. Accordingly, the apparatus components and method steps have been represented where appropriate by conventional symbols in the drawings, showing only those specific details that are pertinent to understanding the embodiments of the present disclosure so as not to obscure the disclosure with details that will be readily apparent to those of ordinary skill in the art having the benefit of the description herein.
Embodiments of the disclosure do not recite the implementation of any commonplace business method aimed at processing business information, nor do they apply a known business process to the particular technological environment of the Internet. Moreover, embodiments of the disclosure do not create or alter contractual relations using generic computer functions and conventional network operations. Quite to the contrary, embodiments of the disclosure employ methods that, when applied to electronic device and/or user interface technology, improve the functioning of the electronic device itself by and improving the overall user experience to overcome problems specifically arising in the realm of the technology associated with electronic device user interaction.
It will be appreciated that embodiments of the disclosure described herein may be comprised of one or more conventional processors and unique stored program instructions that control the one or more processors to implement, in conjunction with certain non-processor circuits, some, most, or all of the functions of controlling or interacting with a deformed touch-sensitive display as described herein. The non-processor circuits may include, but are not limited to, a radio receiver, a radio transmitter, signal drivers, clock circuits, power source circuits, and user input devices. As such, these functions may be interpreted as steps of a method to perform control or actuation of the deformed touch sensitive display. Alternatively, some or all functions could be implemented by a state machine that has no stored program instructions, or in one or more application specific integrated circuits (ASICs), in which each function or some combinations of certain of the functions are implemented as custom logic. Of course, a combination of the two approaches could be used. Thus, methods and means for these functions have been described herein. Further, it is expected that one of ordinary skill, notwithstanding possibly significant effort and many design choices motivated by, for example, available time, current technology, and economic considerations, when guided by the concepts and principles disclosed herein will be readily capable of generating such software instructions and programs and ICs with minimal experimentation.
Embodiments of the disclosure are now described in detail. Referring to the drawings, like numbers indicate like parts throughout the views. As used in the description herein and throughout the claims, the following terms take the meanings explicitly associated herein, unless the context clearly dictates otherwise: the meaning of “a,” “an,” and “the” includes plural reference, the meaning of “in” includes “in” and “on.” Relational terms such as first and second, top and bottom, and the like may be used solely to distinguish one entity or action from another entity or action without necessarily requiring or implying any actual such relationship or order between such entities or actions.
As used herein, components may be “operatively coupled” when information can be sent between such components, even though there may be one or more intermediate or intervening components between, or along the connection path. The terms “substantially” and “about” are used to refer to dimensions, orientations, or alignments inclusive of manufacturing tolerances. Thus, a “substantially orthogonal” angle with a manufacturing tolerance of plus or minus two degrees would include all angles between 88 and 92, inclusive. Also, reference designators shown herein in parenthesis indicate components shown in a figure other than the one in discussion. For example, talking about a device (10) while discussing figure A would refer to an element, 10, shown in figure other than figure A.
As noted above, in many modern electronic devices, touch-sensitive displays are included instead of physical keyboards. These touch-sensitive displays serve as the primary user interface for the electronic device. Embodiments of the disclosure contemplate that when a user is manipulating such a touch-sensitive display, one or more of the user's finger or the user's hand can obscure the display during the interaction. This obstruction of content or user actuation targets can negatively affect the user experience. Moreover, the visual obstruction can potentially make it difficult to perform an operation. Illustrating by example, when a user is editing text, the fact that their finger covers a word being edited makes it difficult to make accurate changes.
Embodiments of the disclosure advantageously work to solve this problem by providing an electronic device with a flexible display that can be physically deformed by one or more bends or folds. A deformable housing supports the flexible display. Using the flexible display, embodiments of the disclosure advantageously leverage the touch-sensitivity of the flexible display to allow control of a first portion of the flexible display, e.g., the portion to the right of a bend or deformation location, by interacting with a second portion of the flexible display, e.g., the portion to the left of a bend or deformation location. Embodiments of the disclosure advantageously map touch input on one side of the deformed device to control the other, thereby providing an intuitive user-interface that is easily controlled, and that obviates the “finger obstruction” problem described above.
In one or more embodiments, a flexible display includes a touch sensor that maps user input along a first portion of the flexible display disposed to a first side of the deformation location to control a second portion of the flexible display disposed to a second side of the deformation location. In one or more embodiments, touch input at the first portion of the flexible display provides control features, such as tracking, single taps, double taps, swipes, and so forth, along the second portion of the flexible display. Tracking of finger locations along the first portion can be rendered along the second portion so that they are easily seen, such as by a ring, color shift, cursor, or other visual indicia. A user can actuate a user actuation target by tapping a location on the first portion of the flexible display that is mapped to a corresponding location on the second portion of the flexible display.
In one or more embodiments, the housing of the electronic device is deformable. Internal and external components can be flexible as well. For instance, flexible batteries and flexible circuit boards can support various components within the electronic device. Touch sensors and substrates can be flexible as well. Remaining or other components disposed within the electronic device, such as one or more processors, other sensors, and other devices, are arranged such that a user can flex, bend, and/or fold the electronic device by executing a bending operation that physically deforms one or more of the housing or display into a deformed geometry.
In other embodiments, the housing may include rigid components that are linked together by one or more hinges. Such hinges can provide a solution offering needed system flexibility by providing support and movement for the flexible display during bending or folding operations. A multi-link hinge with support beams disposed beneath the flexible display, for example, can support the flexible display while allowing portions of the housing to pivot about an axis of the hinge.
In one or more embodiments, when the flexible display is deformed by one or more bends at a deformation portion, one or more processors operable with the flexible display are configured to divide the flexible display into a first portion disposed to one side of the bend and a second portion disposed to a second side of the bend. In one or more embodiments, the one or more processors then present content on one of the first portion or the second portion, detect user input along another of the first portion or the second portion, and control the content in response to the user input.
For instance, when the electronic device is folded in half with one side facing the user, i.e., the “front side,” and another side facing away from the user, i.e, the rear side, touch input along the rear side provides a tracking state indicating where the user's finger is along the rear side on the front side via a ring, cursor, or color shift. Actuation of user actuation targets presented on the front side can be accomplished, in one embodiment, by releasing and tapping the same spot on the rear side. In one or more embodiments, rear-side touch of the deformed, flexible display can be calibrated in accordance with user preferences. This calibration can be refined during actual use to make control and actuation more accurate and efficient.
Rear-side touch and front-side control can further, in one or more embodiments, be activated and/or deactivated based on how the electronic device is used. Illustrating by example, in one or more embodiments an image capture device can detect the presence of a user to one side of the electronic device. The one or more processors, in response to this detection, can then present content, information, and user actuation targets along the side facing the user, while using the other side only for control and omitting the presentation of content, information, and user actuation targets on the rear side. In another embodiment, rear-side touch and front-side content display are activated and/or deactivated when the user selects which side will be designated as the “front side” by a user interface interaction such as swiping the active “presentation” display side to the opposite side as desired.
The deformability of embodiments of the disclosure not only offer unique ways of controlling content with rear-side touch and front-side presentation, but can additionally make the device easier to use. For example, by being able to stand an otherwise thin electronic device on its side, the display can be easily viewable despite the fact that the electronic device is out of the user's hand. Bending the device into folded and multifold shapes allows the device to transform into a “self-standing” device, which can free a user's hands for other activities.
In one or more embodiments, the electronic device can be bent or deformed into different deformed geometries. Illustrating by example, if the electronic device is bent with a single fold, when placed on a table the electronic device can resemble a card folded into a “tent fold.” Where this occurs, one or more processors of the electronic device can partition the display into two parts, with each part being on a different side of the “tent,” using one for user input receipt and the other for the presentation of content, information, and user actuation targets. In other embodiments, a number of bends can be used to partition the display. Where this is the case, the one or more processors may be able to present four images in a predefined aspect ratio to provide rear-side user input reception and front-side content presentation capabilities to two different users. More permutations are completely possible. Still others will be obvious to those of ordinary skill in the art having the benefit of this disclosure.
Turning now to
In one embodiment, the display 102 is configured as an organic light emitting diode (OLED) display fabricated on a flexible plastic substrate. However, it should be noted that other types of displays would be obvious to those of ordinary skill in the art having the benefit of this disclosure. In one or more embodiments, an OLED is constructed on flexible plastic substrates can allow the display 102 to become flexible in one or more embodiments with various bending radii. For example, some embodiments allow bending radii of between thirty and six hundred millimeters to provide a bendable display. Other substrates allow bending radii of around five millimeters to provide a display that is foldable through active bending. Other displays can be configured to accommodate both bends and folds. In one or more embodiments the display 102 may be formed from multiple layers of flexible material such as flexible sheets of polymer or other materials.
The explanatory electronic device 100 of
Where the housing 101 is a deformable housing, it can be manufactured from a single flexible housing member or from multiple flexible housing members. In this illustrative embodiment, a user interface component 114, which may be a button or touch sensitive surface, can also be disposed along the housing 101 to facilitate control of the electronic device 100. Other features can be added, and can be located on the front of the housing 101, sides of the housing 101, or the rear of the housing 101. Illustrating by example, in one or more embodiments a first image capture device 135 can be disposed on one side of the electronic device 100, while a second image capture device 136 is disposed on another side of the electronic device 100.
A block diagram schematic 115 of the electronic device 100 is also shown in
In one or more embodiments when the electronic device 100 is deformed by a bend at a deflection portion, the one or more processors 116 divide the display 102 into a first portion disposed to one side of the bend and a second portion disposed to a second side of the bend. As will be described in more detail below, in one or more embodiments the one or more processors 116 then present content on one of the first portion or the second portion, while detecting user input along another of the first portion or the second portion. For example, the one or more processors 116 can present content on a portion to the left of the bend, while receiving user input on a portion to the right of the bend. In one or more embodiments, the one or more processors 116 then control the content in response to the user input. Accordingly, the user can touch the display 102 to one side of the bend and control content presented on another side of the bend.
In one or more embodiments, the one or more processors 116 are further responsible for performing the primary functions of the electronic device 100. For example, in one embodiment the one or more processors 116 comprise one or more circuits operable to present presentation information, such as images, text, and video, on the display 102. The executable software code used by the one or more processors 116 can be configured as one or more modules 120 that are operable with the one or more processors 116. Such modules 120 can store instructions, control algorithms, and so forth.
In one embodiment, the one or more processors 116 are responsible for running the operating system environment 121. The operating system environment 121 can include a kernel, one or more drivers 122, and an application service layer 123, and an application layer 124. The operating system environment 121 can be configured as executable code operating on one or more processors or control circuits of the electronic device 100.
In one or more embodiments, the one or more processors 116 are responsible for managing the applications of the electronic device 100. In one or more embodiments, the one or more processors 116 are also responsible for launching, monitoring and killing the various applications and the various application service modules. The applications of the application layer 124 can be configured as clients of the application service layer 123 to communicate with services through application program interfaces (APIs), messages, events, or other inter-process communication interfaces.
In this illustrative embodiment, the electronic device 100 also includes a communication circuit 125 that can be configured for wired or wireless communication with one or more other devices or networks. The networks can include a wide area network, a local area network, and/or personal area network. Examples of wide area networks include GSM, CDMA, W-CDMA, CDMA-2000, iDEN, TDMA, 2.5 Generation 3GPP GSM networks, 3rd Generation 3GPP WCDMA networks, 3GPP Long Term Evolution (LTE) networks, and 3GPP2 CDMA communication networks, UMTS networks, E-UTRA networks, GPRS networks, iDEN networks, and other networks.
The communication circuit 125 may also utilize wireless technology for communication, such as, but are not limited to, peer-to-peer or ad hoc communications such as HomeRF, Bluetooth and IEEE 802.11 (a, b, g or n), and other forms of wireless communication such as infrared technology. The communication circuit 125 can include wireless communication circuitry, one of a receiver, a transmitter, or transceiver, and one or more antennas 126.
As shown in
In one embodiment, the electronic device 100 includes one or more flex sensors 112, supported by the housing 101 and operable with the one or more processors 116, to detect a bending operation deforming one or more of the housing 101 or the display 102 into a deformed geometry, such as that shown in
In one embodiment, the flex sensors 112 each comprise passive resistive devices manufactured from a material with an impedance that changes when the material is bent, deformed, or flexed. By detecting changes in the impedance as a function of resistance, the one or more processors 116 can use the one or more flex sensors 112 to detect bending or flexing. In one or more embodiments, each flex sensor 112 comprises a bi-directional flex sensor that can detect flexing or bending in two directions. In one embodiment, the one or more flex sensors 112 have an impedance that increases in an amount that is proportional with the amount it is deformed or bent.
In one embodiment, each flex sensor 112 is manufactured from a series of layers combined together in a stacked structure. In one embodiment, at least one layer is conductive, and is manufactured from a metal foil such as copper. A resistive material provides another layer. These layers can be adhesively coupled together in one or more embodiments. The resistive material can be manufactured from a variety of partially conductive materials, including paper-based materials, plastic-based materials, metallic materials, and textile-based materials. In one embodiment, a thermoplastic such as polyethylene can be impregnated with carbon or metal so as to be partially conductive, while at the same time being flexible.
In one embodiment, the resistive layer is sandwiched between two conductive layers. Electrical current flows into one conductive layer, through the resistive layer, and out of the other conductive layer. As the flex sensor 112 bends, the impedance of the resistive layer changes, thereby altering the flow of current for a given voltage. The one or more processors 116 can detect this change to determine an amount of bending. Taps can be added along each flex sensor 112 to determine other information, including the number of folds, the degree of each fold, the location of the folds, the direction of the folds, and so forth. The flex sensor 112 can further be driven by time-varying signals to increase the amount of information obtained from the flex sensor 112 as well.
While a multi-layered device as a flex sensor 112 is one configuration suitable for detecting a bending operation occurring to deform the electronic device 100 and a geometry of the electronic device 100 after the bending operation, others can be used as well. For example, in another embodiment the proximity sensors can be used to detect how far a first end of the electronic device 100 is from a second end of the electronic device 100. In yet another embodiment, Hall effect sensors can be used to sense open and/or closed state of the electronic device 100 as well. Still other types of flex sensors 112 will be obvious to those of ordinary skill in the art having the benefit of this disclosure.
In one or more embodiments, each of the first image capture device 135 and the second image capture device 136 comprises an intelligent imager 138. An intelligent imager 138 can capture one or more images of environments about the electronic device 100 and determine whether the object matches predetermined criteria. For example, the intelligent imager 138 operate as an identification module configured with optical recognition such as include image recognition, character recognition, visual recognition, facial recognition, color recognition, shape recognition and the like. Advantageously, the intelligent imager 138 can recognize whether a user's face or eyes are disposed to a first side of the electronic device 100 when it is folded or to a second side. Similarly, the intelligent imager 138, in one embodiment, can detect whether the user 130 is gazing toward a portion of the display 102 disposed to a first side of a bend or another portion of the display 102 disposed to a second side of a bend. In yet another embodiment, the intelligent imager 138 can determine where a user's eyes or face are located in three-dimensional space relative to the electronic device 100.
In addition to, or instead of the intelligent imager 138, one or more proximity sensors 139 can determine to which side of the electronic device 100 the user 130 is positioned when the electronic device 100 is deformed. The proximity sensors 139 can include one or more proximity sensor components. The proximity sensors 139 can also include one or more proximity detector components. In one embodiment, the proximity sensor components comprise only signal receivers. By contrast, the proximity detector components include a signal receiver and a corresponding signal transmitter.
While each proximity detector component can be any one of various types of proximity sensors, such as but not limited to, capacitive, magnetic, inductive, optical/photoelectric, imager, laser, acoustic/sonic, radar-based, Doppler-based, thermal, and radiation-based proximity sensors, in one or more embodiments the proximity detector components comprise infrared transmitters and receivers. The infrared transmitters are configured, in one embodiment, to transmit infrared signals having wavelengths of about 860 nanometers, which is one to two orders of magnitude shorter than the wavelengths received by the proximity sensor components. The proximity detector components can have signal receivers that receive similar wavelengths, i.e., about 860 nanometers.
In one or more embodiments the proximity sensor components have a longer detection range than do the proximity detector components due to the fact that the proximity sensor components detect heat directly emanating from a person's body (as opposed to reflecting off the person's body) while the proximity detector components rely upon reflections of infrared light emitted from the signal transmitter. For example, the proximity sensor component may be able to detect a person's body heat from a distance of about ten feet, while the signal receiver of the proximity detector component may only be able to detect reflected signals from the transmitter at a distance of about one to two feet.
In one embodiment, the proximity sensor components comprise an infrared signal receiver so as to be able to detect infrared emissions from a person. Accordingly, the proximity sensor components require no transmitter since objects disposed external to the housing 101 of the electronic device 100 deliver emissions that are received by the infrared receiver. As no transmitter is required, each proximity sensor component can operate at a very low power level. Evaluations show that a group of infrared signal receivers can operate with a total current drain of just a few microamps (˜10 microamps per sensor). By contrast, a proximity detector component, which includes a signal transmitter, may draw hundreds of microamps to a few milliamps.
In one embodiment, one or more proximity detector components can each include a signal receiver and a corresponding signal transmitter. The signal transmitter can transmit a beam of infrared light that reflects from a nearby object and is received by a corresponding signal receiver. The proximity detector components can be used, for example, to compute the distance to any nearby object from characteristics associated with the reflected signals. The reflected signals are detected by the corresponding signal receiver, which may be an infrared photodiode used to detect reflected light emitting diode (LED) light, respond to modulated infrared signals, and/or perform triangulation of received infrared signals.
In one embodiment, the one or more processors 116 may generate commands or execute control operations based on information received from the various sensors, including the one or more flex sensors 112, the user interface 113, or the other sensors 127. The one or more processors 116 may also generate commands or execute control operations based upon information received from a combination of the one or more flex sensors 112, the user interface 113, or the other sensors 127. Alternatively, the one or more processors 116 can generate commands or execute control operations based upon information received from the one or more flex sensors 112 or the user interface 113 alone. Moreover, the one or more processors 116 may process the received information alone or in combination with other data, such as the information stored in the memory 118.
The other sensors 127 may include a microphone, an earpiece speaker, a loudspeaker, key selection sensors, a touch pad sensor, a touch screen sensor, a capacitive touch sensor, and one or more switches. Touch sensors may used to indicate whether any of the user actuation targets present on the display 102 are being actuated. Alternatively, touch sensors disposed in the housing 101 can be used to determine whether the electronic device 100 is being touched at side edges or major faces of the electronic device 100 are being performed by a user 130. The touch sensors can include surface and/or housing capacitive sensors in one embodiment. The other sensors 127 can also include video sensors (such as a camera).
The other sensors 127 can also include motion detectors, such as one or more accelerometers or gyroscopes. For example, an accelerometer may be embedded in the electronic circuitry of the electronic device 100 to show vertical orientation, constant tilt and/or whether the electronic device 100 is stationary. The measurement of tilt relative to gravity is referred to as “static acceleration,” while the measurement of motion and/or vibration is referred to as “dynamic acceleration.” A gyroscope can be used in a similar fashion. In one embodiment the motion detectors are also operable to detect movement, and direction of movement, of the electronic device 100 by a user 130.
In one or more embodiments, the other sensors 127 include a gravity detector 140. For example, as one or more accelerometers and/or gyroscopes may be used to show vertical orientation, constant, or a measurement of tilt relative to gravity 141. Accordingly, in one or more embodiments, the one or more processors 116 can use the gravity detector 140 to determine an orientation of the electronic device 100 in three-dimensional space 142 relative to the direction of gravity 141. If, for example, the direction of gravity 141 flows from a first portion of the display 102 to a second portion of the display 102 when the electronic device 100 is folded, the one or more processors 116 can conclude that the first portion of the display 102 is facing upward. By contrast, if the direction of gravity 141 flows from the second portion to the first, the opposite would be true, i.e., the second portion of the display 102 would be facing upward.
Other components 128 operable with the one or more processors 116 can include output components such as video outputs, audio outputs, and/or mechanical outputs. Examples of output components include audio outputs, an earpiece speaker, haptic devices, or other alarms and/or buzzers and/or a mechanical output component such as vibrating or motion-based mechanisms. Still other components will be obvious to those of ordinary skill in the art having the benefit of this disclosure.
It is to be understood that
Now that the various hardware components have been described, attention will be turned to methods, systems, and use cases in accordance with one or more embodiments of the disclosure. Beginning with
Turning now to
In other embodiments, rather than relying upon the manual application of force, the electronic device can include a mechanical actuator 304, operable with the one or more processors (116), to deform the display 102 by one or more bends. For example, a motor or other mechanical actuator can be operable with structural components to bend the electronic device 100 to predetermined angles and physical configurations in one or more embodiments. The use of a mechanical actuator 304 allows a precise bend angle or predefined deformed physical configurations to be repeatedly achieved without the user 130 having to make adjustments. However, in other embodiments the mechanical actuator 304 will be omitted to reduce component cost.
Regardless of whether the bending operation 301 is a manual one or is instead one performed by a mechanical actuator 304, it results in the display 102 being deformed by one or more bends. One result 400 of the bending operation 301 is shown in
In one embodiment, the one or more processors (116) of the electronic device 100 are operable to detect that a bending operation 301 is occurring by detecting a change in an impedance of the one or more flex sensors (112). The one or more processors (116) can detect this bending operation 301 in other ways as well. For example, the touch sensors can detect touch and pressure from the user (130). Alternatively, the proximity sensors can detect the first side 402 and the second side 403 of the electronic device 100 getting closer together. Force sensors can detect an amount of force that the user (130) is applying to the housing 101 as well. The user (130) can input information indicating that the electronic device 100 has been bent using the display 102 or other user interface (113). Other techniques for detecting that the bending operation (301) has occurred will be obvious to those of ordinary skill in the art having the benefit of this disclosure.
Several advantages offered by the “bendability” of embodiments of the disclosure are illustrated in
In one or more embodiments, the one or more processors (116) are operable to detect the number of folds in the electronic device 100 resulting from the bending operation 301. In one embodiment, after determining the number of folds, the one or more processors (116) can partition the display 102 of the electronic device 100 as another function of the one or more folds. Since there is a single bend 401 here, in this embodiment the display 102 has been partitioned into a first portion and a second portion, with each portion being disposed on opposite sides of the “tent.”
In one or more embodiments, the bending operation 301 can continue from the physical configuration of
Turning now to
In this illustration, one or more processors (116), operable with the one or more flex sensors (112), present content 600 on portions, e.g., portion 402, of the display 102 disposed to one side of the deflection defined by the bend 401. Here the content 600 comprises a smartphone home screen and a picture of the user's dog, Buster. Additionally, user actuation targets 601,602,603,604 and other content 605 are being presented on the display 102. Accordingly, the one or more processors (116) allows the user 300 to look at his dog, Buster, while using the electronic device 100 as a smartphone at the same time.
In one or more embodiments, the one or more processors (116) then detect user input along a second portion (403) of the display 102 disposed to a second side of the deformation portion 305, which is facing the user's palm in
Turning now to
In one or more embodiments, the one or more processors (116) map the vertical and horizontal dimensions of the second portion 403 of the display 102 to the first portion 402 of the display 102. In one embodiment, this comprises mapping X and Y coordinates of the second portion 403 of the display 102 to a mirror image of those coordinates along the first portion 402 of the display 102. Accordingly, the x-y coordinates on the first portion 402 of the display are aligned to the x-y coordinates on the second portion 403 of the display by determining the geometric position of the two portions relative to the deformation portion 305. Spatial calibration or other techniques could also be used. Thus, when the user 130 moves a finger along the second portion 403 of the display 102, control 702 of content, e.g., navigation to user actuation target 704, occurs on the first portion 402 of the display 102.
In this illustrative embodiment, the one or more processors (116) also present visible indicia 705 corresponding to the touch input along the first portion 402 of the display 102. In this illustrative embodiment, the visible indicia 705 comprise a halo. However, the visible indicia 705 could also comprise a cursor 706 or visible marker 707 identifying where the user's finger is touching the second portion 403 of the display 102. Still other visible indicia will be obvious to those of ordinary skill in the art having the benefit of this disclosure.
As shown in
In one or more embodiments, the user 130 can further control the content presented upon the first portion 402 of the display 102 by delivering user input 701 to the second portion 403 of the display 102. Turning now to
As shown in
Turning now to
In this illustrative embodiment, the user 130 is also delivering other user input 904 to the first portion 402 of the display 102. Here, the content 902 comprises a music player application and the user 130 is delivering the other user input 904 to play a song. Thus, as shown, the one or more processors (116) detect the other user input 904 along the first portion 402 of the display 102 and control the content 902, e.g., play a song, in response to the other user input 904.
While the application and reason for delivering the other user input 904 will vary, this example makes it clear that the one or more processors (116) can detect, in one or more embodiments, user input 901 along the second portion 403 of the display 102 and other user input 904 along the first portion 402 of the display 102 to control content 902. Other applications where this “dual” user input may be useful would be to provide multi-touch capabilities that enable “gestures” such as pinching content 902 or pulling content 902 to create actions such as moving an icon, locking the screen, and so forth. Still other applications and gestures will be obvious to those of ordinary skill in the art having the benefit of this disclosure.
Illustrating by example, in one or more embodiments the user input 901 and the other user input 904 can be used together to make complex gestures. For example, the user input 901 may occur in a first direction, while the other user input 904 may occur in an opposite direction, with the user's fingers moving apart, thereby creating a combined “pull” gesture action to, for instance, magnify the content 902, scale the content 902, or otherwise manipulate the content 902. Similarly, the user input 901 may occur in a first direction, while the other user input 904 may occur in an opposite direction, with the user's fingers moving together, thereby creating a combined “pinch” gesture action to, for instance, reduce the content 902, descale the content 902, or otherwise manipulate the content 902.
Moreover, in other embodiments the user input 901 and the other user input 904 can be used in tandem to allow multi-touch control of the content 902. For instance, the user 130 might generate new actuation mechanisms through gesture controls to perform various electronic device operations, such as moving content 902, manipulating the content 902, locking or unlocking the electronic device 100, scrolling through content 902, or perform other operations. Thus, in one or more embodiments the combined user input 901 and other user input 902 are combined to define a multi-touch gesture to control the electronic device 100. These examples are illustrative only, as numerous others will be obvious to those of ordinary skill in the art having the benefit of this disclosure.
In
As shown in
As noted above, in one or more embodiments, when the one or more processors (116) of the electronic device 100 detect deformation of the display 102 by a bend at a deformation location, e.g., the tent fold 1001 of
As also noted above, in one or more embodiments the one or more processors (116) present content, e.g., the alarm clock indicia shown in
As shown in
In one or more embodiments, the one or more processors (116) determine whether to present the content to the first side 1102 or to the second side (facing the user's palm) by capturing images 1103 of a user 130 to determine whether the first side 1102 or the second side is facing the user 130. For example, in one embodiment the one or more processors (116) or intelligent imager (138) can analyze the images 1103 captured by the first image capture device 136 to determine if the user's face or eyes are detected. If the user's face or eyes are detected, the one or more processors (116) present content on the first side 1102 of the display 102 disposed to a first side of the deformable region 1101 after detecting a bend. Similarly, if the one or more processors (116) or intelligent imager (138) detect the user's face or eyes in images captured by the second image capture device (135), content could be presented on the second side of the display 102 disposed to a second side of the deformable region 1101 after detecting a bend.
Alternatively, since an intelligent imager (138) may consume relatively high amounts of current, one or more proximity sensors (139) can be used in conjunction with the intelligent imager (138) to conserve power. Specifically, the one or more proximity sensors (139) can detect whether the user 130 is facing the first side 1102 or the second side, and can present content on that side. The intelligent imager can confirm this position of the user to avoid false positives. Since the user 130 is holding the electronic device 100 in the hand, the proximity sensors (139) on one side will saturate, while the proximity sensors (139) on the other side will detect the user's face. Accordingly, in one embodiment, the one or more processors (116) can present content on the side having unsaturated proximity sensors (139).
In still another embodiment, an accelerometer, gravity detector (140), or gyroscope can determine the orientation of the electronic device 100 in three-dimensional space (142). The one or more processors (116) can monitor the gravity detector (140) to detect, for example, a direction of gravity 141. Since the direction of gravity 141 flows from the first side 1102 to the second side in this example, the one or more processors (116) can present content on the first side 1102, concluding that it is facing upward and toward the user's face. The examples above are illustrative only, as numerous others will be obvious to those of ordinary skill in the art having the benefit of this disclosure.
In other embodiments, the presentation of content can default to a predefined side of a deformation region defined by a bend. Since the default side may not be the side ultimately desired by the user, it can be changed in one or more embodiments. Turning now to
In
In this illustrative embodiment, the one or more processors (116) detect user input 1202 beginning on the second side of the display 102 and sliding to the first side 1102 of the display. Accordingly, the one or more processors (116) cause the content to translate 1203 from the second side to the first side of the display 102. Essentially, using a swipe gesture, the user 130 can quickly and easily cause the content to translate 1203 from one side to the other as desired. Other techniques for moving content from one side of a bend 1201 to the other will be obvious to those of ordinary skill in the art having the benefit of this disclosure.
Turning now to
Where the electronic device is deformed at a bend or deformation region, in one embodiment the method 1300 detects a user's presence to one side of the bend or deformation region at step 1303. As described above this step 1303 can be performed in any of a number of ways, including using an intelligent imager, proximity sensors, gravity detectors, or other techniques.
At step 1304, once the user's presence to one side of the bend or deformation region is detected, the method 1300 assigns the portion of the display disposed to that side of the bend or deformation region as the “front side” and presents content on the front side. At step 1305, the method 1300 omits the presentation of content on the other side, i.e., the “rear side,” and instead receives touch input. At step 1305, this touch input is mapped from the rear side to the front side as well. In one embodiment, this mapping comprises reflecting the touch input to the front side.
At step 1306, the method can include monitoring and calibration of the mapping. For example, a user may refold or re-bend or otherwise alter or change the location of the bend or deformation region. Step 1306 monitors such changes and adjusts the mapping of touch input from the rear side to control content on the front side.
Turning now to
At optional step 1404, the method 1400 determines whether to present the content on the first side or the second side by determining which of the first side or the second side faces a user. This can be done in a variety of ways, including by using an intelligent imager, proximity sensors, gravity detection, spatial orientation of the electronic device, or by other techniques.
At step 1405, the method 1400 detects, with a touch sensor operable with the flexible display, touch input on another of the first side or the second side. At step 1406, the method 1400 manipulates, with the one or more processors, the content as a function of the user input.
At optional step 1407, the method 1400 includes presenting indicia of the user input on the one of the first side or the second side. The presentation of this user input can be in the form of a halo, a cursor, a cross-hair, a spyglass, a magnifying glass, a visible indicator, or by other techniques.
At step 1408, the method 1400 optionally comprises detecting other touch input along another of the first side or the second side, and moving the content from the one of the first side or the second side to the another of the first side or the second side. This allows a user to change which side of the display is presenting content and which side is receiving touch input, as previously described.
At step 1409, the method 1400 comprises detecting, with the one or more flex sensors, removal of the bend. Embodiments of the disclosure contemplate that when a deformable electronic device is bent, the user will eventually want to “unbend” it to use it in a planar mode again. Accordingly, this unbending is detected at step 1409. Additionally, step 1409 can include again uniting the first side and the second side, and presenting content across the entirety of the display.
In the foregoing specification, specific embodiments of the present disclosure have been described. However, one of ordinary skill in the art appreciates that various modifications and changes can be made without departing from the scope of the present disclosure as set forth in the claims below. Thus, while preferred embodiments of the disclosure have been illustrated and described, it is clear that the disclosure is not so limited. Numerous modifications, changes, variations, substitutions, and equivalents will occur to those skilled in the art without departing from the spirit and scope of the present disclosure as defined by the following claims. Accordingly, the specification and figures are to be regarded in an illustrative rather than a restrictive sense, and all such modifications are intended to be included within the scope of present disclosure. The benefits, advantages, solutions to problems, and any element(s) that may cause any benefit, advantage, or solution to occur or become more pronounced are not to be construed as a critical, required, or essential features or elements of any or all the claims.