An example embodiment of the invention relates generally to user interface technology and, more particularly, relates to a method, apparatus, and computer program product for providing a user-friendly and efficient manner in which to manage different visual variants of objects via a user interface.
The modern communications era has brought about a tremendous expansion of wireline and wireless networks. Computer networks, television networks, and telephony networks are experiencing an unprecedented technological expansion, fueled by consumer demand. Wireless and mobile networking technologies have addressed related consumer demands, while providing more flexibility and immediacy of information transfer.
Current and future networking technologies continue to facilitate ease of information transfer and convenience to users. Due to the now ubiquitous nature of electronic communication devices, people of all ages and education levels are utilizing electronic devices to communicate with other individuals or contacts, receive services and/or share information, media and other content. One area in which there is a demand to increase convenience to users relates to improving a user's ability to effectively interface with the user's communication device. Accordingly, numerous user interface mechanisms have been developed to attempt to enable a user to more easily accomplish tasks or otherwise improve the user's experience in using the device. In this regard, for example, a user's experience during certain applications such as, for example, web browsing or interactions with applications may be enhanced by using a touch screen display as the user interface.
For instance, the user interface may have application launchers in a home screen area of the user interface that typically provides a user with a way of launching, switching between, and monitoring the execution of programs or applications. These applications may be represented as displayed icons in which access is provided to functions of the communication device that may interact with an operating system.
At present, typical operating systems of smart communication devices may allow some customization layout of interactive objects of a user interface. However, currently, it may be difficult for a user to organize huge icons and widgets in a home screen and/or a main menu of a user interface of a communication device. Meanwhile, the same application or program may have different visual variants in different views. For example, an icon and a widget may be utilized to point to a same application but may be quite different in their visualization. In this regard, these visual variants may lack a tangible relationship for user understanding. As such, a user may not recognize that these different visualizations relate to the same application or program.
Although, at present, operation systems of smart communication devices allow some customization layout of icons and widgets, these existing operating systems may have some drawbacks. For example, some operating systems may allow a user to manage items in a home screen or main menu of a smart communication device in an organize mode. However, the user may only be able to manage items (e.g., icons, shortcuts) in a current level of a user interface or view and the manner in which to enable editing of the items may not be very intuitive or user friendly. For instance, a long tap to create a pop-up menu for selection of an icon or alternatively to select a new icon from a long list may be required. Also, in some current operating systems such as, for example, Android™, a user may be able to add widgets in a home menu but may lack the continuity and smoothness between a same entity's (e.g., an application) different visual variants (e.g., an icon, a widget) in different views of a user interface.
In addition, some existing operating systems such as, for example, the iPhone™ Operating System (iOS)™ may allow a user to use an icon panel 3 in the bottom of a home area as shown in
In view of the foregoing drawbacks, it may be desirable to provide an alternative mechanism in which to enable objects of a user interface to be more efficiently managed and utilized in different views of a user interface.
A method, apparatus and computer program product are therefore provided for providing a user-friendly, efficient and reliable manner in which to enable management of different visual variants of objects via a user interface of a communication device.
An example embodiment may provide a touchable user interface structure having different levels including, but not limited to, an idle screen, a home screen, an application (e.g., shell) screen, a main menu screen to applications of the user interface, etc. Additionally, an example embodiment of the invention may provide a user some freedom to customize layout, organize icon buttons, widgets, shortcuts and other interactive objects in and between different user interface views, screens, or levels. An example embodiment may provide a manner in which to enable a user to manage user interface objects and to maintain continuity between a merging area(s) of various views of a user interface. As referred to herein, a merging area(s) may, but need not, be an area of a user interface that maintains shortcuts to applications (e.g., designated applications, default applications, favorite applications, etc.). The merging area(s) may be visible in one or more screens of a user interface and may, but need not, serve as a link between the screens.
In this regard, an example embodiment may enable a same entity (e.g., an application, program, or the like) of a user interface to change one or more visual variants representing the entity based in part on a location(s) of the visual variants in different views of the user interface. For instance, based on a division of a merging area and another area/space of a specific user interface view, the same entity of the user interface may have different visual variants to represent the same entity. In one example embodiment, these different visual variants may be in different locations, views or screens of the user interface. The visual variants may be items of visible indicia, including but not limited to icons, widgets, shortcuts, a list(s), a window(s) or other visualizations. Some visual variants (e.g., an icon(s), widget(s), etc.) may correspond to the same entity (e.g., an application, a program, etc.).
In this regard, for example, an example embodiment may enable a user to drag an item of visible indicia (e.g., an object (e.g., a widget) from an area or space of a user interface to a merging area of a user interface. In response to dropping the item of visible indicia at a location (e.g., a location of an item of visual indicia (e.g., an icon) of the merging area, an example embodiment may change or transform the item of visible indicia to another suitable variant item of visible indicia (e.g., another object (e.g., an icon)).
Additionally, an example embodiment may enable an item of visible indicia (e.g., an object, (e.g., an icon)) that is being moved or replaced from the merging area to be changed or transformed to a different item of visible indicia (e.g., another object (e.g. a widget)) in response to being moved to an area/space (e.g., a home screen) outside of the merging area in a screen of the user interface. The different item of visible indicia may be a different type of visible data and in one example embodiment the different item of visible indicia (e.g., a widget) may depict visual information that is associated with functions (e.g., widget functions) of an application corresponding to the different item of visible indicia in an instance in which the application is being executed. The functions may, but need not, be widget functions including, but not limited to, TV widget functions, instant message (IM) widget functions, weather widget functions, music widget functions, and any other suitable widget functions.
The location of the merging area may be moved in different areas in one or more specific views of a user interface based in part on a user interface transition such as, for example, a selection by a user of the merging area to move the merging area to another area of a screen of the user interface. In this regard, an example embodiment may maintain continuity of the merging area in multiple views of a user interface to enable sharing of objects and functions.
By utilizing example embodiments of the invention, users may interact with user interface objects between a merging area and another space or area of a screen(s) of a user interface to organize the user interface. For instance, the user may interact with user interface objects between a merging area and another space of the screen to organize the user interface. In some example embodiments, utilization of the merging area may enable a user to switch between different applications, programs or the like.
In one example embodiment, a method for providing a user-friendly and reliable manner for managing different visual variants of objects via a user interface is provided. The method may include positioning a generated merging area at a first area of a screen of a user interface. The merging area may include one or more items of visible indicia corresponding to shortcuts to respective applications. The method may also include changing an item of visual indicia associated with an application to a transformed item of visible indicia in response to detecting movement of the item of visual indicia from a second area of the user interface to a location of an item of visible indicia of the merging area.
In another example embodiment, an apparatus for providing a user-friendly and reliable manner for managing different visual variants of objects via a user interface is provided. The apparatus may include a processor and a memory including computer program code. The memory and the computer program code are configured to, with the processor, cause the apparatus to at least perform operations including positioning a generated merging area at a first area of a screen of a user interface. The merging area may include one or more items of visible indicia corresponding to shortcuts to respective applications. The memory and the computer program code are also configured to, with the processor, cause the apparatus to change an item of visual indicia associated with an application to a transformed item of visible indicia in response to detecting movement of the item of visual indicia from a second area of the user interface to a location of an item of visible indicia of the merging area.
In another example embodiment, a computer program product for providing a user-friendly and reliable manner for managing different visual variants of objects via a user interface is provided. The computer program product includes at least one computer-readable storage medium having computer-executable program code instructions stored therein. The computer executable program code instructions may include program code instructions configured to position a generated merging area at a first area of a screen of a user interface. The merging area may include one or more items of visible indicia corresponding to shortcuts to respective applications. The program code instructions may also change an item of visual indicia associated with an application to a transformed item of visible indicia in response to detecting movement of the item of visual indicia from a second area of the user interface to a location of an item of visible indicia of the merging area.
In another example embodiment, an apparatus for providing a user-friendly and reliable manner for managing different visual variants of objects via a user interface is provided. The apparatus may include means for positioning a generated merging area at a first area of a screen of a user interface. The merging area may include one or more items of visible indicia corresponding to shortcuts to respective applications. The apparatus may also include means for changing an item of visual indicia associated with an application to a transformed item of visible indicia in response to detecting movement of the item of visual indicia from a second area of the user interface to a location of an item of visible indicia of the merging area.
An example embodiment of the invention may provide a better user experience given the ease and efficiency in enabling management of visual variants of objects via a user interface. As a result, device users may enjoy improved capabilities with respect to applications and services accessible via the device.
Having thus described the invention in general terms, reference will now be made to the accompanying drawings, which are not necessarily drawn to scale, and wherein:
Some embodiments of the invention will now be described more fully hereinafter with reference to the accompanying drawings, in which some, but not all embodiments of the invention are shown. Indeed, various embodiments of the invention may be embodied in many different forms and should not be construed as limited to the embodiments set forth herein. Like reference numerals refer to like elements throughout. As used herein, the terms “data,” “content,” “information” and similar terms may be used interchangeably to refer to data capable of being transmitted, received and/or stored in accordance with embodiments of the invention. Moreover, the term “exemplary”, as used herein, is not provided to convey any qualitative assessment, but instead merely to convey an illustration of an example. Thus, use of any such terms should not be taken to limit the spirit and scope of embodiments of the invention.
Additionally, as used herein, the term ‘circuitry’ refers to (a) hardware-only circuit implementations (e.g., implementations in analog circuitry and/or digital circuitry); (b) combinations of circuits and computer program product(s) comprising software and/or firmware instructions stored on one or more computer readable memories that work together to cause an apparatus to perform one or more functions described herein; and (c) circuits, such as, for example, a microprocessor(s) or a portion of a microprocessor(s), that require software or firmware for operation even if the software or firmware is not physically present. This definition of ‘circuitry’ applies to all uses of this term herein, including in any claims. As a further example, as used herein, the term ‘circuitry’ also includes an implementation comprising one or more processors and/or portion(s) thereof and accompanying software and/or firmware. As another example, the term ‘circuitry’ as used herein also includes, for example, a baseband integrated circuit or applications processor integrated circuit for a mobile phone or a similar integrated circuit in a server, a cellular network device, other network device, and/or other computing device.
As defined herein a “computer-readable storage medium,” which refers to a non-transitory, physical or tangible storage medium (e.g., volatile or non-volatile memory device), may be differentiated from a “computer-readable transmission medium,” which refers to an electromagnetic signal.
As referred to herein, a merging area may, but need not be, an area of a user interface that maintains shortcuts to applications (e.g., designated applications, default applications, favorite applications, etc.), which may be visible both in one or more screens (e.g., a home screen, an application screen (also referred to herein as application view, etc.) of a user interface and which may, but need not, serve as a link between the screens. The relationship of the link may, but need not, be based on (1) the previous/next steps in a use flow, (2) the upper/lower levels in a hierarchical structure, or (3) other similarities such as operating a same type of content or information.
As referred to herein, a widget(s) may be a graphical element(s), an item(s) of visual (also referred to herein as visible) indicia, information or the like of a user interface that may be displayed and may provide an interaction point(s) for accessing data associated with an application(s), program(s), or the like. As referred to herein, a pointer(s) may include, but is not limited to, one or more body parts such as, for example, a finger(s), a hand(s) etc., or a mechanical and/or electronic pointing device(s) (e.g., a stylus, pen, mouse, joystick, etc.) configured to enable a user(s) to input items of data to a communication device.
As referred to herein, a home screen, an idle screen or an auto-screen, may be a screen of a user interface that is initially enabled for display via a device of a user interface in an instance in which the device is turned on and which may remain as an active screen of the user interface even after the device is turned on.
The mobile terminal 10 may be any of multiple types of mobile communication and/or computing devices such as, for example, portable digital assistants (PDAs), pagers, mobile televisions, mobile telephones, gaming devices, wearable devices, head mounted devices, laptop computers, touch surface devices, cameras, camera phones, video recorders, audio/video players, radios, global positioning system (GPS) devices, or any combination of the aforementioned, and other types of voice and text communications systems. The network 30 may include a collection of various different nodes, devices or functions that may be in communication with each other via corresponding wired and/or wireless interfaces. As such, the illustration of
Although not necessary, in some embodiments, the network 30 may be capable of supporting communication in accordance with any one or more of a number of First-Generation (1G), Second-Generation (2G), 2.5G, Third-Generation (3G), 3.5G, 3.9G, Fourth-Generation (4G) mobile communication protocols, Long Term Evolution (LTE), LTE advanced (LTE-A) and/or the like. Thus, the network 30 may be a cellular network, a mobile network and/or a data network, such as a Local Area Network (LAN), a Metropolitan Area Network (MAN), and/or a Wide Area Network (WAN), e.g., the Internet. In turn, other devices such as processing elements (e.g., personal computers, server computers or the like) may be included in or coupled to the network 30. By directly or indirectly connecting the mobile terminal 10 and the other devices (e.g., service platform 20, or other mobile terminals or devices such as the communication device 15) to the network 30, the mobile terminal 10 and/or the other devices may be enabled to communicate with each other, for example, according to numerous communication protocols, to thereby carry out various communication or other functions of the mobile terminal 10 and the other devices, respectively. As such, the mobile terminal 10 and the other devices may be enabled to communicate with the network 30 and/or each other by any of numerous different access mechanisms. For example, mobile access mechanisms such as Wideband Code Division Multiple Access (W-CDMA), CDMA2000, Global System for Mobile communications (GSM), General Packet Radio Service (GPRS) and/or the like may be supported as well as wireless access mechanisms such as Wireless LAN (WLAN), Worldwide Interoperability for Microwave Access (WiMAX), WiFi, Ultra-Wide Band (UWB), Wibree techniques and/or the like and fixed access mechanisms such as Digital Subscriber Line (DSL), cable modems, Ethernet and/or the like.
In an example embodiment, the service platform 20 may be a device or node such as a server or other processing element. The service platform 20 may have any number of functions or associations with various services. As such, for example, the service platform 20 may be a platform such as a dedicated server (or server bank) associated with a particular information source or service (e.g., a service associated with provision of content elements (e.g., applications, widgets, etc.)), or the service platform 20 may be a backend server associated with one or more other functions or services. As such, the service platform 20 represents a potential host for a plurality of different services or information sources. In one embodiment, the functionality of the service platform 20 is provided by hardware and/or software components configured to operate in accordance with known techniques for the provision of information to users of communication devices. However, at least some of the functionality provided by the service platform 20 may be data processing and/or service provision functionality provided in accordance with an example embodiment of the invention.
In an example embodiment, the mobile terminal 10 may employ an apparatus (e.g., the apparatus 40 of
Referring now to
The apparatus 40 may, in one embodiment, be a mobile terminal (e.g., mobile terminal 10) or a fixed communication device or computing device configured to employ an example embodiment of the invention. However, in one embodiment, the apparatus 40 may be embodied as a chip or chip set. In other words, the apparatus 40 may comprise one or more physical packages (e.g., chips) including materials, components and/or wires on a structural assembly (e.g., a baseboard). The structural assembly may provide physical strength, conservation of size, and/or limitation of electrical interaction for component circuitry included thereon. The apparatus 40 may therefore, in some cases, be configured to implement an embodiment of the invention on a single chip or as a single “system on a chip.” As such, in some cases, a chip or chipset may constitute means for performing one or more operations for providing the functionalities described herein. Additionally or alternatively, the chip or chipset may constitute means for enabling user interface navigation with respect to the functionalities and/or services described herein.
The processor 52 may be embodied in a number of different ways. For example, the processor 52 may be embodied as one or more of various processing means such as a coprocessor, microprocessor, a controller, a digital signal processor (DSP), processing circuitry with or without an accompanying DSP, or various other processing devices including integrated circuits such as, for example, an ASIC (application specific integrated circuit), an FPGA (field programmable gate array), a microcontroller unit (MCU), a hardware accelerator, a special-purpose computer chip, or the like. In an example embodiment, the processor 52 may be configured to execute instructions stored in the memory device 58 or otherwise accessible to the processor 52. As such, whether configured by hardware or software methods, or by a combination thereof, the processor 52 may represent an entity (e.g., physically embodied in circuitry) capable of performing operations according to an embodiment of the invention while configured accordingly. Thus, for example, when the processor 52 is embodied as an ASIC, FPGA or the like, the processor 52 may be specifically configured hardware for conducting the operations described herein. Alternatively, as another example, when the processor 52 is embodied as an executor of software instructions, the instructions may specifically configure the processor 52 to perform the algorithms and operations described herein when the instructions are executed. However, in some cases, the processor 52 may be a processor of a specific device (e.g., a mobile terminal or network device) adapted for employing an embodiment of the invention by further configuration of the processor 52 by instructions for performing the algorithms and operations described herein. The processor 52 may include, among other things, a clock, an arithmetic logic unit (ALU) and logic gates configured to support operation of the processor 52.
In an example embodiment, the processor 52 may be configured to operate a connectivity program, such as a browser, Web browser or the like. In this regard, the connectivity program may enable the apparatus 40 to transmit and receive Web content such as, for example, location-based content or any other suitable content, according to a Wireless Application Protocol (WAP), for example. It should be pointed out that the processor 52 may also be in communication with the touch screen display 50 and may instruct the display to illustrate any suitable information, data, content (e.g., media content) or the like.
Meanwhile, the communication interface 56 may be any means such as a device or circuitry embodied in either hardware, a computer program product, or a combination of hardware and software that is configured to receive and/or transmit data from/to a network and/or any other device or module in communication with the apparatus 40. In this regard, the communication interface 56 may include, for example, an antenna (or multiple antennas) and supporting hardware and/or software for enabling communications with a wireless communication network (e.g., network 30). In fixed environments, the communication interface 56 may alternatively or also support wired communication. As such, the communication interface 56 may include a communication modem and/or other hardware/software for supporting communication via cable, Digital Subscriber Line (DSL), Universal Serial Bus (USB), Ethernet, High-Definition Multimedia Interface (HDMI) or other mechanisms. Furthermore, the communication interface 56 may include hardware and/or software for supporting communication mechanisms such as Bluetooth, Infrared, Ultra-Wideband (UWB), WiFi and/or the like.
The touch screen display 50 may be configured to enable touch recognition by any suitable technique, such as resistive, capacitive, infrared, strain gauge, surface wave, optical imaging, dispersive signal technology, acoustic pulse recognition, or other like techniques. The touch screen display 50 may also detect pointer (e.g., finger) movements just above the touch screen display even in an instance in which the pointer (e.g., finger) may not actually touch the touch screen of the display 50. The touch screen interface 54 may be in communication with the touch screen display 50 to receive indications of user inputs at the touch screen display 50 and to modify a response to such indications based on corresponding user actions that may be inferred or otherwise determined responsive to the indications. In this regard, the touch screen interface 54 may be any device or means embodied in either hardware, software, or a combination of hardware and software configured to perform the respective functions associated with the touch screen interface 54 as described below. In an example embodiment, the touch screen interface 54 may be embodied in software as instructions that are stored in the memory device 58 and executed by the processor 52. Alternatively, the touch screen interface 54 may be embodied as the processor 52 configured to perform the functions of the touch screen interface 54.
The touch screen interface 54 may be configured to receive an indication of an input in the form of a touch event at the touch screen display 50. Following recognition of the touch event, the touch screen interface 54 may be configured to thereafter determine a stroke event or other input gesture and provide a corresponding indication on the touch screen display 50 based on the stroke event. In this regard, for example, the touch screen interface 54 may include a detector 60 to receive indications of user inputs in order to recognize and/or determine a touch event based on each input received at the detector 60.
In an example embodiment, one or more sensors (e.g., sensor 72) may be in communication with the detector 60. The sensors may be any of various devices or modules configured to sense one or more conditions. In this regard, for example, a condition(s) that may be monitored by the sensor 72 may include pressure (e.g., an amount of pressure exerted by a touch event) and any other suitable parameters (e.g., an amount of time in which the touch screen of the display 50 was pressed (e.g., a long press, a swipe), or a size of an area of the touch screen of the display 50 that was pressed).
A touch event may be defined as a detection of an object, such as a stylus, finger, pen, pencil or any other pointing device, coming into contact with a portion of the touch screen display in a manner sufficient to register as a touch (or registering of a detection of an object just above the touch screen display (e.g., hovering of a finger). In this regard, for example, a touch event could be a detection of pressure on the screen of touch screen display 50 above a particular pressure threshold over a given area. In one alternative embodiment, a touch event may be a detection of pressure on the screen of touch screen display 50 above a particular threshold time. Subsequent to each touch event, the touch screen interface 54 (e.g., via the detector 60) may be further configured to recognize and/or determine a corresponding stroke event or input gesture. A stroke event (which may also be referred to as an input gesture) may be defined as a touch event followed immediately by motion of the object initiating the touch event while the object remains in contact with the touch screen display 50. In other words, the stroke event or input gesture may be defined by motion following a touch event thereby forming a continuous, moving touch event defining a moving series of instantaneous touch positions. The stroke event or input gesture may represent a series of unbroken touch events, or in some cases a combination of separate touch events. For purposes of the description above, the term immediately should not necessarily be understood to correspond to a temporal limitation. Rather, the term immediately, while it may generally correspond to relatively short time after the touch event in many instances, instead is indicative of no intervening actions between the touch event and the motion of the object defining the touch positions while such object remains in contact with the touch screen display 50. In this regard, it should be pointed out that no intervening actions cause operation or function of the touch screen. However, in some instances in which a touch event that is held for a threshold period of time triggers a corresponding function, the term immediately may also have a temporal component associated in that the motion of the object causing the touch event must occur before the expiration of the threshold period of time.
In an example embodiment, the detector 60 may be configured to communicate detection information regarding the recognition or detection of a stroke event or input gesture as well as a selection of one or more items of data (e.g., images, text, graphical elements, etc.) to an input analyzer 62. The input analyzer 62 may communicate with a replication module 78. In one embodiment, the input analyzer 62 (along with the detector 60) may be a portion of the touch screen interface 54. In an example embodiment, the touch screen interface 54 may be embodied by a processor, controller of the like. Furthermore, the input analyzer 62 and the detector 60 may each be embodied as any means such as a device or circuitry embodied in hardware, software or a combination of hardware and software that is configured to perform corresponding functions of the input analyzer 62 and the detector 60, respectively.
The input analyzer 62 may be configured to compare an input gesture or stroke event to various profiles of previously received or predefined input gestures and/or stroke events in order to determine whether a particular input gesture or stroke event corresponds to a known or previously received input gesture or stroke event. If a correspondence is determined, the input analyzer may identify the recognized or determined input gesture or stroke event to the visual variant module 78. In one embodiment, the input analyzer 62 is configured to determine stroke or line orientations (e.g., vertical, horizontal, diagonal, etc.) and various other stroke characteristics such as length, curvature, shape, and/or the like. The determined characteristics may be compared to characteristics of other input gestures either of this user or generic in nature, to determine or identify a particular input gesture or stroke event based on similarity to know input gestures.
In an example embodiment, the processor 52 may be embodied as, include or otherwise control the visual variant module 78. The visual variant module 78 may be any means such as a device or circuitry operating in accordance with software or otherwise embodied in hardware or a combination of hardware and software (e.g., processor 52 operating under software control, the processor 52 embodied as an ASIC or FPGA specifically configured to perform the operations described herein, or a combination thereof) thereby configuring the device or structure to perform the corresponding functions of the visual variant module 78, as described below. Thus, in an example in which software is employed, a device or circuitry (e.g., the processor 52 in one example) executing the software forms the structure associated with such means.
The visual variant module 78 may communicate with the detector 60 and the input analyzer 62. The visual variant module 78 may generate a merging area (e.g., merging area 5 of
The items of visible indicia of the merging area generated by the visible variant module 78 may be indexed as shortcuts to allow quicker access to applications, programs, files, or the like without requiring opening of specific folder, menu or the like for accessing the application(s), program(s), file(s), etc. The visual variant module 78 may change items of visual indicia of a merging area with a different/variant item of visual indicia. The item of visual indicia of the merging area and the different/variant item of visual indicia may correspond to and/or point (e.g., a shortcut) to a same application, program or the like, as described more fully below. Additionally, the visual variant module 78 may change or transform items of visual indicia being moved from the merging area to another area/space of a user interface (e.g., touch screen interface 54), as described more fully below.
For purposes of illustration and not of limitation, the visual variant module 78 may for example, enable visible indicia such as widgets to be dragged from an area (e.g., a home screen) of a user interface (e.g., touch screen interface 54) and dropped in a merging area. In response, the visual variant module 78 may change the visible indicia depicting the appearance of the widget to other visible indicia such as, for example, an icon and may place the icon in the merging area. The icon being placed in the merging area by the visual variant module 78 may replace another icon of the merging area. The icon being replaced may be transformed/changed by the visual variant module 78 to other visible indicia such as, for example, a widget and this widget may be placed in the previous location of the widget that was transformed to an icon, as described more fully below.
The items of visible indicia of the merging area may, but need not, be associated with one or more favorite applications of a user of the apparatus 40. In an example embodiment, the visual variant module 78 may generate a merging area as part of the touch screen interface 54. The merging area (e.g., merging area 5 of
Additionally, the visual variant module 78 may generate the merging area to enable the merging area to be accessible and viewable via a previously accessed view (e.g., screen) of the touch screen interface 54 as well as a view of a home screen to enable continuity between different views of the touch screen interface 54. The merging area may be viewable in a same position (e.g., at a bottom position, a middle position, a top position, etc.) of the previously accessed screen as well as the home screen. For instance, the previously accessed view may be a screen of the touch screen interface 54 preceding a home screen of the touch screen interface 54. Moreover, the visual variant module 78 may generate a merging area that accessible via a screen of the touch screen interface 54 that is next or subsequent to a home screen of the touch screen interface 54 to enable the merging area to be viewable in a same position of a home screen as well as the next or subsequent views of the touch screen interface 54.
In addition, the visual variant module 78 may enable a merging area to be moved to different areas of a home screen or any other screen of the touch screen interface 54 which may be displayed by the touch screen display 50. In this regard, by enabling the merging area to be displayed via the touch screen display 50 in a home screen of the touch screen interface 54 as well as a next or subsequent screen(s) of the touch screen interface 54, the visual variant module 78 may maintain continuity between different user interface views. In this manner, the merging area may share a particular functional attribute in multiple views of the touch screen interface based in part on the continuity and may provide flexibility to a user of the apparatus 40 to manage the objects of the touch screen interface 54.
Additionally, the visual variant module 78 may enable a selected item(s) of visible indicia (e.g., icons) in a portion of a screen of the touch screen interface 54 to be moved into the merging area which may replace an item of visible indicia (e.g., an icon) of the merging area with the item of visible indicia being moved into the merging area. The item(s) of visible indicia being replaced in the merging area may be moved by the visual variant module 78 and/or the detector 60 to the location in which the items being moved to the merging area previously occupied in a screen of the touch screen interface. For more information regarding a manner in which one or more items are moved to a merging area(s) and in which the merging area(s) may be moved in different areas of a user interface such as, for example, a touch screen interface, see PCT Patent Application No. PCT/CN2011/083979, entitled: Methods, Apparatuses And Computer Program Products For Merging Areas In Views Of User Interfaces, filed concurrently herewith. The contents of the foregoing patent application are hereby incorporated by reference in its entirety.
Referring now to
In some example embodiments, in an instance in which the visual variant module and/or a processor (e.g., visual variant module 78 and/or the processor 52) detects a selection of an item of visible indicia of the merging area 5, the visual variant module and/or a processor of the apparatus 340 may execute the corresponding application.
In the example embodiment of
Referring now to
In addition, the visual variant module may include the changed visible indicia associated with the TV icon in the merging area 5. Although the visible indicia associated with the TV widget 4 is changed to other visible indicia corresponding to a TV icon, in this example, by the visual variant module of the apparatus 340, the visible variant module may enable the TV icon to be linked or point to the same application (e.g., a TV application) that the TV widget pointed to or was linked with.
Referring now to
Referring now to
In one example embodiment of
As such, in an example embodiment, in an instance in which the visual variant module detects that a pointer moves (e.g., scrolls) the home screen of touch screen interface 354 of
Additionally, in an instance in which a pointer is moved or scrolls to the left or right of the home screen to access previous or subsequent screens/views of the touch screen interface 354, the visual variant module may enable the merging area 5 to maintain continuity with the home screen and the visual variant module may enable the merging area 5 to keep the same position on the newly accessed screen as the position of the merging area 5 on the home screen. In this regard, the merging area 5 and its items of visible indicia (e.g., icons) may be shared between multiple different screens/views of the touch screen interface 354. The merging area 5 maintaining continuity between views of the touch screen interface 354 may include a transformed item of visible indicia (e.g., TV icon 9, etc.) in the manner described above.
Referring now to
Referring now to
In the example embodiment of
Referring now to
In the example embodiment of
In one example embodiment, the merging area 19 may optionally include a content element 27 (e.g., an item of visual indicia (e.g., an ) that may be generated based in part on transforming/changing a content element 23 that was moved into the merging area 19 from a space of the touch screen interface 754. In other words, the visual variant module of the apparatus 740 may change the content element 23 to a transformed content element 27 (e.g., an item of visual indicia (e.g., an icon)) in response to detection of a pointer releasing the content element 23 at a location (e.g., a location of an icon that was previously positioned at the location) of the merging area 19.
Referring now to
Optionally, at operation 810, an apparatus (e.g., apparatus 40) may include means such as the processor 52, the visual variant module 78 and/or the like, for linking the transformed item of visible indicia to the application. Optionally, at operation 815, an apparatus (e.g., apparatus 40) may include means such as the processor 52, the visual variant module 78 and/or the like for, replacing the item of visible indicia (e.g., an icon) of the merging area with the transformed item of visible indicia (e.g., a different icon) at the location.
It should be pointed out that
Accordingly, blocks of the flowchart support combinations of means for performing the specified functions. It will also be understood that one or more blocks of the flowchart, and combinations of blocks in the flowchart, can be implemented by special purpose hardware-based computer systems which perform the specified functions, or combinations of special purpose hardware and computer instructions.
In an example embodiment, an apparatus for performing the method of
Many modifications and other embodiments of the inventions set forth herein will come to mind to one skilled in the art to which these inventions pertain having the benefit of the teachings presented in the foregoing descriptions and the associated drawings. Therefore, it is to be understood that the inventions are not to be limited to the specific embodiments disclosed and that modifications and other embodiments are intended to be included within the scope of the appended claims. Moreover, although the foregoing descriptions and the associated drawings describe example embodiments in the context of certain example combinations of elements and/or functions, it should be appreciated that different combinations of elements and/or functions may be provided by alternative embodiments without departing from the scope of the appended claims. In this regard, for example, different combinations of elements and/or functions than those explicitly described above are also contemplated as may be set forth in some of the appended claims. Although specific terms are employed herein, they are used in a generic and descriptive sense only and not for purposes of limitation.
Filing Document | Filing Date | Country | Kind | 371c Date |
---|---|---|---|---|
PCT/CN2011/084007 | 12/14/2011 | WO | 00 | 6/12/2014 |