Closing applications

Information

  • Patent Grant
  • 10191633
  • Patent Number
    10,191,633
  • Date Filed
    Monday, December 28, 2015
    9 years ago
  • Date Issued
    Tuesday, January 29, 2019
    6 years ago
Abstract
Application closing techniques are described. In one or more implementations, a computing device recognizes an input as involving selection of an application displayed in a display environment by the computing device and subsequent movement of a point of the selection toward an edge of the display environment. Responsive to the recognizing of the input, the selected application is closed by the computing device.
Description

This application claims priority to U.S. patent application Ser. No. 13/335,001 filed on Dec. 22, 2011 and entitled “Closing Applications”, and issued as U.S. Pat. No. 9,223,472 on Dec. 29, 2015, the entirety of which is incorporated herein by reference.


BACKGROUND

Users have access to computing devices that may assume a wide variety of configurations. For example, the first computing devices that were traditionally available to common users were configured as desktop personal computers. Users typically interacted with these first computing devices using a keyboard and later a mouse to view data on a monitor.


Computing devices continued to evolve as users desired to interact with the devices in an increasing variety of circumstances. One such example was mobile phones that evolved from basic telephone functionality to the ability to run applications. Another such example is a slate computer, which may also be referred to as a tablet. Both of these examples are configured to be held by a user during interaction and thus techniques were developed to address this configuration.


Because of the different configurations of the computing devices, however, a variety of different techniques may be employed for interaction. Thus, even typical users may be confronted with a wide range a different techniques, even to utilize matching functionality of the devices, which could lead to user frustration and even cause the users to forgo use of computing devices having the different configurations due to user confusion. Further, these conventional techniques could be intrusive when employed by the different configurations, such as convention techniques that were used to close an application.


SUMMARY

Application closing techniques are described. In one or more implementations, a computing device recognizes an input as involving selection of an application displayed in a display environment by the computing device and subsequent movement of a point of the selection toward an edge of the display environment. Responsive to the recognizing of the input, the selected application is closed by the computing device.


In one or more implementations, a computing device recognizes a selection input as initiating selection of an application displayed in a display environment of the computing device, the selection made at a portion of the application displayed proximal to a first edge of the display environment. Responsive to the recognition of the selection input, a display characteristic of the application is changed to indicate the selection of the application. The display of the application having the changed display characteristic is also caused to follow subsequent movement of the selection as indicated by one or more inputs. A closing input is recognized through the subsequent movement as crossing a threshold defined in relation to a second edge of the display environment. Responsive to the recognizing of the closing input, a further change in one or more display characteristics of the application is caused to indicate availability of an operation to initiate closing of the application.


In one or more implementations, a computing device enters a mode to close an application responsive to recognition of one or more inputs to select an application displayed in a display environment by the computing device, the mode operable to change a display characteristic of the application and cause the display of the application to follow subsequent movement described by an input device used to select the application. The computing device exits the mode to close the application responsive to detection that the one or more inputs have ceased.


This Summary is provided to introduce a selection of concepts in a simplified form that are further described below in the Detailed Description. This Summary is not intended to identify key features or essential features of the claimed subject matter, nor is it intended to be used as an aid in determining the scope of the claimed subject matter.





BRIEF DESCRIPTION OF THE DRAWINGS

The detailed description is described with reference to the accompanying figures. In the figures, the left-most digit(s) of a reference number identifies the figure in which the reference number first appears. The use of the same reference numbers in different instances in the description and the figures may indicate similar or identical items.



FIG. 1 is an illustration of an environment in an example implementation that is operable to implement application closing techniques described herein.



FIG. 2 illustrates a system in an example implementation of computing device of FIG. 1 as supporting selection of an application in a display environment.



FIG. 3 depicts a system in an example implementation in which a computing device indicates availability of an operation to close the application of FIG. 2.



FIG. 4 depicts a system in an example implementation in which a user has initiated the closing of an application.



FIG. 5 is an illustration of a system in an example implementation showing areas of a display environment defined by thresholds to perform different actions after selection of an application as shown in FIG. 2.



FIG. 6 depicts a system in an example implementation in which an application closing technique is applied to a docked application.



FIG. 7 depicts a system in an example implementation in which close techniques are used to close an application that is available via a back stack maintained by the operating system.



FIG. 8 depicts a system in an example implementation in which a “close all” operation is supported by an operating system of the computing device.



FIG. 9 is a flow diagram depicting a procedure in an example implementation in which one or more inputs are recognized as indicating initiation of a closing operation.



FIG. 10 is a flow diagram depicting a procedure in an example implementation in which display characteristics are changed to indicate selection and availability of a close operation to close an application.



FIG. 11 is a flow diagram depicting a procedure in an example implementation in which a mode is entered that is usable to close one or more applications.



FIG. 12 illustrates an example system that includes the computing device as described with reference to FIG. 1.



FIG. 13 illustrates various components of an example device that can be implemented as any type of computing device as described with reference to FIGS. 1-8 and 12 to implement embodiments of the techniques described herein.





DETAILED DESCRIPTION

Overview


The diversity of computing devices with which a user interacts even in a typical day is growing rapidly. However, conventional techniques via which a user interacted with the devices were often optimized for these configurations. Thus, different techniques could be employed by different computing device configurations, even to access matching functionality. Additionally, development of new configurations could confound users due to the new ways in which a user could interact with the device, which could even result in a user forgoing interaction with unfamiliar device configurations.


Application closing techniques are described. In one or more implementations, techniques are described in which a user may cause a computing device to close one or more applications in an intuitive manner. This may include recognition of selection of the application by a user, such as by selecting a top edge of the application as displayed in a display environment using a cursor control device, a finger of a user's hand using touchscreen functionality, and so forth. Once selected, a point at which the display of the application was selected (e.g., a cursor or point of contact) may then be moved toward an edge of the display environment, such as a bottom edge implemented by one or more display devices.


The computing device, e.g., through execution of an operating system of the computing device, may then recognize an intent of the user to close the application and initiate this process. In one or more implementations, animations and display characteristics may be utilized to provide user feedback, such as to indicate selection of the application, availability of the close operation when the subsequent movement has reached a defined threshold, and so on. In this way, a user is provided with an intuitive technique to close applications that may be leveraged in a variety of different computing environments, examples of which are described in relation to FIG. 12.


In the following discussion, an example environment is first described that may employ the application closing techniques described herein. Example procedures are then described which may be performed in the example environment as well as other environments. Consequently, performance of the example procedures is not limited to the example environment and the example environment is not limited to performance of the example procedures.


Example Environment



FIG. 1 is an illustration of an environment 100 in an example implementation that is operable to employ application closing techniques described herein. The illustrated environment 100 includes a computing device 102 having a processing system 104 and a computer-readable storage medium that is illustrated as a memory 106 although other confirmations are also contemplated as further described below.


The computing device 102 may be configured in a variety of ways. For example, a computing device may be configured as a computer that is capable of communicating over a network, such as a desktop computer, a mobile station, an entertainment appliance, a set-top box communicatively coupled to a display device, a wireless phone, a game console, and so forth. Thus, the computing device 102 may range from full resource devices with substantial memory and processor resources (e.g., personal computers, game consoles) to a low-resource device with limited memory and/or processing resources (e.g., traditional set-top boxes, hand-held game consoles) as further described in relation to FIG. 12. Additionally, although a single computing device 102 is shown, the computing device 102 may be representative of a plurality of different devices, such as multiple servers utilized by a business to perform operations such as by a web service, a remote control and set-top box combination, an image capture device and a game console configured to capture gestures, and so on.


The computing device 102 is further illustrated as including an operating system 108. The operating system 108 is configured to abstract underlying functionality of the computing device 102 to applications 110 that are executable on the computing device 102. For example, the operating system 108 may abstract the processing system 104, memory 106, network, and/or display device 112 functionality of the computing device 102 such that the applications 110 may be written without knowing “how” this underlying functionality is implemented. The application 110, for instance, may provide data to the operating system 108 to be rendered and displayed by the display device 112 without understanding how this rendering will be performed. The operating system 108 may also represent a variety of other functionality, such as to manage a file system and user interface that is navigable by a user of the computing device 102.


The operating system 108 is also illustrated as including an environment module 114 which is representative of functionality of the computing device 102 to provide an environment via which a user may interact with the applications 110 and other data of the computing device 102, both local to the device as well as remotely via a network. A variety of different environments may be supported by the environment module 114, which may be configured to accept inputs to interact with the operating system 108 and applications 110 of the computing device 102 to access functionality of the computing device 102, such as the abstracted functionality described above.


One such environment is referred to as an immersive environment which may be configured to support interaction with the applications 110 with little to no window frame. Additionally, the immersive environment may support interaction with the applications 110 without requiring the user to manage a corresponding window frame's layout, primacy of the window with respect to other windows (e.g., whether a window is active, in front of behind other windows, an order of the windows, and so on). Although illustrated as part of the operating system 108, the environment module 114 may be implemented in a variety of other ways, such as a stand-alone module, remotely via a network, and so forth.


In one or more implementations, the immersive environment of the operating system 108 is configured such that it is not closeable or capable of being uninstalled apart from the operating system 108. Additionally, the immersive environment may be configured to consume a significant portion of an available display area of the display device 112. A user may interact with the immersive environment in a variety of ways, such as via a cursor control device, using one or more gestures involving a touch input, using speech recognition, capture using one or more depth-sensing cameras, and so on. Thus, the environment module 114 may manage the immersive environment in which content of applications 110 may be presented and the presentation may be performed without requiring a user to manage size, location, primacy, and so on of windows used to display the content of the applications 110.


For example, as shown on a user interface displayed by the display device 112, a user interface is shown that is configured to display data 116, 118 from two applications in a “docked” configuration. In this example, both of the applications that correspond to the data are enabled to actively execute by the computing device 102 while execution is suspended for other of the applications 110 that do not currently display data. A gutter 120 is disposed between the displays of the data 116, 118 that may be moveable to change an amount of display area consumed by applications on the display device 112, respectively.


The immersive environment may support a wide variety of functionality to provide an immersive experience for a user to access the applications 110. In the following discussion, this functionality is discussed in relation to leveraging these techniques to close an application. It should be readily apparent, however, that these techniques may be employed for managing interaction in other environments and/or with the applications 110 themselves without departing from the spirit and scope thereof.


For example, the environment module 114 may also be configured to support a desktop environment. The desktop environment is representative of another configuration of a user interface output by the operating system 108 when in this example is to interact with the applications 110 and other data. For example, the desktop environment may be configured to present applications and corresponding data through windows having frames. These frames may provide controls through which a user may interact with an application as well as controls enabling a user to move and size the window. The desktop environment may also support techniques to navigate through a hierarchical file structure through the use of folders and represent the data and applications through use of icons. In one or more implementations, the desktop environment may also be utilized to access application s 110 that are configured specifically for interaction via the desktop environment and not configured for access via the immersive shell 116, although other implementation are also contemplated.


Thus, it should be readily apparent that the techniques described to close applications herein may be supported in a wide range of environments, such as an immersive environment, a desktop environment, or other environment as further described in relation to FIG. 2.


Further, these techniques are not limited to inputs provided by a user. For example, the operating system 108 may incorporate a process lifetime manager (PLM) that is configured to automatically manage whether applications 110 are actively executed, suspended (e.g., frozen—still open but not consuming processing system 104 resources), or terminated (e.g., closed). As applications 110 transition between these parts of the “application lifecycle,” the application may receive two events—visibility and suspension—that indicate whether the application 110 is “visible” on-screen and/or whether the application 110 is about to become “suspended.” These events help to make the application 110 aware as to what to expect next, which may be used to indicate to the application 110 to save certain types of data. Accordingly, the environment module 114 may be configured to ensure that the applications 110 proceed through the typical “application lifecycle” and therefore receives these same events when closed. In one or more implementations, this is different than functionality employed by a task manager to “kill” (e.g., stop execution) of an application, which does not involve sending events to the application 110. The result of closing an app through the typical “application lifecycle” is that the application 110 is able to save user data and exit gracefully.


Generally, any of the functions described herein can be implemented using software, firmware, hardware (e.g., fixed logic circuitry), or a combination of these implementations. The terms “module,” “functionality,” and “engine” as used herein generally represent software, firmware, hardware, or a combination thereof. In the case of a software implementation, the module, functionality, or engine represents program code that performs specified tasks when executed on a processor (e.g., CPU or CPUs). The program code can be stored in one or more computer readable memory devices. The features of the techniques described below are platform-independent, meaning that the techniques may be implemented on a variety of commercial computing platforms having a variety of processors.


For example, the computing device 102 may also include an entity (e.g., software) that causes hardware of the computing device 102 to perform operations, e.g., processors, functional blocks, and so on. For example, the computing device 102 may include a computer-readable medium that may be configured to maintain instructions that cause the computing device, and more particularly hardware of the computing device 102 to perform operations. Thus, the instructions function to configure the hardware to perform the operations and in this way result in transformation of the hardware to perform functions. The instructions may be provided by the computer-readable medium to the computing device 102 through a variety of different configurations.


One such configuration of a computer-readable medium is signal bearing medium and thus is configured to transmit the instructions (e.g., as a carrier wave) to the hardware of the computing device, such as via a network. The computer-readable medium may also be configured as a computer-readable storage medium and thus is not a signal bearing medium. Examples of a computer-readable storage medium include a random-access memory (RAM), read-only memory (ROM), an optical disc, flash memory, hard disk memory, and other memory devices that may use magnetic, optical, and other techniques to store instructions and other data.



FIG. 2 illustrates a system 200 in an example implementation of the computing device 102 of FIG. 1 as supporting selection of application in a display environment. The system 200 is illustrated as including first and second stages 202, 204. At the first stage 202, the computing device 102 is illustrated as outputting a user interface 206 of a social network application in an immersive environment and displaying the user interface 206 on a display device, although other environments are contemplated as previously described. Further, it should be noted that although a display environment implemented using a single display device is shown, it should be readily apparent that a display environment may leverage a plurality of display devices without departing from the spirit and scope thereof.


A finger of the user's hand 208 is illustrated in the first stage as selecting a portion of a display of the application that is near a top edge 210 of a display environment. The selection may be detected using touchscreen functionality of the display device 112, such as to contact the display device 112. A variety of other examples are also contemplated, such as through positioning of a cursor through use of a cursor control device near the top edge 210 and then performing a “click and hold” in which a button of the cursor control device is “held down.”


As shown in the second stage 204 of FIG. 2, the point at which the application was selected is then moved away from the top edge 210 and toward a bottom edge 212. In response to this movement the display of the application is configured to follow this subsequent movement of the selection of the application. Further, display characteristics of the application may be changed to indicate the selection, such as lessening an amount of display area of the display environment consumed by the display of the application, thereby exposing a background 214 of the display environment, e.g., a background utilized by the operating system 108. Other characteristics are also contemplated, such as changing a transparency of the display, color, and so forth. Thus, these characteristics may be used to give feedback to the user regarding selection of the application. This selection may then be used as a basis to manage how and if the application is displayed, as further described in relation to the following figure.



FIG. 3 depicts a system 300 in an example implementation in which a computing device indicates availability of an operation to close the application of FIG. 2. In this illustration, subsequent movement of the finger of the user's hand 208 has continued toward the bottom edge 212 of the display environment.


Once movement past a threshold has been detected by the environment module 114, the environment module 114 may again cause display characteristics of the display of the application to change to indicate availability of a close operation. As before, this may include changing a size as illustrated, transparency, color, a “speed bump” in which the movement of the selected display of the application is slowed to indicate crossing of the threshold, and so on of the display of the application. Thus, at this point a user is made aware of the availability of the close operation and may then choose to do so as further described in relation to the following figure.



FIG. 4 depicts a system 400 in an example implementation in which a user has initiated the closing of an application. In this example, a user has lifted a finger of the user's hand 208 away from the display device 112. This causes the environment module 114 to exit the selection mode in which the user could reposition the display of the application. Further, because this selection occurred while the display of the application indicated availability of the close operation (due to passing of the threshold), the environment module 114 may initiate the operation to close the application. The may include removing the application from availability for execution, such that if execution was desired the application would first be reloaded by the operating system 108.


The close operation may be accompanied by one or more animations or other display techniques to indicate that the operation has been initiated to provide feedback to the user. This may include further shrinking and movement of the display of the application “off” the display device 112 as indicated by the arrows in FIG. 4. In this example, a user initiated the operation to close the application. The environment module 114 may also include functionality such that a user may cease initiation of this operation, of further described in relation to the following figure.



FIG. 5 is an illustration of a system 500 in an example implementation showing areas of a display environment defined by thresholds to perform different actions after selection of an application as shown in FIG. 2. A close 502 region is shown as defined between a threshold 504 and a bottom edge 212 of the display environment. In this example, the threshold 504 is defined statically, e.g., at a defined distance from the bottom edge 212, such as for a gesture recognized using touch functionality of the display device 112. Other examples are also contemplated, such as a dynamic threshold defined as a percentage of movement between edges (e.g., the top and bottom edges 210, 212), such as twenty five percent for movement input using a cursor control device.


Thus, the close 502 region may be used to define “when” the close operation is available for an application that is “moved within” the region. This may include changing display characteristics to indicate availability as described in relation to FIG. 3.


Use of regions may also be used to allow a user to exit this mode, such as in an instance in which a user reconsiders, inadvertently entered the mode the close the application, and so on. For example, once moving into the close 502 region and upon receiving feedback that the close operation is available through the change in display characteristics, the user may reconsider. Accordingly, the user may then move a point of contact (e.g., a finger of the user's hand 208) back outside of the close 502 region. Accordingly, the environment module 114 may return the display of the application to the selection state as shown in FIG. 2.


Other regions may also be defined for implementation as part of the display environment. For example, a discard 506 region may be defined along another edge of the display environment, e.g., the left edge in the illustrated example. Movement of the display of the application in the selected state to the discard 506 region may cause the environment module 114 to exit the selection mode, e.g., to return to a state in the display environment as was achieved before the selection in FIG. 2.


In another example, snap 508 regions may be defined to “dock” a display of an application along an edge of the display environment. As described and shown in relation to FIG. 1, for instance, an immersive environment may support a concurrent display of data 116, 118 from applications that is separated by a gutter 120 that is configured to adjust an amount of display area of the display device 112 that is consumed by the respective applications. In a further example, a switch region may be defined, such as to move an application from a snap 508 region in a selected state to the switch 510 region. A variety of other examples are also contemplated. The close techniques may also be implemented to support these different views of the applications, such as for a docked application as described in relation to the following figure.



FIG. 6 depicts a system 600 in an example implementation in which an application closing technique is applied to a docked application. This system 600 is illustrated through use of first and second stages 602, 604. At the first stage 602, a user interface displayed by the display device 112 includes a first region and second region separated by a gutter 120. The first region in this example includes a primary display by the display device 112 and the second region in this example includes a “docked” application as described previously.


The first and second regions are illustrated as rendering data from a search service 606 (e.g., via a browser application) and a social network application 608, respectively. A finger of the user's hand 208 is illustrated as selecting the display of the social network application 608 from the docked region, such as by interacting with a top edge 210 of the display environment that coincides with the display of the social network application in the second region.


Responsive to subsequent movement of a point of contact away from the top edge 210 and toward the opposing bottom edge 212 of the display environment, the display characteristics of the application may change as shown in the second stage 604. Further, the display of the application in the first region (e.g., the search application 606) may be adjusted to consume a background “behind” the display of the social application, e.g., as an application in an immersive environment.


As before, the display of the application may be configured to follow subsequent movement of the finger of the user's hand 208. The user may then manipulate this display toward a close region 502 for output of an indication that a close operation is available, to initiate the close operation, and so on as described in relation to FIGS. 3-5. Other examples are also contemplated, at least one of which is described in relation to the following figure.



FIG. 7 depicts a system 700 in an example implementation in which the close techniques are used to close an application that is available via a back stack maintained by the operating system 108. This system 700 is shown through use of first and second stages 702, 704. As previously described, the operating system 108 may maintain an immersive environment in which a display of one or more applications may substantially consume a display environment of one or more display devices. However, applications may also “remain available” by the operating system 108 through use of a back stack in which execution of the application is suspended until it appears on the display device 112. A user, for instance, may make a swipe gesture through one or more fingers of the users hand, a “click and drag” gesture, and so on to replace one or more of the applications currently being displayed with an application from the back stack.


The close techniques described herein may also be used to close applications from the back stack. Further, these techniques may also be used without executing the application, thereby conserving resources of the computing device 102. For example, as shown at the first stage 702, a finger of the user's hand 208 is indicated as being placed proximal to a left edge of a display environment and moved away from the edge through use of an arrow 706.


In response, the environment module 114 may output a display of a “next available” application from the back stack, such as when the back stack is configured for navigation similar to a carousel. The next available application in this example is a display of the social network application 608 and is displayed as in a selection mode as previously described to follow subsequent movement of a point of selection. The selection mode may be used to support a variety of functionality as previously described, such as to specify “where” the display of the application is to be performed in the display environment, such as in the snap 508 or switch 510 regions described in relation to FIG. 5.


This selection may also be used to close the application and therefore remove it as available via the back stack. For example, subsequent movement of a point of selection may be made toward a bottom edge 212 of the display environment (as illustrated by an arrow in this example) to close the application as described in relation to FIGS. 3-5. In this way, a user may “clean up” a back stack to lessen a number of applications that are made available via scrolling. As before, this mode may also be exited through moving back to the discard 506 region and so on as previously described.



FIG. 8 depicts a system 800 in an example implementation in which a “close all” operation is supported by an operating system 108 of the computing device 102. This system 800 is also illustrated through use of first and second stages 802, 804. At the first stage 802, a cursor is shown as selecting a point near an edge of a display environment and including subsequent movement away from that edge, as shown through use of an arrow 806. Responsive to this movement, a menu 808 is output by the environment module 114.


The menu 808 in this example includes representations of commands that may be initiated from both the applications 110 and/or the environments supported by the environment module 114. Examples of such commands are illustrated as including “search,” “share,” “start,” “connect,” and “settings.”


In this example, selection of the “settings” command at the first stage 802 causes output of another command that is selected to close all applications 810 as shown in the second stage 804. Selection of the close all applications 810 command may cause removal of an ability to execute each of the applications 110 by the computing device 102 without re-launching (as opposed to restoring) the applications 110. For example, the application may be removed from both a current display as well as a back stack supported by the operating system 108. A variety of other examples are also contemplated.


Example Procedures


The following discussion describes application closing techniques that may be implemented utilizing the previously described systems and devices. Aspects of each of the procedures may be implemented in hardware, firmware, or software, or a combination thereof. The procedures are shown as a set of blocks that specify operations performed by one or more devices and are not necessarily limited to the orders shown for performing the operations by the respective blocks. In portions of the following discussion, reference will be made to the environment 100 of FIG. 1 and systems 200-800 of FIGS. 2-8, respectively.



FIG. 9 depicts a procedure 900 in an example implementation in which one or more inputs are recognized as indicating initiation of a closing operation. A computing device recognizes an input as involving selection of an application displayed in a display environment by the computing device and subsequent movement of a point of the selection toward an edge of the display environment (block 902). The input, for instance, may be received via a cursor control device, through touchscreen functionality of a display device 112, detected using one or more cameras such as through a natural user interface, and so on. As previously described, although movement toward a bottom edge 212 was discussed, a variety of different edges of a display environment may be utilized. Additionally, the display environment may be supported by one or more display devices and thus is not limited to a physical edge nor a single display device.


Responsive to the recognizing of the input, the selected application is closed by the computing device (block 904). The operating system 108, for instance, may remove the application 110 from volatile memory, e.g., RAM, removed from storage as part of state maintained by the operating system 108 such that the application 110 is not available without “re-launching” the application 110 (e.g., as part of a back stack), and so forth.



FIG. 10 depicts a procedure 1000 in an example implementation in which display characteristics are changed to indicate selection and availability of a close operation to close an application. A computing device recognizes a selection input as initiating selection of an application displayed in a display environment of the computing device, the selection made at a portion of the application displayed proximal to at a first edge of the display environment (block 1002). The selection, for instance, may involve contact detected using touchscreen functionality of the display device 112 (e.g., a finger or stylus press), a click and hold made by a cursor control device, and so forth.


Responsive to the recognition of the selection input, a display characteristic of the application is changed to indicate the selection of the application (block 1004). The environment module 114, for instance, may change a size, transparency, and so on of the display of the application in the user interface.


The display of the application having the changed display characteristic is also caused to follow subsequent movement of the selection as indicated by one or more inputs (block 1006). Continuing with the previous example, the subsequent movement may follow subsequent movement defined by a cursor control device while “clicked,” subsequent movement of a contact as continued across a display device 112 using touchscreen functionality, and so forth.


A closing input is recognized through the subsequent movement as crossing a threshold defined in relation to a second edge of the display environment (block 1008). The closing input, for instance, may be detected by crossing a static threshold (e.g., a threshold defined at a particular location in a display environment for a touch contact), a dynamic threshold (e.g., defined as an amount of movement in relation to a display environment, such as for a cursor control device), and so forth.


Responsive to the recognizing of the closing input, a further change in one or more display characteristics of the application is caused to indicate availability to initiate closing of the application (block 1010). The environment module 114, for instance, may cause the change (e.g., a further reduction in size, increase in transparency, and so forth) to indicate when the close operation is available. A user may then lift a button of a cursor control device, remove a finger of the user's hand 208 from contacted the display device 112, and so on to initiate the operation, further discussion of which may be found in relation to the following figure.



FIG. 11 depicts a procedure 1100 in an example implementation in which a mode is entered that is usable to close one or more applications. A computing device enters a mode to close an application responsive to recognition of one or more inputs to select an application displayed in a display environment by the computing device, the mode operable to change a display characteristic of the application and cause the display of the application to follow subsequent movement described by an input device used to select the application as defined by the inputs (block 1102). A user, for instance, may enter the mode by contacting a display of an application on a display device 112 near a top edge 210 of the device, although other edges and locations in the display environment are also contemplated. As previously described, a cursor control device may also be used.


The computing device exits the mode to close the application responsive to detection that the one or more inputs have ceased (block 1104). The user as shown in FIG. 4, for instance, may remove a finger of the user's hand 208 from the display device 112 while the display of the application 206 is positioned within the close region 502 of the display environment, thereby causing initiation of the close operation and exit of the mode. Other examples are also contemplated, such as through positioned of a point of the selection in the discard 506 region, snap 508 regions, switch 510 regions, and so on.


Example System and Device



FIG. 12 illustrates an example system 1200 that includes the computing device 102 as described with reference to FIG. 1. The example system 1200 enables ubiquitous environments for a seamless user experience when running applications on a personal computer (PC), a television device, and/or a mobile device. Services and applications run substantially similar in all three environments for a common user experience when transitioning from one device to the next while utilizing an application, playing a video game, watching a video, and so on.


In the example system 1200, multiple devices are interconnected through a central computing device. The central computing device may be local to the multiple devices or may be located remotely from the multiple devices. In one embodiment, the central computing device may be a cloud of one or more server computers that are connected to the multiple devices through a network, the Internet, or other data communication link. In one embodiment, this interconnection architecture enables functionality to be delivered across multiple devices to provide a common and seamless experience to a user of the multiple devices. Each of the multiple devices may have different physical requirements and capabilities, and the central computing device uses a platform to enable the delivery of an experience to the device that is both tailored to the device and yet common to all devices. In one embodiment, a class of target devices is created and experiences are tailored to the generic class of devices. A class of devices may be defined by physical features, types of usage, or other common characteristics of the devices.


In various implementations, the computing device 102 may assume a variety of different configurations, such as for computer 1202, mobile 1204, and television 1206 uses. Each of these configurations includes devices that may have generally different constructs and capabilities, and thus the computing device 102 may be configured according to one or more of the different device classes. For instance, the computing device 102 may be implemented as the computer 1202 class of a device that includes a personal computer, desktop computer, a multi-screen computer, laptop computer, netbook, and so on.


The computing device 102 may also be implemented as the mobile 1204 class of device that includes mobile devices, such as a mobile phone, portable music player, portable gaming device, a tablet computer, a multi-screen computer, and so on. The computing device 102 may also be implemented as the television 1206 class of device that includes devices having or connected to generally larger screens in casual viewing environments. These devices include televisions, set-top boxes, gaming consoles, and so on. The techniques described herein may be supported by these various configurations of the computing device 102 and are not limited to the specific examples the techniques described herein.


The cloud 1208 includes and/or is representative of a platform 1210 for content services 1212. The platform 1210 abstracts underlying functionality of hardware (e.g., servers) and software resources of the cloud 1208. The content services 1212 may include applications and/or data that can be utilized while computer processing is executed on servers that are remote from the computing device 102. Content services 1212 can be provided as a service over the Internet and/or through a subscriber network, such as a cellular or Wi-Fi network.


The platform 1210 may abstract resources and functions to connect the computing device 102 with other computing devices. The platform 1210 may also serve to abstract scaling of resources to provide a corresponding level of scale to encountered demand for the content services 1212 that are implemented via the platform 1210. Accordingly, in an interconnected device embodiment, implementation of functionality of the functionality described herein may be distributed throughout the system 1200. For example, the functionality may be implemented in part on the computing device 102 as well as via the platform 1210 that abstracts the functionality of the cloud 1208.



FIG. 13 illustrates various components of an example device 1300 that can be implemented as any type of computing device as described with reference to FIGS. 1, 2, and 12 to implement embodiments of the techniques described herein. Device 1300 includes communication devices 1302 that enable wired and/or wireless communication of device data 1304 (e.g., received data, data that is being received, data scheduled for broadcast, data packets of the data, etc.). The device data 1304 or other device content can include configuration settings of the device, media content stored on the device, and/or information associated with a user of the device. Media content stored on device 1300 can include any type of audio, video, and/or image data. Device 1300 includes one or more data inputs 1306 via which any type of data, media content, and/or inputs can be received, such as user-selectable inputs, messages, music, television media content, recorded video content, and any other type of audio, video, and/or image data received from any content and/or data source.


Device 1300 also includes communication interfaces 1308 that can be implemented as any one or more of a serial and/or parallel interface, a wireless interface, any type of network interface, a modem, and as any other type of communication interface. The communication interfaces 1308 provide a connection and/or communication links between device 1300 and a communication network by which other electronic, computing, and communication devices communicate data with device 1300.


Device 1300 includes one or more processors 1310 (e.g., any of microprocessors, controllers, and the like) which process various computer-executable instructions to control the operation of device 1300 and to implement embodiments of the techniques described herein. Alternatively or in addition, device 1300 can be implemented with any one or combination of hardware, firmware, or fixed logic circuitry that is implemented in connection with processing and control circuits which are generally identified at 1312. Although not shown, device 1300 can include a system bus or data transfer system that couples the various components within the device. A system bus can include any one or combination of different bus structures, such as a memory bus or memory controller, a peripheral bus, a universal serial bus, and/or a processor or local bus that utilizes any of a variety of bus architectures.


Device 1300 also includes computer-readable media 1314, such as one or more memory components, examples of which include random access memory (RAM), non-volatile memory (e.g., any one or more of a read-only memory (ROM), flash memory, EPROM, EEPROM, etc.), and a disk storage device. A disk storage device may be implemented as any type of magnetic or optical storage device, such as a hard disk drive, a recordable and/or rewriteable compact disc (CD), any type of a digital versatile disc (DVD), and the like. Device 1300 can also include a mass storage media device 1316.


Computer-readable media 1314 provides data storage mechanisms to store the device data 1304, as well as various device applications 1318 and any other types of information and/or data related to operational aspects of device 1300. For example, an operating system 1320 can be maintained as a computer application with the computer-readable media 1314 and executed on processors 1310. The device applications 1318 can include a device manager (e.g., a control application, software application, signal processing and control module, code that is native to a particular device, a hardware abstraction layer for a particular device, etc.). The device applications 1318 also include any system components or modules to implement embodiments of the techniques described herein. In this example, the device applications 1318 include an interface application 1322 and an input/output module 1324 that are shown as software modules and/or computer applications. The input/output module 1324 is representative of software that is used to provide an interface with a device configured to capture inputs, such as a touchscreen, track pad, camera, microphone, and so on. Alternatively or in addition, the interface application 1322 and the input/output module 1324 can be implemented as hardware, software, firmware, or any combination thereof. Additionally, the input/output module 1324 may be configured to support multiple input devices, such as separate devices to capture visual and audio inputs, respectively.


Device 1300 also includes an audio and/or video input-output system 1326 that provides audio data to an audio system 1328 and/or provides video data to a display system 1330. The audio system 1328 and/or the display system 1330 can include any devices that process, display, and/or otherwise render audio, video, and image data. Video signals and audio signals can be communicated from device 1300 to an audio device and/or to a display device via an RF (radio frequency) link, S-video link, composite video link, component video link, DVI (digital video interface), analog audio connection, or other similar communication link. In an embodiment, the audio system 1328 and/or the display system 1330 are implemented as external components to device 1300. Alternatively, the audio system 1328 and/or the display system 1330 are implemented as integrated components of example device 1300.


CONCLUSION

Although the invention has been described in language specific to structural features and/or methodological acts, it is to be understood that the invention defined in the appended claims is not necessarily limited to the specific features or acts described. Rather, the specific features and acts are disclosed as example forms of implementing the claimed invention.

Claims
  • 1. A device comprising: one or more processors; andmemory storing instructions, that when executed on the one or more processors, cause the device to: detect a selection of at least a portion of an application displayed on a display screen of the device, the application consuming a first amount of display area within the display screen, wherein the selection corresponds to a touch input from a user of the device;responsive to the detection of the selection, reduce the first amount of display area consumed by the application within the display screen to a second amount of display area, the reducing to indicate the selection of the application;detect a movement of the touch input from an initial position of the selection towards an edge of the display screen;responsive to the detection of the movement of the touch input, cause the second amount of display area to follow the touch input from the initial position of the selection towards the edge of the display screen;determine whether the touch input crossed a threshold defined in relation to the edge of the display screen during the movement of the touch input from the initial position of the selection towards the edge of the display screen, wherein the threshold outlines a close region of the display screen;responsive to the determination that the touch input has crossed the threshold, cause a display characteristic of a current display of the application to change to indicate an availability of a close operation of the application, wherein the close operation removes the application from availability for execution;detect a removal of the touch input from the display screen while the touch input is located in the close region; andresponsive to the detection of the removal of the touch input from the display screen while the touch input is located in the close region, execute the close operation of the application.
  • 2. The device of claim 1, wherein: the first amount of display area comprises a first display transparency;the second amount of display area comprises a second display transparency; andthe instructions further configure the device to change from the first display transparency to the second display transparency responsive to the recognizing of the touch input.
  • 3. The device of claim 1, wherein the touch input comprises a swiping gesture towards the edge of the display screen.
  • 4. The device of claim 1, wherein the instructions further configure the device to recognize that the touch input moves at a velocity that is indicative of a desire to close the application, wherein the closing of the application is further based at least in part on the recognizing that the touch input moves at the velocity that is indicative of the desire to close the application.
  • 5. The device of claim 1, wherein the threshold comprises a dynamic threshold defined as a percentage of a distance between (i) a location where the touch input is initiated and (ii) the edge of the display screen.
  • 6. The device of claim 1, wherein the threshold is defined as a predetermined distance from the edge of the display screen.
  • 7. The device of claim 1, wherein the threshold is defined as a predetermined proportion of an available display area of the display screen from the edge of the display screen.
  • 8. The device of claim 1, wherein the instructions further configure the device to: detect a subsequent movement of the touch input from the close region towards at least a second region, the second region being different from the close region;determine whether the touch input crossed the threshold during the subsequent movement of the touch input from the close region towards the second region; andresponsive to the determination that the touch input crossed the threshold, cause the display characteristic of the current display of the application to cease the indication of the availability of the close operation of the application.
  • 9. A system comprising: a display screen;one or more processors; andmemory storing instructions, that when executed on the one or more processors, perform operations comprising: detecting a selection of at least a portion of an application displayed on the display screen, the application consuming a first amount of display area within the display screen, wherein the selection corresponds to an input from a user of the device;responsive to the detection of the selection, reducing the first amount of display area consumed by the application within the display screen to a second amount of display area;detecting a movement of the input from an initial position of the selection towards an edge of the display screen;responsive to the detection of the movement of the input, causing the second amount of display area to follow the input from the initial position of the selection towards the edge of the display screen;determining whether the input crossed a threshold defined in relation to the edge of the display screen during the movement of the input from the initial position of the selection towards the edge of the display screen, wherein the threshold outlines a close region of the display screen;responsive to the determination that the input has crossed the threshold, causing a display characteristic of a current display of the application to change to indicate an availability of a close operation of the application, wherein the close operation removes the application from availability for execution;detecting a removal of the input from the display screen while the input is located in the close region; andresponsive to the detection of the removal of the input from the display screen while the input is located in the close region, execute the closing operation of the application.
  • 10. The system of claim 9, wherein: the first amount of display area comprises a first display transparency;the second amount of display area comprises a second display transparency; andthe operations further comprise changing from the first display transparency to the second display transparency responsive to the recognizing of the input.
  • 11. The system of claim 9, wherein the input comprises a swiping gesture towards an edge of the display screen.
  • 12. The system of claim 9, wherein the operations further comprise recognizing that the input moves at a velocity that is indicative of a desire to close the application, wherein the closing of the application is further based at least in part on the recognizing that the input moves at the velocity that is indicative of the desire to close the application.
  • 13. The system of claim 9, wherein the threshold comprises a dynamic threshold defined as a percentage of a distance between (i) a location where the input is initiated and (ii) the edge of the display screen.
  • 14. The system of claim 9, wherein the threshold is defined as a predetermined distance from the edge of the display screen.
  • 15. The system of claim 9, wherein the threshold is defined as a predetermined proportion of an available display area of the display screen from the edge of the display screen.
  • 16. The system of claim 9, wherein the operations further comprise: detecting a subsequent movement of the input from the close region towards at least a second region, the second region being different from the close region;determining whether the input crossed the threshold during the subsequent movement of the input from the close region towards the second region; andresponsive to the determination that the input crossed the threshold, causing the display characteristic of the current display of the application to cease the indication of the availability of the close operation of the application.
  • 17. A method comprising: detecting, by one or more processors, a selection of an application displayed on the display screen, the application consuming a first amount of display area within the display screen, wherein the selection corresponds to an input from a user of the device;responsive to the detection of the detection of the selection, reducing the first amount of display area consumed by the application within the display screen to a second amount of display area;detecting a movement of the input from an initial position of the selection towards an edge of the display screen;responsive to the detection of the movement of the input, causing the second amount of display area to follow the input from the initial position of the selection towards the edge of the display screen;determining whether the touch input crossed a threshold defined in relation to the edge of the display screen during the movement of the input from the initial position of the selection towards the edge of the display screen, wherein the threshold outlines a close region of the display screen;responsive to the determination that the input has crossed the threshold, causing a display characteristic of a current display of the application to change to indicate an availability of a close operation of the application, wherein the close operation removes the application from availability for execution;detecting a removal of the input from the display screen while the input is located in the close region; andresponsive to the detection of the removal of the input from the display screen while the input is located in the close region, execute the closing of the application.
  • 18. The method of claim 17, further comprising closing the application after the input has passed a threshold defined in relation to an edge of the display screen.
  • 19. The method of claim 18, wherein the threshold is defined as a predetermined distance from the edge of the display screen.
  • 20. The method of claim 17, wherein the input comprises a swiping gesture towards an edge of the display screen.
US Referenced Citations (830)
Number Name Date Kind
4823283 Diehm et al. Apr 1989 A
5045997 Watanabe Sep 1991 A
5046001 Barker et al. Sep 1991 A
5189732 Kondo Feb 1993 A
5258748 Jones Nov 1993 A
5297032 Trojan et al. Mar 1994 A
5321750 Nadan Jun 1994 A
5339392 Risberg et al. Aug 1994 A
5406307 Hirayama et al. Apr 1995 A
5432932 Chen et al. Jul 1995 A
5463725 Henckel et al. Oct 1995 A
5485197 Hoarty Jan 1996 A
5487143 Southgate Jan 1996 A
5495566 Kwatinetz Feb 1996 A
5515495 Ikemoto May 1996 A
5574836 Broemmelsiek Nov 1996 A
5598523 Fujita Jan 1997 A
5603053 Gough et al. Feb 1997 A
5611060 Belfiore et al. Mar 1997 A
5623613 Rowe et al. Apr 1997 A
5640176 Mundt et al. Jun 1997 A
5650827 Tsumori et al. Jul 1997 A
5657049 Ludolph et al. Aug 1997 A
5675329 Barker et al. Oct 1997 A
5687331 Volk et al. Nov 1997 A
5712995 Cohn Jan 1998 A
5771042 Santos-Gomez Jun 1998 A
5793415 Gregory, III et al. Aug 1998 A
5819284 Farber et al. Oct 1998 A
5844547 Minakuchi et al. Dec 1998 A
5859636 Pandit Jan 1999 A
5860073 Ferrel et al. Jan 1999 A
5889517 Ueda et al. Mar 1999 A
5905492 Straub et al. May 1999 A
5914720 Maples et al. Jun 1999 A
5940076 Sommers et al. Aug 1999 A
5959621 Nawaz et al. Sep 1999 A
5963204 Ikeda et al. Oct 1999 A
6008809 Brooks Dec 1999 A
6008816 Eisler et al. Dec 1999 A
6009519 Jones et al. Dec 1999 A
6011542 Durrani et al. Jan 2000 A
6028600 Rosin et al. Feb 2000 A
6057839 Advani et al. May 2000 A
6064383 Skelly May 2000 A
6104418 Tanaka et al. Aug 2000 A
6108003 Hall, Jr. et al. Aug 2000 A
6111585 Choi Aug 2000 A
6115040 Bladow et al. Sep 2000 A
6166736 Hugh Dec 2000 A
6188405 Czerwinski et al. Feb 2001 B1
6211921 Cherian et al. Apr 2001 B1
6212564 Harter et al. Apr 2001 B1
6216141 Straub et al. Apr 2001 B1
6266098 Cove et al. Jul 2001 B1
6278448 Brown et al. Aug 2001 B1
6281940 Sciammarella Aug 2001 B1
6311058 Wecker et al. Oct 2001 B1
6346935 Nakajima et al. Feb 2002 B1
6369837 Schirmer Apr 2002 B1
6385630 Ejerhed May 2002 B1
6396963 Shaffer et al. May 2002 B2
6411307 Rosin et al. Jun 2002 B1
6424338 Anderson Jul 2002 B1
6426753 Migdal Jul 2002 B1
6433789 Rosman Aug 2002 B1
6448987 Easty et al. Sep 2002 B1
6449638 Wecker et al. Sep 2002 B1
6456334 Duhault Sep 2002 B1
6489977 Sone Dec 2002 B2
6505243 Lortz Jan 2003 B1
6507643 Groner Jan 2003 B1
6510144 Dommety et al. Jan 2003 B1
6510466 Cox et al. Jan 2003 B1
6510553 Hazra Jan 2003 B1
6538635 Ringot Mar 2003 B1
6570582 Sciammarella et al. May 2003 B1
6570597 Seki et al. May 2003 B1
6577323 Jamieson et al. Jun 2003 B1
6577350 Proehl et al. Jun 2003 B1
6591244 Jim et al. Jul 2003 B2
6597374 Baker et al. Jul 2003 B1
6628309 Dodson et al. Sep 2003 B1
6636246 Gallo et al. Oct 2003 B1
6662023 Helle Dec 2003 B1
6690387 Zimmerman et al. Feb 2004 B2
6697825 Underwood et al. Feb 2004 B1
6707449 Hinckley et al. Mar 2004 B2
6710771 Yamaguchi et al. Mar 2004 B1
6721958 Dureau Apr 2004 B1
6724403 Santoro et al. Apr 2004 B1
6784925 Tomat et al. Aug 2004 B1
6798421 Baldwin Sep 2004 B2
6801203 Hussain Oct 2004 B1
6807558 Hassett et al. Oct 2004 B1
6832355 Duperrouzel et al. Dec 2004 B1
6857104 Cahn Feb 2005 B1
6865297 Loui et al. Mar 2005 B2
6873329 Cohen et al. Mar 2005 B2
6876312 Yu Apr 2005 B2
6885974 Holle Apr 2005 B2
6904597 Jin Jun 2005 B2
6920445 Bae Jul 2005 B2
6938101 Hayes et al. Aug 2005 B2
6961731 Holbrook Nov 2005 B2
6971067 Karson et al. Nov 2005 B1
6972776 Davis et al. Dec 2005 B2
6975306 Hinckley et al. Dec 2005 B2
6976210 Silva et al. Dec 2005 B1
6978303 McCreesh et al. Dec 2005 B1
6983310 Rouse et al. Jan 2006 B2
6987991 Nelson Jan 2006 B2
7013041 Miyamoto Mar 2006 B2
7017119 Johnston et al. Mar 2006 B1
7019757 Brown et al. Mar 2006 B2
7028264 Santoro et al. Apr 2006 B2
7032187 Keely, Jr. et al. Apr 2006 B2
7036090 Nguyen Apr 2006 B1
7036091 Nguyen Apr 2006 B1
7042460 Hussain et al. May 2006 B2
7051291 Sciammarella et al. May 2006 B2
7058955 Porkka Jun 2006 B2
7065385 Jarrad et al. Jun 2006 B2
7065386 Smethers Jun 2006 B1
7075535 Aguera y Arcas Jul 2006 B2
7089507 Lection et al. Aug 2006 B2
7091998 Miller-Smith Aug 2006 B2
7093201 Duarte Aug 2006 B2
7106349 Baar et al. Sep 2006 B2
7111044 Lee Sep 2006 B2
7133707 Rak et al. Nov 2006 B1
7133859 Wong Nov 2006 B1
7139800 Bellotti et al. Nov 2006 B2
7146573 Brown et al. Dec 2006 B2
7155729 Andrew et al. Dec 2006 B1
7158123 Myers et al. Jan 2007 B2
7158135 Santodomingo et al. Jan 2007 B2
7178111 Glein et al. Feb 2007 B2
7197702 Niyogi et al. Mar 2007 B2
7210099 Rohrabaugh et al. Apr 2007 B2
7216588 Suess May 2007 B2
7249326 Stoakley et al. Jul 2007 B2
7251782 Albers et al. Jul 2007 B1
7262775 Calkins et al. Aug 2007 B2
7263668 Lentz Aug 2007 B1
7280097 Chen et al. Oct 2007 B2
7283620 Adamczyk Oct 2007 B2
7289806 Morris et al. Oct 2007 B2
7296184 Derks et al. Nov 2007 B2
7296242 Agata et al. Nov 2007 B2
7304638 Murphy Dec 2007 B2
7310100 Hussain Dec 2007 B2
7333092 Zadesky et al. Feb 2008 B2
7333120 Venolia Feb 2008 B2
7336263 Valikangas Feb 2008 B2
7347567 Deter et al. Mar 2008 B2
7369647 Gao et al. May 2008 B2
7376907 Santoro et al. May 2008 B2
7386807 Cummins et al. Jun 2008 B2
7388578 Tao Jun 2008 B2
7403191 Sinclair Jul 2008 B2
7408538 Hinckley et al. Aug 2008 B2
7412663 Lindsay et al. Aug 2008 B2
7433920 Blagsvedt et al. Oct 2008 B2
7447520 Scott Nov 2008 B2
7461151 Colson et al. Dec 2008 B2
7469380 Wessling et al. Dec 2008 B2
7469381 Ording Dec 2008 B2
7478326 Holecek et al. Jan 2009 B2
7479949 Jobs et al. Jan 2009 B2
7480870 Anzures et al. Jan 2009 B2
7483418 Maurer Jan 2009 B2
7484182 Smith Jan 2009 B1
7487467 Kawahara et al. Feb 2009 B1
7496830 Rubin et al. Feb 2009 B2
7512966 Lyons, Jr. et al. Mar 2009 B2
7536650 Robertson May 2009 B1
7577918 Lindsay Aug 2009 B2
7581034 Polivy et al. Aug 2009 B2
7593995 He et al. Sep 2009 B1
7599790 Rasmussen et al. Oct 2009 B2
7600189 Fujisawa Oct 2009 B2
7600234 Dobrowski et al. Oct 2009 B2
7606714 Williams et al. Oct 2009 B2
7607096 Oreizy et al. Oct 2009 B2
7607106 Ernst et al. Oct 2009 B2
7610563 Nelson et al. Oct 2009 B2
7614018 Ohazama et al. Nov 2009 B1
7619615 Donoghue Nov 2009 B1
7640518 Forlines et al. Dec 2009 B2
7653883 Hotelling et al. Jan 2010 B2
7657849 Chaudhri et al. Feb 2010 B2
7663607 Hotelling et al. Feb 2010 B2
7664067 Pointer Feb 2010 B2
7671756 Herz et al. Mar 2010 B2
7676907 Nies Mar 2010 B2
7681138 Grasser et al. Mar 2010 B2
7702683 Kirshenbaum Apr 2010 B1
7730425 de los Reyes et al. Jun 2010 B2
7746388 Jeon Jun 2010 B2
7755674 Kaminaga Jul 2010 B2
7765490 Lai et al. Jul 2010 B2
7834861 Lee Nov 2010 B2
7840979 Poling, Jr. et al. Nov 2010 B2
7877707 Westerman et al. Jan 2011 B2
7880728 de los Reyes et al. Feb 2011 B2
7889180 Byun et al. Feb 2011 B2
7895309 Belali et al. Feb 2011 B2
7898529 Fitzmaurice et al. Mar 2011 B2
7924271 Christie et al. Apr 2011 B2
7933632 Flynt et al. Apr 2011 B2
7962281 Rasmussen et al. Jun 2011 B2
7983718 Roka Jul 2011 B1
7987431 Santoro et al. Jul 2011 B2
8006276 Nakagawa et al. Aug 2011 B2
8065628 Oshiro et al. Nov 2011 B2
8086275 Wykes et al. Dec 2011 B2
8108781 Laansoo et al. Jan 2012 B2
8131808 Aoki et al. Mar 2012 B2
8134727 Shmunis et al. Mar 2012 B1
8150924 Buchheit et al. Apr 2012 B2
8175653 Smuga et al. May 2012 B2
8176438 Zaman et al. May 2012 B2
8190707 Trivedi et al. May 2012 B2
8209623 Barletta et al. Jun 2012 B2
8225193 Kleinschnitz et al. Jul 2012 B1
8238876 Teng et al. Aug 2012 B2
8245152 Brunner et al. Aug 2012 B2
8245156 Mouilleseaux et al. Aug 2012 B2
8250494 Butcher et al. Aug 2012 B2
8255473 Eren et al. Aug 2012 B2
8255812 Parparita et al. Aug 2012 B1
8255818 Bales et al. Aug 2012 B2
8269736 Wilairat Sep 2012 B2
8280901 McDonald Oct 2012 B2
8289688 Behar et al. Oct 2012 B2
8312383 Gilfix Nov 2012 B2
8334871 Hamilton, II et al. Dec 2012 B2
8355698 Teng et al. Jan 2013 B2
8385952 Friedman et al. Feb 2013 B2
8411046 Kruzeniski et al. Apr 2013 B2
8429565 Agarawala et al. Apr 2013 B2
8448083 Migos et al. May 2013 B1
8473870 Hinckley et al. Jun 2013 B2
8527892 Sirpal et al. Sep 2013 B2
8539384 Hinckley et al. Sep 2013 B2
8548431 Teng et al. Oct 2013 B2
8549430 Russell et al. Oct 2013 B2
8560959 Zaman et al. Oct 2013 B2
8589815 Fong et al. Nov 2013 B2
8612874 Zaman et al. Dec 2013 B2
8615713 Sun et al. Dec 2013 B2
8639819 Pohja et al. Jan 2014 B2
8687023 Markiewicz et al. Apr 2014 B2
8689123 Zaman et al. Apr 2014 B2
9015606 Zaman et al. Apr 2015 B2
20010022621 Squibbs Sep 2001 A1
20020000963 Yoshida et al. Jan 2002 A1
20020018051 Singh Feb 2002 A1
20020035607 Checkoway et al. Mar 2002 A1
20020054117 van Dantzich et al. May 2002 A1
20020060701 Naughton et al. May 2002 A1
20020070961 Xu et al. Jun 2002 A1
20020077156 Smethers Jun 2002 A1
20020091755 Narin Jul 2002 A1
20020097264 Dutta et al. Jul 2002 A1
20020105531 Niemi Aug 2002 A1
20020105553 Segre Aug 2002 A1
20020115476 Padawer et al. Aug 2002 A1
20020128036 Yach et al. Sep 2002 A1
20020129061 Swart et al. Sep 2002 A1
20020138248 Corston-Oliver et al. Sep 2002 A1
20020142762 Chmaytelli et al. Oct 2002 A1
20020145631 Arbab et al. Oct 2002 A1
20020152305 Jackson et al. Oct 2002 A1
20020154176 Barksdale et al. Oct 2002 A1
20020161634 Kaars Oct 2002 A1
20020186251 Himmel et al. Dec 2002 A1
20020194385 Linder et al. Dec 2002 A1
20030003899 Tashiro et al. Jan 2003 A1
20030008686 Park et al. Jan 2003 A1
20030011643 Nishihata Jan 2003 A1
20030020671 Santoro et al. Jan 2003 A1
20030040300 Bodic et al. Feb 2003 A1
20030046396 Richter et al. Mar 2003 A1
20030073414 P. Capps Apr 2003 A1
20030096604 Vollandt May 2003 A1
20030105827 Tan et al. Jun 2003 A1
20030135582 Allen et al. Jul 2003 A1
20030187996 Cardina et al. Oct 2003 A1
20030222907 Heikes et al. Dec 2003 A1
20030225846 Heikes et al. Dec 2003 A1
20030234799 Lee Dec 2003 A1
20040015553 Griffin et al. Jan 2004 A1
20040066414 Czerwinski et al. Apr 2004 A1
20040066418 Tosey Apr 2004 A1
20040068543 Seifert Apr 2004 A1
20040078299 Down-Logan et al. Apr 2004 A1
20040111673 Bowman et al. Jun 2004 A1
20040155908 Wagner Aug 2004 A1
20040165010 Robertson et al. Aug 2004 A1
20040185883 Rukman Sep 2004 A1
20040212586 Denny, III Oct 2004 A1
20040212617 Fitzmaurice et al. Oct 2004 A1
20040217954 O'Gorman et al. Nov 2004 A1
20040237048 Tojo et al. Nov 2004 A1
20040250217 Tojo et al. Dec 2004 A1
20050005241 Hunleth et al. Jan 2005 A1
20050028208 Ellis et al. Feb 2005 A1
20050044058 Matthews et al. Feb 2005 A1
20050054384 Pasquale et al. Mar 2005 A1
20050060647 Doan et al. Mar 2005 A1
20050060658 Tsukiori Mar 2005 A1
20050060665 Rekimoto Mar 2005 A1
20050079896 Kokko et al. Apr 2005 A1
20050085215 Kokko et al. Apr 2005 A1
20050085272 Anderson et al. Apr 2005 A1
20050108655 Andrea et al. May 2005 A1
20050114788 Fabritius May 2005 A1
20050120306 Klassen et al. Jun 2005 A1
20050125739 Thompson et al. Jun 2005 A1
20050136953 Jo et al. Jun 2005 A1
20050143138 Lee et al. Jun 2005 A1
20050182798 Todd et al. Aug 2005 A1
20050183021 Allen et al. Aug 2005 A1
20050184999 Daioku Aug 2005 A1
20050198159 Kirsch Sep 2005 A1
20050198584 Matthews et al. Sep 2005 A1
20050200762 Barletta et al. Sep 2005 A1
20050216300 Appelman et al. Sep 2005 A1
20050223057 McNulty Oct 2005 A1
20050223069 Cooperman et al. Oct 2005 A1
20050232166 Nierhaus Oct 2005 A1
20050250547 Salman et al. Nov 2005 A1
20050268237 Crane et al. Dec 2005 A1
20050273614 Ahuja et al. Dec 2005 A1
20050280719 Kim et al. Dec 2005 A1
20060004685 Pyhalammi et al. Jan 2006 A1
20060005207 Louch et al. Jan 2006 A1
20060010394 Chaudhri et al. Jan 2006 A1
20060015562 Kilian-Kehr et al. Jan 2006 A1
20060015736 Callas et al. Jan 2006 A1
20060015812 Cunningham et al. Jan 2006 A1
20060015820 Wood Jan 2006 A1
20060026013 Kraft Feb 2006 A1
20060026521 Hotelling et al. Feb 2006 A1
20060036425 Le Cocq et al. Feb 2006 A1
20060048073 Jarrett et al. Mar 2006 A1
20060048101 Krassovsky et al. Mar 2006 A1
20060059430 Bells et al. Mar 2006 A1
20060070005 Gilbert et al. Mar 2006 A1
20060074735 Shukla et al. Apr 2006 A1
20060074771 Kim et al. Apr 2006 A1
20060075360 Bixler Apr 2006 A1
20060103623 Davis May 2006 A1
20060107231 Matthews et al. May 2006 A1
20060112354 Park et al. May 2006 A1
20060114239 Nakajima Jun 2006 A1
20060129543 Bates et al. Jun 2006 A1
20060135220 Kim et al. Jun 2006 A1
20060136773 Kespohl et al. Jun 2006 A1
20060152803 Provitola Jul 2006 A1
20060156247 McCormack et al. Jul 2006 A1
20060172724 Linkert et al. Aug 2006 A1
20060173911 Levin et al. Aug 2006 A1
20060184901 Dietz Aug 2006 A1
20060190833 SanGiovanni et al. Aug 2006 A1
20060199598 Lee et al. Sep 2006 A1
20060212806 Griffin et al. Sep 2006 A1
20060218234 Deng et al. Sep 2006 A1
20060218501 Wilson et al. Sep 2006 A1
20060224993 Wong et al. Oct 2006 A1
20060246955 Nirhamo et al. Nov 2006 A1
20060248471 Lindsay et al. Nov 2006 A1
20060253801 Okaro et al. Nov 2006 A1
20060259870 Hewitt et al. Nov 2006 A1
20060259873 Mister Nov 2006 A1
20060262134 Hamiter et al. Nov 2006 A1
20060268100 Karukka et al. Nov 2006 A1
20060271520 Ragan Nov 2006 A1
20060281448 Plestid et al. Dec 2006 A1
20060293088 Kokubo Dec 2006 A1
20060294063 Ali et al. Dec 2006 A1
20060294396 Witman et al. Dec 2006 A1
20070005716 LeVasseur et al. Jan 2007 A1
20070006094 Canfield et al. Jan 2007 A1
20070011610 Sethi et al. Jan 2007 A1
20070015532 Deelman Jan 2007 A1
20070024646 Saarinen et al. Feb 2007 A1
20070035513 Sherrard et al. Feb 2007 A1
20070038567 Allaire et al. Feb 2007 A1
20070050724 Lee et al. Mar 2007 A1
20070054679 Cho et al. Mar 2007 A1
20070061306 Pell et al. Mar 2007 A1
20070061488 Alagappan et al. Mar 2007 A1
20070061714 Stuple et al. Mar 2007 A1
20070063995 Bailey et al. Mar 2007 A1
20070067272 Flynt et al. Mar 2007 A1
20070067737 Zielinski et al. Mar 2007 A1
20070067798 Wroblewski Mar 2007 A1
20070073718 Ramer et al. Mar 2007 A1
20070076013 Campbell et al. Apr 2007 A1
20070080954 Griffin et al. Apr 2007 A1
20070082707 Flynt et al. Apr 2007 A1
20070082708 Griffin Apr 2007 A1
20070083746 Fallon et al. Apr 2007 A1
20070083821 Garbow et al. Apr 2007 A1
20070094597 Rostom Apr 2007 A1
20070106635 Frieden et al. May 2007 A1
20070120835 Sato May 2007 A1
20070127638 Doulton Jun 2007 A1
20070136677 Agarwal Jun 2007 A1
20070150826 Anzures et al. Jun 2007 A1
20070150842 Chaudhri et al. Jun 2007 A1
20070157089 Van Os et al. Jul 2007 A1
20070157099 Haug Jul 2007 A1
20070171192 Seo et al. Jul 2007 A1
20070171238 Ubillos et al. Jul 2007 A1
20070180401 Singh et al. Aug 2007 A1
20070182595 Ghasabian Aug 2007 A1
20070182999 Anthony et al. Aug 2007 A1
20070185847 Budzik et al. Aug 2007 A1
20070192707 Maeda et al. Aug 2007 A1
20070192730 Simila et al. Aug 2007 A1
20070192733 Horiuchi Aug 2007 A1
20070192739 Hunleth et al. Aug 2007 A1
20070197196 Shenfield et al. Aug 2007 A1
20070198420 Goldstein Aug 2007 A1
20070208840 McConville et al. Sep 2007 A1
20070211034 Griffin et al. Sep 2007 A1
20070214422 Agarwal et al. Sep 2007 A1
20070214429 Lyudovyk et al. Sep 2007 A1
20070214454 Edwards et al. Sep 2007 A1
20070216651 Patel Sep 2007 A1
20070216661 Chen et al. Sep 2007 A1
20070220444 Sunday Sep 2007 A1
20070225022 Satake Sep 2007 A1
20070233654 Karlson et al. Oct 2007 A1
20070236468 Tuli Oct 2007 A1
20070238487 Kuhl et al. Oct 2007 A1
20070238488 Scott Oct 2007 A1
20070247435 Benko et al. Oct 2007 A1
20070250583 Hardy et al. Oct 2007 A1
20070250787 Kawahara et al. Oct 2007 A1
20070253758 Suess Nov 2007 A1
20070255831 Hayashi et al. Nov 2007 A1
20070256029 Maxwell Nov 2007 A1
20070257891 Esenther et al. Nov 2007 A1
20070257933 Klassen et al. Nov 2007 A1
20070260674 Shenfield Nov 2007 A1
20070262964 Zotov et al. Nov 2007 A1
20070263843 Foxenland Nov 2007 A1
20070273663 Park et al. Nov 2007 A1
20070273668 Park et al. Nov 2007 A1
20070280457 Aberethy et al. Dec 2007 A1
20070281747 Pletikosa et al. Dec 2007 A1
20070291007 Forlines et al. Dec 2007 A1
20080005668 Mavinkurve et al. Jan 2008 A1
20080028294 Sell et al. Jan 2008 A1
20080032681 West Feb 2008 A1
20080036743 Westerman et al. Feb 2008 A1
20080040692 Sunday et al. Feb 2008 A1
20080048986 Khoo Feb 2008 A1
20080052370 Snyder Feb 2008 A1
20080057910 Thoresson et al. Mar 2008 A1
20080057926 Forstall et al. Mar 2008 A1
20080059896 Anderson et al. Mar 2008 A1
20080066010 Brodersen et al. Mar 2008 A1
20080072173 Brunner et al. Mar 2008 A1
20080076472 Hyatt Mar 2008 A1
20080082934 Kocienda et al. Apr 2008 A1
20080084970 Harper Apr 2008 A1
20080085700 Arora Apr 2008 A1
20080092054 Bhumkar et al. Apr 2008 A1
20080092057 Monson et al. Apr 2008 A1
20080095100 Cleveland et al. Apr 2008 A1
20080102863 Hardy et al. May 2008 A1
20080104544 Collins et al. May 2008 A1
20080107057 Kannan et al. May 2008 A1
20080113656 Lee et al. May 2008 A1
20080114535 Nesbitt May 2008 A1
20080120569 Mann et al. May 2008 A1
20080120571 Chang et al. May 2008 A1
20080122796 Jobs et al. May 2008 A1
20080132252 Altman et al. Jun 2008 A1
20080141153 Samson et al. Jun 2008 A1
20080153551 Baek et al. Jun 2008 A1
20080155425 Murthy et al. Jun 2008 A1
20080162651 Madnani Jul 2008 A1
20080163104 Haug Jul 2008 A1
20080165132 Weiss et al. Jul 2008 A1
20080165136 Christie et al. Jul 2008 A1
20080165141 Christie Jul 2008 A1
20080165163 Bathiche Jul 2008 A1
20080167058 Lee et al. Jul 2008 A1
20080168349 Lamiraux et al. Jul 2008 A1
20080168379 Forstall et al. Jul 2008 A1
20080168382 Louch et al. Jul 2008 A1
20080168402 Blumenberg Jul 2008 A1
20080168403 Westerman et al. Jul 2008 A1
20080172609 Rytivaara Jul 2008 A1
20080174570 Jobs et al. Jul 2008 A1
20080180399 Cheng Jul 2008 A1
20080182628 Lee et al. Jul 2008 A1
20080184112 Chiang et al. Jul 2008 A1
20080189653 Taylor et al. Aug 2008 A1
20080189658 Jeong et al. Aug 2008 A1
20080192056 Robertson et al. Aug 2008 A1
20080198141 Lee et al. Aug 2008 A1
20080200142 Abdel-Kader et al. Aug 2008 A1
20080208973 Hayashi et al. Aug 2008 A1
20080222273 Lakshmanan et al. Sep 2008 A1
20080222545 Lemay et al. Sep 2008 A1
20080222547 Wong et al. Sep 2008 A1
20080222560 Harrison Sep 2008 A1
20080222569 Champion et al. Sep 2008 A1
20080242362 Duarte Oct 2008 A1
20080250354 Park Oct 2008 A1
20080259042 Thorn Oct 2008 A1
20080261513 Shin et al. Oct 2008 A1
20080261660 Huh et al. Oct 2008 A1
20080263457 Kim et al. Oct 2008 A1
20080270558 Ma et al. Oct 2008 A1
20080297475 Woolf et al. Dec 2008 A1
20080299999 Lockhart et al. Dec 2008 A1
20080301046 Martinez et al. Dec 2008 A1
20080301575 Fermon Dec 2008 A1
20080307351 Louch et al. Dec 2008 A1
20080307364 Chaudhri et al. Dec 2008 A1
20080309626 Westerman et al. Dec 2008 A1
20080313538 Hudson Dec 2008 A1
20080316177 Tseng Dec 2008 A1
20080317240 Chang et al. Dec 2008 A1
20080320413 Oshiro et al. Dec 2008 A1
20090007009 Luneau et al. Jan 2009 A1
20090007017 Anzures et al. Jan 2009 A1
20090012952 Fredriksson Jan 2009 A1
20090029736 Kim et al. Jan 2009 A1
20090031247 Walter et al. Jan 2009 A1
20090037469 Kirsch Feb 2009 A1
20090037846 Spalink et al. Feb 2009 A1
20090051671 Konstas Feb 2009 A1
20090058821 Chaudhri Mar 2009 A1
20090061837 Chaudhri et al. Mar 2009 A1
20090061948 Lee et al. Mar 2009 A1
20090064055 Chaudhri et al. Mar 2009 A1
20090070673 Barkan et al. Mar 2009 A1
20090077649 Lockhart et al. Mar 2009 A1
20090083656 Dukhon et al. Mar 2009 A1
20090085851 Lim et al. Apr 2009 A1
20090085878 Heubel et al. Apr 2009 A1
20090089215 Newton Apr 2009 A1
20090089459 Jeyaseelan et al. Apr 2009 A1
20090094562 Jeong et al. Apr 2009 A1
20090103515 Pointer Apr 2009 A1
20090106694 Kraft et al. Apr 2009 A1
20090106696 Duarte Apr 2009 A1
20090109243 Kraft et al. Apr 2009 A1
20090117942 Boningue et al. May 2009 A1
20090125844 Weir et al. May 2009 A1
20090140061 Schultz et al. Jun 2009 A1
20090140986 Karkkainen et al. Jun 2009 A1
20090144642 Crystal Jun 2009 A1
20090144653 Ubillos Jun 2009 A1
20090146962 Ahonen et al. Jun 2009 A1
20090153492 Popp Jun 2009 A1
20090158144 Griffin Jun 2009 A1
20090160809 Yang et al. Jun 2009 A1
20090163182 Gatti et al. Jun 2009 A1
20090164888 Phan Jun 2009 A1
20090164928 Brown et al. Jun 2009 A1
20090164936 Kawaguchi Jun 2009 A1
20090174679 Westerman Jul 2009 A1
20090178007 Matas et al. Jul 2009 A1
20090182788 Chung et al. Jul 2009 A1
20090184939 Wohlstadter et al. Jul 2009 A1
20090199122 Deutsch et al. Aug 2009 A1
20090199128 Matthews et al. Aug 2009 A1
20090199130 Tsern et al. Aug 2009 A1
20090205041 Michalske Aug 2009 A1
20090215504 Lando Aug 2009 A1
20090228825 Van Os et al. Sep 2009 A1
20090228841 Hildreth Sep 2009 A1
20090235200 Deutsch et al. Sep 2009 A1
20090235203 Iizuka Sep 2009 A1
20090249257 Bove et al. Oct 2009 A1
20090265662 Bamford Oct 2009 A1
20090271778 Mandyam et al. Oct 2009 A1
20090284482 Chin Nov 2009 A1
20090288044 Matthews et al. Nov 2009 A1
20090292989 Matthews et al. Nov 2009 A1
20090293007 Duarte et al. Nov 2009 A1
20090293013 O'Shaugnessy et al. Nov 2009 A1
20090293014 Meuninck et al. Nov 2009 A1
20090298547 Kim et al. Dec 2009 A1
20090303231 Robinet et al. Dec 2009 A1
20090305732 Marcellino et al. Dec 2009 A1
20090307105 Lemay et al. Dec 2009 A1
20090307589 Inose et al. Dec 2009 A1
20090307623 Agarawala et al. Dec 2009 A1
20090313584 Kerr et al. Dec 2009 A1
20090315839 Wilson et al. Dec 2009 A1
20090315847 Fujii Dec 2009 A1
20090318171 Backholm et al. Dec 2009 A1
20090322760 Kwiatkowski et al. Dec 2009 A1
20090327969 Estrada Dec 2009 A1
20100008490 Gharachorloo et al. Jan 2010 A1
20100013782 Liu et al. Jan 2010 A1
20100020025 Lemort et al. Jan 2010 A1
20100020091 Rasmussen et al. Jan 2010 A1
20100031186 Tseng et al. Feb 2010 A1
20100042911 Wormald et al. Feb 2010 A1
20100050076 Roth Feb 2010 A1
20100058248 Park Mar 2010 A1
20100066698 Seo Mar 2010 A1
20100066705 Keely et al. Mar 2010 A1
20100070931 Nichols Mar 2010 A1
20100073380 Kaplan et al. Mar 2010 A1
20100075628 Ye Mar 2010 A1
20100077058 Messer Mar 2010 A1
20100077310 Karachale et al. Mar 2010 A1
20100077330 Kaplan et al. Mar 2010 A1
20100079413 Kawashima et al. Apr 2010 A1
20100081475 Chiang et al. Apr 2010 A1
20100087169 Lin et al. Apr 2010 A1
20100087173 Lin et al. Apr 2010 A1
20100088635 Louch Apr 2010 A1
20100100839 Tseng et al. Apr 2010 A1
20100103118 Townsend et al. Apr 2010 A1
20100103124 Kruzeniski et al. Apr 2010 A1
20100105370 Kruzeniski et al. Apr 2010 A1
20100105424 Smuga et al. Apr 2010 A1
20100105438 Wykes et al. Apr 2010 A1
20100105439 Friedman et al. Apr 2010 A1
20100105440 Kruzeniski et al. Apr 2010 A1
20100105441 Voss et al. Apr 2010 A1
20100106607 Riddiford et al. Apr 2010 A1
20100106915 Krishnaprasad et al. Apr 2010 A1
20100107067 Vaisanen Apr 2010 A1
20100107068 Butcher et al. Apr 2010 A1
20100107100 Schneekloth et al. Apr 2010 A1
20100122110 Ordogh May 2010 A1
20100131901 Takahashi et al. May 2010 A1
20100138767 Wang et al. Jun 2010 A1
20100138782 Rainisto Jun 2010 A1
20100138834 Agarwal et al. Jun 2010 A1
20100145675 Lloyd et al. Jun 2010 A1
20100146384 Peev et al. Jun 2010 A1
20100146437 Woodcock et al. Jun 2010 A1
20100157157 Yi et al. Jun 2010 A1
20100159966 Friedman et al. Jun 2010 A1
20100159994 Stallings et al. Jun 2010 A1
20100159995 Stallings et al. Jun 2010 A1
20100162180 Dunnam et al. Jun 2010 A1
20100167699 Sigmund et al. Jul 2010 A1
20100169766 Duarte et al. Jul 2010 A1
20100169772 Stallings et al. Jul 2010 A1
20100175018 Petschnigg et al. Jul 2010 A1
20100175029 Williams Jul 2010 A1
20100180233 Kruzeniski et al. Jul 2010 A1
20100185932 Coffman et al. Jul 2010 A1
20100192102 Chmielewski et al. Jul 2010 A1
20100216491 Winkler et al. Aug 2010 A1
20100223569 Vuong et al. Sep 2010 A1
20100248688 Teng et al. Sep 2010 A1
20100248689 Teng et al. Sep 2010 A1
20100248741 Setlur et al. Sep 2010 A1
20100248787 Smuga et al. Sep 2010 A1
20100251153 SanGiovanni et al. Sep 2010 A1
20100265196 Lee et al. Oct 2010 A1
20100281402 Staikos et al. Nov 2010 A1
20100281409 Rainisto et al. Nov 2010 A1
20100283743 Coddington Nov 2010 A1
20100289806 Lao et al. Nov 2010 A1
20100293056 Flynt et al. Nov 2010 A1
20100295795 Wilairat Nov 2010 A1
20100298034 Shin et al. Nov 2010 A1
20100302148 Tanabe et al. Dec 2010 A1
20100302172 Wilairat Dec 2010 A1
20100302176 Nikula et al. Dec 2010 A1
20100302278 Shaffer et al. Dec 2010 A1
20100302712 Zednicek et al. Dec 2010 A1
20100309148 Fleizach et al. Dec 2010 A1
20100311470 Seo et al. Dec 2010 A1
20100313165 Louch et al. Dec 2010 A1
20100321403 Inadome Dec 2010 A1
20100325411 Jung et al. Dec 2010 A1
20100328431 Kim et al. Dec 2010 A1
20100329642 Kam et al. Dec 2010 A1
20100333008 Taylor Dec 2010 A1
20110004839 Cha et al. Jan 2011 A1
20110004845 Ciabarra Jan 2011 A1
20110018806 Yano Jan 2011 A1
20110029598 Arnold et al. Feb 2011 A1
20110029904 Smith et al. Feb 2011 A1
20110029927 Lietzke et al. Feb 2011 A1
20110029934 Locker et al. Feb 2011 A1
20110043527 Ording et al. Feb 2011 A1
20110055773 Agarawala et al. Mar 2011 A1
20110057953 Horodezky Mar 2011 A1
20110074699 Marr et al. Mar 2011 A1
20110074710 Weeldreyer et al. Mar 2011 A1
20110074719 Yeh et al. Mar 2011 A1
20110078624 Missig et al. Mar 2011 A1
20110087739 Lin et al. Apr 2011 A1
20110087988 Ray et al. Apr 2011 A1
20110093778 Kim et al. Apr 2011 A1
20110093816 Chang et al. Apr 2011 A1
20110093821 Wigdor et al. Apr 2011 A1
20110107272 Aguilar May 2011 A1
20110109563 Liu May 2011 A1
20110113337 Liu et al. May 2011 A1
20110113486 Hunt et al. May 2011 A1
20110119586 Blinnikka et al. May 2011 A1
20110126156 Krishnaraj et al. May 2011 A1
20110154235 Min et al. Jun 2011 A1
20110157027 Rissa Jun 2011 A1
20110161845 Stallings et al. Jun 2011 A1
20110163968 Hogan Jul 2011 A1
20110167341 Cranfill Jul 2011 A1
20110173556 Czerwinski et al. Jul 2011 A1
20110173568 Royal, Jr. et al. Jul 2011 A1
20110175930 Hwang et al. Jul 2011 A1
20110179386 Shaffer et al. Jul 2011 A1
20110179387 Shaffer et al. Jul 2011 A1
20110185318 Hinckley et al. Jul 2011 A1
20110191675 Kauranen Aug 2011 A1
20110202866 Huang et al. Aug 2011 A1
20110209039 Hinckley et al. Aug 2011 A1
20110209089 Hinckley et al. Aug 2011 A1
20110209099 Hinckley et al. Aug 2011 A1
20110209100 Hinckley et al. Aug 2011 A1
20110209101 Hinckley et al. Aug 2011 A1
20110209102 Hinckley et al. Aug 2011 A1
20110209103 Hinckley et al. Aug 2011 A1
20110209104 Hinckley et al. Aug 2011 A1
20110216064 Dahl et al. Sep 2011 A1
20110225547 Fong et al. Sep 2011 A1
20110231796 Vigil Sep 2011 A1
20110252346 Chaudhri Oct 2011 A1
20110252380 Chaudhri Oct 2011 A1
20110252381 Chaudhri Oct 2011 A1
20110276864 Oules Nov 2011 A1
20110279461 Hamilton, II et al. Nov 2011 A1
20110316884 Giambalvo et al. Dec 2011 A1
20110320978 Horodezky et al. Dec 2011 A1
20120005584 Seago et al. Jan 2012 A1
20120009903 Schultz et al. Jan 2012 A1
20120011437 James et al. Jan 2012 A1
20120028687 Wykes et al. Feb 2012 A1
20120050185 Davydov et al. Mar 2012 A1
20120050332 Nikara et al. Mar 2012 A1
20120072853 Krigstrom et al. Mar 2012 A1
20120072953 James et al. Mar 2012 A1
20120084704 Lee et al. Apr 2012 A1
20120102433 Falkenburg Apr 2012 A1
20120151397 Oberstein et al. Jun 2012 A1
20120154303 Lazaridis Jun 2012 A1
20120159395 Deutsch et al. Jun 2012 A1
20120159402 Nurmi et al. Jun 2012 A1
20120161791 Shaw Jun 2012 A1
20120162266 Douglas et al. Jun 2012 A1
20120167008 Zaman et al. Jun 2012 A1
20120167011 Zaman et al. Jun 2012 A1
20120174005 Deutsch et al. Jul 2012 A1
20120174029 Bastide et al. Jul 2012 A1
20120174034 Chae et al. Jul 2012 A1
20120179992 Smuga et al. Jul 2012 A1
20120180001 Griffin et al. Jul 2012 A1
20120208501 Tsuda Aug 2012 A1
20120210265 Delia et al. Aug 2012 A1
20120212495 Butcher et al. Aug 2012 A1
20120216139 Ording et al. Aug 2012 A1
20120233571 Wever et al. Sep 2012 A1
20120235930 Lazaridis et al. Sep 2012 A1
20120244841 Teng et al. Sep 2012 A1
20120254780 Mouton et al. Oct 2012 A1
20120254808 Gildfind Oct 2012 A1
20120265644 Roa et al. Oct 2012 A1
20120272181 Rogers et al. Oct 2012 A1
20120278727 Ananthakrishnan et al. Nov 2012 A1
20120290962 Zielinski et al. Nov 2012 A1
20120299968 Wong et al. Nov 2012 A1
20120304068 Zaman et al. Nov 2012 A1
20120304092 Jarrett et al. Nov 2012 A1
20120304102 LeVee et al. Nov 2012 A1
20120304106 LeVee et al. Nov 2012 A1
20120304107 Nan et al. Nov 2012 A1
20120304108 Jarrett et al. Nov 2012 A1
20120304113 Patten et al. Nov 2012 A1
20120304114 Wong et al. Nov 2012 A1
20120304116 Donahue et al. Nov 2012 A1
20120304117 Donahue et al. Nov 2012 A1
20120304118 Donahue et al. Nov 2012 A1
20120304131 Nan et al. Nov 2012 A1
20120304132 Sareen et al. Nov 2012 A1
20120304133 Nan et al. Nov 2012 A1
20120311485 Caliendo, Jr. et al. Dec 2012 A1
20120322527 Aoki et al. Dec 2012 A1
20120322537 Antkowiak et al. Dec 2012 A1
20120323992 Brobst et al. Dec 2012 A1
20130033525 Markiewicz et al. Feb 2013 A1
20130042203 Wong et al. Feb 2013 A1
20130042206 Zaman et al. Feb 2013 A1
20130044136 Matthews et al. Feb 2013 A1
20130044141 Markiewicz et al. Feb 2013 A1
20130047079 Kroeger et al. Feb 2013 A1
20130047105 Jarrett et al. Feb 2013 A1
20130047117 Deutsch et al. Feb 2013 A1
20130047126 Sareen et al. Feb 2013 A1
20130057587 Leonard et al. Mar 2013 A1
20130057588 Leonard et al. Mar 2013 A1
20130063442 Zaman et al. Mar 2013 A1
20130063443 Garside et al. Mar 2013 A1
20130063465 Zaman et al. Mar 2013 A1
20130063490 Zaman et al. Mar 2013 A1
20130067381 Yalovsky et al. Mar 2013 A1
20130067390 Kwiatkowski et al. Mar 2013 A1
20130067391 Pittappilly et al. Mar 2013 A1
20130067398 Pittappilly et al. Mar 2013 A1
20130067399 Elliott et al. Mar 2013 A1
20130067412 Leonard et al. Mar 2013 A1
20130067420 Pittappilly et al. Mar 2013 A1
20130093757 Cornell Apr 2013 A1
20130102366 Teng et al. Apr 2013 A1
20130179781 Nan et al. Jul 2013 A1
20130268422 Ram et al. Oct 2013 A1
20130297700 Hayton et al. Nov 2013 A1
20140082552 Zaman et al. Mar 2014 A1
20140298245 Tappen et al. Oct 2014 A1
Foreign Referenced Citations (71)
Number Date Country
1512301 Jul 2004 CN
1735856 Feb 2006 CN
1749936 Mar 2006 CN
1786906 Jun 2006 CN
1851641 Oct 2006 CN
1936797 Mar 2007 CN
101006425 May 2007 CN
101075174 Nov 2007 CN
101114303 Jan 2008 CN
101228570 Jul 2008 CN
101770332 Jul 2010 CN
101809531 Aug 2010 CN
102197702 Sep 2011 CN
102624970 Aug 2012 CN
103105995 May 2013 CN
103197837 Jul 2013 CN
103475784 Dec 2013 CN
103562858 Feb 2014 CN
103562917 May 2014 CN
0583060 Feb 1994 EP
1337093 Aug 2003 EP
1526463 Apr 2005 EP
1752868 Feb 2007 EP
1939718 Jul 2008 EP
2535809 Dec 2012 EP
2350991 Dec 2000 GB
2004227393 Aug 2004 JP
2004357257 Dec 2004 JP
2006139615 Jun 2006 JP
200303655 Feb 2002 KR
20060019198 Mar 2006 KR
100854333 May 2006 KR
20070036114 Apr 2007 KR
20070093585 Sep 2007 KR
20070098337 Oct 2007 KR
20070120368 Dec 2007 KR
20080025951 Mar 2008 KR
20080041809 May 2008 KR
20080076390 Aug 2008 KR
20080084156 Sep 2008 KR
20080113913 Dec 2008 KR
20090002951 Jan 2009 KR
20090041635 Apr 2009 KR
20090053143 May 2009 KR
20100010072 Feb 2010 KR
20100048375 May 2010 KR
20100056369 May 2010 KR
201023026 Jun 2010 TW
WO9926127 May 1999 WO
WO2005026931 Mar 2005 WO
WO2005027506 Mar 2005 WO
WO2006019639 Feb 2006 WO
WO2008031871 Mar 2007 WO
WO2007121557 Nov 2007 WO
WO2007134623 Nov 2007 WO
WO2008030608 Mar 2008 WO
WO2008035831 Mar 2008 WO
WO2008084211 Jul 2008 WO
WO2008146784 Dec 2008 WO
WO2009000043 Dec 2008 WO
WO2009012398 Jan 2009 WO
WO2009049331 Apr 2009 WO
WO2009054809 Apr 2009 WO
WO2010024969 Mar 2010 WO
WO2010048229 Apr 2010 WO
WO2010048448 Apr 2010 WO
WO2010048519 Apr 2010 WO
WO2010117643 Oct 2010 WO
WO2010117661 Oct 2010 WO
WO2010135155 Nov 2010 WO
WO2011041885 Apr 2011 WO
Non-Patent Literature Citations (437)
Entry
Chinese Office Action dated Dec. 24, 2013 for Chinese patent application No. 201080015788.3, 10 pages.
Chinese Office Action dated Dec. 26, 2013 for Chinese patent application No. 201080015728.1, 9 pages.
Final Rejection Issued in U.S. Appl. No. 13/335,001, dated Oct. 21, 2014, 35 Pages.
Final Rejection Issued in U.S. Appl. No. 13/335,001, dated Nov. 22, 2013, 19 Pages.
Final Rejection Issued in U.S. Appl. No. 13/335,001, dated May 22, 2014, 20 Pages.
Non-Final Rejection Issued in U.S. Appl. No. 13/335,001, dated Jan. 8, 2014, 18 Pages.
Non-Final Rejection Issued in U.S. Appl. No. 13/335,001, dated Apr. 1, 2015, 26 Pages.
Non-final Office Action Issued in U.S. Appl. No. 13/657,646, dated Jan. 3, 2013, 13 Pages.
Non-Final Rejection Issued in U.S. Appl. No. 13/335,001, dated Aug. 7, 2014, 20 Pages.
Non-Final Rejection Issued in U.S. Appl. No. 13/335,001, dated Sep. 13, 2013, 16 Pages.
Notice of Allowance Issued in U.S. Appl. No. 13/335,001, dated Aug. 26, 2015, 8 Pages.
“25 Apps for Dropbox”, available at: <<http://www.practicalecommerce.com/articles/3984-25-Apps-for-Dropbox>>, Apr. 15, 2013, 10 pages.
Advisory Action Issued in U.S. Appl. No. 12/414,382, dated Jan. 20, 2012, 3 Pages.
Advisory Action Issued in U.S. Appl. No. 12/433,605, dated Apr. 5, 2012, 3 Pages.
Adobe.com, “Adobe Acrobat 8 Standard User Guide”, Adobe Systems Incorporated, 2007, available at: <<http://help.adobe.com/archive/en_US/acrobat/8/standard/acrobat_8_help.pdf>> pp. 34 & 36, 2 pages.
“Advanced Task Killer”, Retrieved From<<http://www.androidtapp.com/advanced-task-killer/>>, Retrieved Date: Sep. 18, 2011, 15 Pages.
Aguilar, “How to Run Multiple Instances of the Same App in Mac OS X”, retrieved on Feb. 20, 2013, available at: <<http://operating-systems.wonderhowto.com/how-to/run-multiple-instances-sameapp-mac-os-x-0140144/>>, Jan. 1, 2013, 4 Pages.
“Alltel Adds Dedicated Search Key to Phones”, available at: <<http://www.phonescoop.com/news/item.php?n=2159>>, Apr. 12, 2007, 2 pages.
Al, “Droptiles—Metro Style Live Tiles Enabled Web 2.0 Dashboard”, retrieved on Mar. 5, 2013, available at: <<http://oazabir.github.com/Droptiles/>>, Jul. 18, 2012, 7 pages.
“Android 2.3 User's Guide”, AUG-2.3-103, Android mobile technology platform 2.3, Dec. 13, 2010, 380 pages.
“Apple iPhone—8GB AT&T”, available at: <<http://nytimes.com.com/smartphones/apple-iphone-8gb-at/4515-6452_7-32309245.html>>, Jun. 29, 2007, 11 pages.
“iPad User Guide”, retrieved on Jun. 17, 2011, available at: <<http://cyndidannerkuhn.info/CDK/iPads_Resources_files/iPad_User_Guide.pdf>>, 154 pages.
Apple.com, “iPod touch User Guide for iPhone OS 3.0 Software”, Apple Inc., 2009, 153 pages.
“Application User Model IDs”, available at: <<http://msdn.microsoft.com/enus/library/dd378459(VS.85).aspx>>, 2010, 6 pages.
“Ask Web Hosting—HTC FUZE From AT&T Fuses Fun and Function With the One-Touch Power of TouchFLO 3D”, Retrieved from <<http://www.askwebhosting.com/story/18501/HTC_FUZE_ATandam_Fuses_Fun_and_Function_Wth_th_%20One-Touch_Power_of_TouchFLO_3D.html>>, Nov. 11, 2008, 3 Pages.
Bamadhaj, Tariq, “Winterface Review”, Retrieved from <<http://web.archive.org/web/20080714033203/http://www.mytodayscreen.com/winterface-review/>>, Jul. 9, 2008, 42 Pages.
“Basics of Your Device: Get Familiar with the Home Screen”, Retrieved from <<http://www.nokia.ca/get-support-and-software/product-support/c6-01/how-to#>>, Retrieved on May 12, 2011, 3 Pages.
Bates, “A Framework to Support Large-Scale Active Applications”, University of Cambridge Computer Laboratory, available at: <<http://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.48.1690&rep=rep1&type=pdf>>, 1996, 8 pages.
Baweja, “OpenHeatMap: Create Heat Maps for Excel Spreadsheets With Geo Data”, available at: <<http://www.makeuseof.com/tag/openheatmap-create-heat-maps/>> Jul. 24, 2010, 4 pages.
Beiber et al.,“Screen Coverage: A Pen-Interaction Problem for PDA's and Touch Screen Computers”, In Proceedings of ICWMC 2007, Mar. 2007, 6 pages.
Best Android Apps Review, “Top Android App: Swipepad”, retrieved on May 11, 2011, available at: <<http://www.bestandroidappsreview.com/2011/01/top-android-app-swipepadlauncher.html>>, 4 pages.
“Better Way to Delete and Close Apps from Task Manager”, Retrieved From<<http://www.ideasproject.com/ideas/14475>>, Retrieved Date: Sep. 21, 2011, 2 Pages.
Bhatnagar, Abhishek, “Locate and Kill Annoying Processes in Windows”, Retrieved From<<http://www.troublefixers.com/locate-and-kill-annoying-processes-in-windows/>>, Apr. 19, 2010, 4 Pages.
Bjork, et al., “Redefining the Focus and Context of Focus+Context Visualizations”, In Proceedings of IEEE Symposium on Information Visualization, Oct. 2000, 9 pages.
“Blackberry Office Tools: Qwerty Convert”, Retrieved from <<http://blackberrysoftwarelist.net/blackberry/download-software/blackberry-office/qwerty_convert.aspx>>, Retrieved on: Nov. 20, 2008, 1 Page.
Blackberry.com, “Class ScrollView”, retrieved on Sep. 28, 2010, available at: <<http://www.blackberry.com/developers/docs/6.0.0api/net/rim/device/api/ui/Scroi1View.html>>, 13 pages.
Blackberryoffice.com,“Blackberry office tools: Qwerty Convert”, retrieved on Nov. 20, 2008, available at: <<http://blackberrysoftwarelist.net/blackberry/download-software/blackberryoffice/qwerty_convert.aspx>>, Nov. 20, 2008, 1 page.
Blankenburg, “31 Days of Mango 1 Day #11: Live Tiles”, retrieved on Mar. 5, 2013, available at: <<http://www.jeffblankenburg.com/2011/11/11/31-days-of-mango-day-11-live-tiles/>>, Nov. 11, 2011, 10 pages.
Bowes, et al., “Transparency for Item Highlighting”, Faculty of Computing Science, Dalhousie University, available at: <<http://torch.cs.dal.ca/-dearman/pubs/GI2003-bowes,dearman,perkins-paper.pdf>>, 2003, 2 pages.
Bruzzese, “Using Windows 7, Managing and Monitoring Windows 7—Chapter 11”, Que Publishing, May 5, 2010, 33 pages.
Buring et al., “User Interaction with Scatterplots on Small Screens—A Comparative Evaluation of Geometric-Semantic Zoom and Fisheye Distortion”, IEEE Transactions on Visualization and Computer Graphics, vol. 12, Issue 5, Sep. 2006, pp. 829-836.
Busmanagement.com, “Content-Centric E-Mail Message Analysis in Litigation Document Reviews”, retrieved on May 6, 2009, available at: <<http://www.busmanagement.com/article/1 ssue-14/Data-ManagemenUContent-Centric-E-Mail-Message-Analysis-in-Litigation-Document-Reviews/>>, 2009, 5 Pages.
Cal, Mike, “How do you dial 1-800-FLOWERS”, Retrieved from <<http://blogs.msdn.com/windowsmobile/archive/2007/02/06/how-do-you-dial-1-800-flowers.aspx>>, Feb. 6, 2007, 24 Pages.
“Calc4M”, Retrieved from <<http://www.hellebo.com/Calc4M.html>>, Sep. 10, 2008, 4 Pages.
Carrera, et al., “Conserving Disk Energy in Network Servers”, available at: <<http://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.84.795&rep=rep1&type=pdf>>, Nov. 2002, 15 pages.
Cawley, “How to Customize Your Windows Phone 7”, retrieved on May 10, 2011, available at: <<http://www. brighthub.com/mobile/windows-mobile-platform/articles/95213.aspx>>, Nov. 12, 2010, 3 pages.
Cawley, “Windows Phone 7 Customization Tips and Tricks”, retrieved on Jun. 20, 2011, available at: <<http://www. brighthub.com/mobile/windows-mobile-platform/articles/95213.aspx>>, May 16, 2011, 2 pages.
Office Action Issued in Chilean Patent Application No. 2379-2011, dated Jul. 3, 2013, 8 Pages.
“Class ScrollView”, Retrieved from <<http://www.blackberry.com/developers/docs/6.0.0api/net/rim/device/api/ui/ScrollView.html>>, Retrieved on: Sep. 28, 2010, 13 Pages.
Office Action Issued in Chinese Patent Application No. 201180071186.4, dated Jun. 2, 2015, 7 Pages.
The Chinese Office Action dated Dec. 4, 2012 for Chinese patent application No. 201080045865.X, 10 pages.
The Chinese Office Action dated Jan. 21, 2013 for Chinese patent application No. 200980142661.5, 12 pages.
The Chinese Office Action dated Jan. 29, 2013 for Chinese patent application No. 200980142632.9, 11 pages.
The Chinese Office Action dated Jan. 6, 2014 for Chinese patent application No. 201110429183.5, 10 pages.
The Chinese Office Action dated Jan. 6, 2014 for Chinese patent application No. 201110437542.1, 10 pages.
First Office Action and Search Report Issued in Chinese Patent Application No. 201080015788.3, dated Dec. 24, 2012, 10 Pages.
First Office Action and Search Report Issued in Chinese Patent Application No. 201080015728.1, dated Dec. 26, 2012, 9 Pages.
Chinese Office Action dated Dec. 27, 2013 for Chinese patent application No. 201110454251.3, 12 pages.
Chinese Office Action dated Dec. 3, 2013 for Chinese patent application No. 201110437572.2, 7 pages.
Chinese Office Action dated Dec. 5, 2012 for Chinese patent application No. 201080023212.1, 10 pages.
Chinese Office Action dated Apr. 3, 2013 for Chinese patent application No. 200980142644.1, 10 pages.
Chinese Office Action dated May 16, 2013 for Chinese patent application No. 201080015728.1, 10 pages.
Chinese Office Action dated Jun. 14, 2013 for Chinese patent application No. 200980142632.9, 6 pages.
Chinese Office Action dated Jun. 4, 2013 for Chinese patent application No. 201080045865.X, 10 pages.
Chinese Office Action dated Jun. 5, 2013 for Chinese patent application No. 201080015788.3, 12 pages.
Chinese Office Action dated Jun. 5, 2013 for Chinese patent application No. 201080023212.1, 8 pages.
Chinese Office Action dated Jul. 1, 2013 for Chinese patent application No. 200980139831.4, 12 pages.
Chinese Office Action dated Aug. 20, 2013 for Chinese patent application No. 200980142644.1, 9 pages.
Chinese Office Action dated Sep. 24, 2012 for Chinese patent application No. 200980142661.5, 8 pages.
Cohen, et al., “Automatic Strategies in the Siemens RTL Tiled Window Manager”, In Proceedings: The 2nd IEEE Conference on Computer Workstations, Mar. 7, 1988, pp. 111-119.
Cohen, et al., “Wang Tiles for Image and Texture Generation”, In Proceedings of ACM Special Interest Group on Computer Graphics and Interactive Techniques, Jul. 2003, 8 pages.
Contextmagic.com,“Email Notification for Microsoft Outlook and Outlook Express”, retrieved on Sep. 29, 2010, available at: <<http://www.contextmagic.com/express-notification/>>, Jul. 21, 2004, 3 pages.
Crouch, “Smartphone Wars: Micron's Slide-to-Unlock Patent”, available at: <<http://patentlyo.com/patent/2013/01/smartphone-wars-microns-slide-to-unlock-patent.html>>, Jan. 30, 2013, 2 pages.
Damien, “7 Ways to Supercharge Multitasking in Android”, retrieved from <http://maketecheasier.com/7-ways-to-supercharge-multitasking-in-android/2011/01/22/>> on May 11, 2011, Jan. 22, 2011, 5 pages.
Danish, “Win7se Brings Mac-Like Screen Corners to Windows 7 & Windows 8”, retrieved on Nov. 23, 2011, available at: <<http://technomondo.com/2011/11/13/win7se-brings-mac-like-screencorners-to-windows7-windows8/>>, Nov. 13, 2011, 4 pages.
“Daphne v1.47”, Retrieved From<<http://www.drk.com.ar/daphne.php>>, Retrieved Date: Sep. 18, 2011, 3 Pages.
Davis, “A WPF Custom Control for Zooming and Panning”, retrieved on Sep. 28, 2010, available at: <<http://www.codeproject.com/KB/WPF/zoomandpancontrol.aspx>>, Jun. 29, 2010, 21 pages.
Delimarsky, “Sending Tile Push Notifications on Windows Phone 7”, retrieved on May 10, 2011, <<http://mobile.dzone.com/articles/sending-tile-push>>, Aug. 25, 2010, 2 pages.
Denoue, et al., “WebNC: Efficient Sharing of Web Applications”, In Proceedings of ACM Special Interest Group on Hypertext, Hypermedia, and Web, Jun. 2009, 2 pages.
“Dial a number”, Retrieved from <<http://www.phonespell.org/dialhelp.html>>, Retrieved on Nov. 20, 2008, 1 Page.
Dixon, “Living in the Online Cloud: The T-Mobile G1 / Google Android Smartphone”, available at: <<http://www.manifest-tech.com/media_pda/t-mobile_g1_android.htm>>, Dec. 2008, 3 pages.
Dolcourt, “Webware”, retrieved on May 5, 2009, available at: <<http://news.cnet.com/webware/?categoryId=2010>>, May 5, 2009, 13 Pages.
Dunsmuir, “Selective Semantic Zoom of a Document Collection”, available at: <http:/ /www.cs.ubc.ca/˜tmm/courses/533/projects/dustin/proposal. pdf>, Oct. 30, 2009, pp. 1-9.
“DuoSense Multi-Touch Gestures”, Retrieved from: <<http://www.n-triq.com/Data/Uploads/Misc/DuoSenseMTG_final.pdf>>, Retrieved on Jul. 2008, 4 Pages.
Egan, “Modifying Live Tiles in a Background Process”, Retrieved on Mar. 5, 2013, available at: <<http://thesociablegeek.com/windows-8/livetiles/modifying-live-tiles-in-a-backgroundprocess/>>, Aug. 31, 2012, 24 pages.
“Elecont Quick Desktop 1.0.43”, retrieved on May 5, 2009, available at: <<http://handheld.softpedia.com/geUSystem-Utilities/Launcher-Applications/Eiecont-Quick-Desktop-72131.shtml>>, Mar. 13, 2009, 2 pages.
“Email Notification for Microsoft Outlook and Outlook Express”, Retrieved from <<http://www.contextmagic.com/express-notification/>>, Retrieved on Sep. 29, 2010, 3 Pages.
“Enhanced IBM Power Systems Software and PowerVM Restructuring”, Retrieved from <<http://www.ibm.com/common/ssi/rep_ca/2/897/ENUS208-082/ENUS208082.PDF>>, Apr. 8, 2008, 19 Pages.
European Office Action dated Aug. 10, 2012 for European patent application No. 10823883.3, 5 pages.
Supplemental European Search Report for Application No. 11867033.0, dated Nov. 27, 2014, 8 pages.
Supplemental European Search Report for Application No. 09822736.6, dated Dec. 18, 2012, 7 pages.
Supplemental European Search Report for Application No. 11866553.8, dated Dec. 9, 2014, 7 pages.
Supplemental European Search Report for Application No. 09818253.8, dated Apr. 10, 2012, 7 pages.
Supplemental European Search Report for Application No. 10762112.0, dated Aug. 2, 2013, 7 pages.
“Exclusive: Windows Mobile 7 to Focus on Touch and Motion Gestures”, Retrieved from <<https://web.archive.org/web/20090622110827/http://anti-linux.blogspot.com/2008/08/exclusive-windows-mobile-7-to-focus-on.html>>, Aug. 1, 2008, 14 pages.
“eXtreme Energy Conservation: Advanced Power-Saving Software for Wireless Devices”, In White paper of Freescale Semiconductor, Inc., Feb. 2006, 15 Pages.
Farrugia, et al., “Cell Phone Mini Challenge: Node-Link Animation Award Animating Multivariate Dynamic Social Networks”, IEEE Symposium on Visual Analytics Science and Technology, Oct. 21, 2008, 2 pages.
Fisher, “Cool Discussion of Push Notifications—Toast and Tile—on Windows Phone”, retrieved on Sep. 29, 2010, available at: <<http://www.windowsphoneexpert.com/Connection/forums/p/4153/18399.aspx>>, May 3, 2010, 3 pages.
Final Office Action Issued in U.S. Appl. No. 12/433,667, dated Jan. 7, 2013, 17 Pages.
Final Office Action issued in U.S. Appl. No. 12/972,967, dated Oct. 11, 2013, 22 Pages.
Final Office Action Issued in U.S. Appl. No. 13/658,694, dated Oct. 30, 2013, 17 Pages.
Final Office Action issued in U.S. Appl. No. 12/983,106, dated Oct. 7, 2013, 21 Pages.
Final Office Action Issued in U.S. Appl. No. 12/413,977, dated Nov. 17, 2011, 16 Pages.
Final Office Action Issued in U.S. Appl. No. 12/469,458, dated Nov. 17, 2011, 15 Pages.
Final Office Action issued in U.S. Appl. No. 13/118,204, dated Nov. 21, 2013, 25 Pages.
Final Office Action Issued in U.S. Appl. No. 12/480,969, dated Nov. 23, 2012, 18 Pages.
Final Office Action Issued in U.S. Appl. No. 12/414,476, dated Dec. 1, 2011, 20 Pages.
Final Office Action issued in U.S. Appl. No. 13/118,321, dated Dec. 19, 2013, 30 Pages.
Final Office Action issued in U.S. Appl. No. 13/118,181, dated Dec. 20, 2013, 19 Pages.
Final Office Action Issued in U.S. Appl. No. 12/414,382, dated Dec. 23, 2011, 7 Pages.
Final Office Action Issued in U.S. Appl. No. 13/118,292, dated Dec. 5, 2014, 30 Pages.
Final Office Action Issued in U.S. Appl. No. 12/244,545, dated Dec. 7, 2011, 16 Pages.
Final Office Action Issued in U.S. Appl. No. 12/469,458, dated Feb. 1, 2013, 19 Pages.
Final Office Action Issued in U.S. Appl. No. 12/433,605, dated Feb. 3, 2012, 11 Pages.
Final Office Action Issued in U.S. Appl. No. 12/469,480, dated Feb. 9, 2012, 17 Pages.
Final Office Action Issued in U.S. Appl. No. 11/502,264, dated Feb. 4, 2010, 15 Pages.
Final Office Action Issued in U.S. Appl. No. 12/560,081, dated Mar. 14, 2012, 16 Pages.
Final Office Action received for U.S. Appl. No. 13/118,221, dated Mar. 17, 2014, 21 Pages.
Final Office Action Issued in U.S. Appl. No. 12/721,422, dated Mar. 7, 2013, 10 Pages.
Final Office Action Issued in U.S. Appl. No. 11/502,264, dated Mar. 29, 2013, 16 Pages.
Final Office Action Issued in U.S. Appl. No. 12/469,480, dated Apr. 10, 2013, 21 Pages.
Final Office Action Issued in U.S. Appl. No. 12/484,799, dated Apr. 30, 2012, 13 Pages.
Final Office Action Issued in U.S. Appl. No. 12/414,476, dated Apr. 8, 2013, 25 Pages.
Final Office Action Issued in U.S. Appl. No. 11/305,789, dated Apr. 1, 2009, 10 Pages.
Final Office Action Issued in U.S. Appl. No. 11/502,264, dated Apr. 3, 2009, 9 Pages.
Final Office Action Issued in U.S. Appl. No. 13/657,646, dated May 6, 2013, 12 Pages.
Final Office Action Issued in U.S. Appl. No. 13/656,354, dated Jun. 17, 2013, 14 Pages.
Final Office Action Issued in U.S. Appl. No. 13/657,789, dated Jun. 21, 2013, 35 Pages.
Final Office Action Issued in U.S. Appl. No. 13/655,386, dated Jun. 6, 2013, 34 Pages.
Final Office Action Issued in U.S. Appl. No. 12/433,605, dated Jul. 17, 2013, 13 Pages.
Final Office Action Issued in U.S. Appl. No. 12/480,969, dated Jul. 24, 2013, 19 Pages.
Final Office Action Issued in U.S. Appl. No. 13/345,383, dated Jul. 30, 2013, 28 Pages.
Final Office Action Issued in U.S. Appl. No. 13/118,347, dated Aug. 15, 2013, 25 Pages.
Final Office Action Issued in U.S. Appl. No. 12/577,400, dated Aug. 16, 2013, 25 Pages.
Final Office Action Issued in U.S. Appl. No. 13/118,339, dated Aug. 22, 2013, 21 Pages.
Final Office Action Issued in U.S. Appl. No. 13/656,574, dated Aug. 23, 2013, 20 Pages.
Final Office Action Issued in U.S. Appl. No. 13/657,621, dated Sep. 10, 2013, 18 Pages.
Final Office Action Issued in U.S. Appl. No. 13/224,258, dated Sep. 11, 2013, 37 Pages.
Final Office Action Issued in U.S. Appl. No. 12/433,667, dated Sep. 13, 2011, 17 Pages.
Final Office Action Issued in U.S. Appl. No. 12/577,400, dated Sep. 14, 2012, 26 Pages.
Final Office Action Issued in U.S. Appl. No. 13/229,693, dated Sep. 4, 2013, 23 Pages.
Final Office Action Issued in U.S. Appl. No. 12/244,545, dated Sep. 7, 2012, 23 Pages.
Freescale Semiconductor, Inc., “eXtreme Energy Conservation: Advanced Power-Saving Software for Wireless Devices”, White Paper, Document No. XTMENRGYCNSVWP, Rev #0, available at <<http://www.freescale.com/files/32bit/doc/white_paper/XTMENRGYCNSVWP.pdf>>, Feb. 2006, 15 pages.
“Freeware.mobi”, retrieved on Nov. 6, 2008, available at: <<http://www.palmfreeware.mobi/downloadpalette.html>>, Oct. 9, 2001, 2 pages.
Gade, “Samsung Alias u740”, retrieved on Nov. 20, 2008, avaiable at: <<http://www.mobiletechreview.com/phones/Samsung-U740.htm>>, Mar. 14, 2007, 6 pages.
Gao, “A General Logging Service for Symbian based Mobile Phones”, Retrieved on Jul. 17, 2008, available at: <<http://www.nada.kth.se/utbildning/grukth/exjobb/rapportlistor/2007/rapporter07/gao_rui_07132.pdf.>>, Feb. 2007, pp. 1-42.
Gnome Developer, “GnomeCanvas”, retrieved on Sep. 28, 2010, available at: <<http://library.gnome.org/devel/libgnomecanvas/unstable/GnomeCanvas.html>>, 11 pages.
“Google Apps versus Office 365: Audit Logs”, Retrieved From <<https://www.linkgard.com/blog/google-apps/google-apps-vs-office-365-audit-logs.html>>, Retrieved on: Jul. 23, 2013, 8 Pages.
Gralla, “Windows XP Hacks, Chapter 13—Hardware Hacks”, O'Reilly Publishing, Feb. 23, 2005, 25 pages.
Guiding Tech, CloseAll Closes All Open Program Windows in One Click, online tech blog, from u http://www.guidingtech.com/2470/close-open-program-windows-in-one-click/, retrieved archive version of 4/24/201 O from wayback machine (http://archive.org/web/web.php).
Ha, et al., “SIMKEYS: An Efficient Keypad Configuration for Mobile Communications”, retrieved on Dec. 17, 2008, available at: <<http://ieeexplore.ieee.org/stamp/stamp.jsp?arnumber=01362557>>, Nov. 2004, 7 Pages.
Harrison, “Symbian OS C++ for Mobile Phones vol. 3”, retrieved on Oct. 23, 2008, available at: <<http://www.amazon.co.uk/Symbian-OS-Mobile-Phones-Press/dp/productdescription/0470066415>>, Symbian Press, Jun. 16, 2003, 4 pages.
Hickey, “Google Android has Landed; T-Mobile, HTC Unveil G1”, retrieved on Nov. 26, 2008, available at: <<http://www.crn.com/retail/21 0603348>>, Sep. 23, 2008, 4 pages.
Horowitz, “Installing and Tweaking Process Explorer part 2”, Retrieved on Mar. 12, 2013, available at: <<http://web.archive.org/web/20110510093838/http://blogs.computerworld.com/16165/installing_and_tweaking_process_explorer_part_2>>, May 23, 2010, 7 pages.
“How Do I Cancel a “Drag” Motion on an Android Seekbar?”, retrieved on Jun. 20, 2011, available at: <<http://stackoverflow.com/questions/2917969/how-do-i-cancel-a-drag-motion-on-an-android-seekbar>>, May 28, 2010, 1 page.
“How do I use Categories with my Weblog?”, retrieved on Sep. 28, 2010, available at: <<http://tpsupport.mtcs.sixapart.com/tp/ustp1/how_do_i_use_categories_with_my_weblog.html>>, Sep. 16, 2009, 3 pages.
“How do you dial 1-800-FLOWERS”, retrieved on Nov. 20, 2008, available at: <<http://blogs.msdn .com/windowsmobile/archive/2007/02/06/how-do-you-d ial-1-800-flowers.aspx>>, Feb. 6, 2007, 24 pages.
“How-to close an application in BlackBerry PlayBook?”, Retrieved From<<http://stackoverflow.com/questions/5277027/how-to-close-an-application-in-blackberry-playbook>>, Retrieved Date: Sep. 18, 2011, 2 Pages.
HTC.com, “Ask Web Hosting”, available at: <<http:/ /www.askwebhosting.com/story/18501 /HTC_FUZE_From_ATandamp T Fuses_Fun_and_Function_With_the_One-Touch_Power_of_ TouchFL0_3D.html>>, Nov. 11, 2008, 3 pages.
IBM, “Enhanced IBM Power Systems Software and PowerVM Restructuring”, IBM United States Announcement 208-082, available at <<http://www.ibm.com/common/ssi/rep_ca/2/897/ENUS208-082/ENUS208082.PDF>>, Apr. 8, 2008, pp. 1-19.
IBM.com, “Remapping the Keyboard”, retrieved on Dec. 11, 2008, available at: <<http://publib.boulder.ibm.com/infocenter/hodhelp/v9r0/index.jsp?topic=/com.ibm.hod9.doc/help/assignkey.html>>, Jul. 15, 2005, 5 Pages.
“IntelliScreen—New iPhone App Shows Today Screen Type Info in Lock Screen”, Retrieved from: <http://justanotheriphoneblog.com/wordpress//2008/05/13/intelliscreennew-iphone-app-shows-today-screen-type-info-on-lock-screen/> on Nov. 12, 2008, May 13, 2008, 11 pages.
“Internet Explorer Window Restrictions”, Retrieved from <<http://technet.microsoft.com/en-us/library/cc759517(WS.10).aspx>>, Jun. 28, 2011, 5 pages.
“Introducing Application Styling for Windows Forms”, In Infragistics Software Manual, Version 7.3.20073.1043, Nov. 2007, 95 pages.
“Introduction to Windows Touch”, Retrieved from <<http://download.microsoft.com/download/a/d/f/adf1347d-08dc-41a4-9084-623b1194d4b2/Win7_touch.docx>>, Dec. 18, 2008, 7 Pages.
“iPad User Guide”, Retrieved from <<http://cyndidannerkuhn.info/CDK/iPads_Resources_files/iPad_User_Guide.pdf>>, Retrieved on Jun. 17, 2011, 154 Pages.
“iPod touch User Guide for iPhone OS 3.0 Software”, Retrieved from <<https://manuals.info.apple.com/MANUALS/1000/MA1135/en_US/iPod_touch_3.0_User_Guide.pdf>>, 2009, 153 Pages.
PCT Search Report and Written Opinion dated Jan. 19, 2011 for PCT Application No. PCT/US2010/038730, 8 pages.
PCT Search Report & Written Opinion for Application No. PCT/CN2014/074825, dated Jan. 5, 2015, 15 pages.
PCT Search Report and Written Opinion dated Oct. 12, 2010 for PCT Application No. PCT/US2010/028555, 10 pages.
PCT Search Report and Written Opinion dated Oct. 4, 2010, for PCT Application No. PCT/US2010/028699, 10 pages.
PCT Search Report and Written Opinion dated Nov. 7, 2013 for PCT Application No. PCT/US2013/059563, 10 Pages.
PCT Search Report and Written Opinion dated Nov. 9, 2010 for PCT Application No. PCT/US2010/028553, 9 pages.
PCT Search Report and Written Opinion dated Dec. 12, 2012 for PCT Application No. PCT/US2011/067075, 10 pages.
PCT Search Report and Written Opinion dated Dec. 27, 2012 for PCT Application No. PCT/US2012/047091, 15 pages.
PCT Search Report and Written Opinion dated Dec. 29, 2010 for PCT Application No. PCT/US2010/034772, 12 pages.
PCT Search Report & Written Opinion for Application No. PCT/CN2014/074788, dated Dec. 30, 2014, 12 pages.
PCT Search Report & Written Opinion for Application No. PCT/CN2014/074783, dated Dec. 31, 2014, 15 pages.
PCT Search Report and Written Opinion dated Feb. 4, 2014 for PCT Application No. PCT/US2013/061066, 11 Pages.
PCT Search Report and Written Opinion dated Mar. 27, 2012 for PCT Application No. PCT/US2011/055513, 8 pages.
PCT Search Report and Written Opinion dated Apr. 24, 2012 for PCT Application No. PCT/US2011/055511, 9 pages.
PCT Search Report and Written Opinion dated May 10, 2012 for PCT Application No. PCT/US2011/055523, 9 pages.
PCT Search Report and Written Opinion dated May 14, 2010 for PCT Application No. PCT/US2009/061864, 10 pages.
PCT Search Report and Written Opinion dated May 15, 2012 for PCT Application No. PCT/US2011/055521, 9 pages.
PCT Search Report and Written Opinion dated May 15, 2012 for PCT Application No. PCT/US2011/055522, 9 pages.
PCT Search Report and Written Opinion dated May 2, 2011 for PCT Application No. PCT/US2010/052119, 8 pages.
PCT Search Report and Written Opinion dated May 22, 2012 for PCT Application No. PCT/US2011/055514, 8 pages.
PCT Search Report and Written Opinion dated May 24, 2012 for PCT Application No. PCT/US2011/055512, 8 pages.
PCT Search Report and Written Opinion dated May 26, 2010 for PCT Application No. PCT/US2009/061382, 10 pages.
PCT Search Report and Written Opinion dated May 9, 2012 for PCT Application No. PCT/US2011/055520, 8 pages.
PCT Search Report and Written Opinion dated Jun. 1, 2012 for PCT Application No. PCT/US2011/055524, 8 pages.
PCT Search Report and Written Opinion dated Jun. 7, 2010 for PCT Application No. PCT/US2009/061735, 11 pages.
PCT Search Report and Written Opinion dated Aug. 29, 2012 for PCT Application No. PCT/US2011/065702, 8 pages.
PCT Search Report and Written Opinion dated Sep. 12, 2012 for PCT Application No. PCT/US2011/055496, 9 Pages.
PCT Search Report and Written Opinion dated Sep. 17, 2012 for PCT Application No. PCT/US2011/055736, 8 pages.
PCT Search Report and Written Opinion dated Sep. 17, 2012 for PCT Application No. PCT/US2011/067073, 8 pages.
PCT Search Report and Written Opinion dated Sep. 21, 2012 for PCT Application No. PCT/US2011/055712, 9 pages.
PCT Search Report and Written Opinion dated Sep. 26, 2012 for PCT Application No. PCT/US2011/055493, 9 pages.
PCT Search Report and Written Opinion dated Sep. 27, 2012 for PCT Application No. PCT/US2011/055725, 10 pages.
PCT Search Report and Written Opinion dated Sep. 27, 2012 for PCT Application No. PCT/US2011/055478, 9 pages.
PCT Search Report and Written Opinion dated Sep. 27, 2012 for PCT Application No. PCT/US2011/055746, 9 pages.
PCT Search Report & Written Opinon for Application No. PCT/C2014/074793, dated Sep. 30, 2014, 15 pages.
Janecek, et al., “An Evaluation of Semantic Fisheye Views for Opportunistic Search in an Annotated Image Collection”, available at: <<http://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.67.3084&rep=rep1&type=pdf>>, Feb. 15, 2005, pp. 1-15.
janti-linux, “Exclusive: Windows Mobile 7 to Focus on Touch and Motion Gestures”, retrieved on May 6, 2009, available at: <<http:/Janti-linux.blogspot.com/2008/08/exclusive-windows-mobile-7-to-focuson.html>>, Aug. 1, 2008, 14 pages.
Office Action Issued in Japanese Patent Application No. 2012-503523, dated Apr. 22, 2013, 5 pages.
The Japanese Office Action dated Jul. 5, 2013 for Japanese patent application No. 2011-533353, 9 pages.
Kandogan, et al., “Elastic Windows: Improved Spatial Layout and Rapid Multiple Window Operations”, In Proceedings of the Workshop on Advanced Visual Interfaces, May 27, 1996, 10 pages.
Kcholi, “Windows CE .NET Interprocess Communication”, Retrieved on Jul. 17, 2008, available at: <<http://msdn.microsoft.com/en-us/library/ms836784.aspx>>, Jan. 2004, 15 Pages.
Keranen, “OpenGL-based User Interface Toolkit for Symbian Mobile Devices”, Master of Science Thesis, Tamere University of Technology, Department of Information Technology, Apr. 6, 2005, 88 pages.
“Keyboard (5)”, retrieved on Dec. 11, 2008, available at: <<http://landru.uwaterloo.ca/cgibin/man .cgi?section=5&topic=keyboard>>, Aug. 11, 1997, 8 Pages.
“Keyboard Shortcuts”, retrieved on Aug. 3, 2009, available at: <<http://www.pctoday.com/editorial/article.asp?article=articles%2F2005%2Ft0311%2F26t11%2F26t11.asp>>, Nov. 2005, 5 pages.
KillerMobile, “SecureMe-Anti-Theft Security Application for S60 3rd”, retrieved on Jun. 28, 2011, available at: <<http:/www.killermobile.com/newsite/mobile-software/s60-applications/secureme-%11-anti% 11theft-security-application-for-s60-3rd.htm>>, Dec. 15, 2008, 3 pages.
“Kiosk Browser Chrome Customization Firefox 2.x”, retrieved on Oct. 22, 2008, available at: <<http://stlouisshopper.com/cgi-bin/mozdev-wiki/,pi?ChromeCustomization>>, Making a new chrome for the kiosk browser, Kiosk Project Kiosk Browser Chrome Customization Firefox-2.x, Aug. 16, 2007, 2 pages.
Kurdi, “Acer GridVista: snap your windows to pre-defined sections on your screen(s)”, retrieved on Jun. 30, 2013, available at: <<http://www.freewaregenius.com/acer-gridvista-snap-your-windows-topre-defined-sections-of-your-screens/>>, Jan. 19, 2010, 6 pages.
Kurdi, “Win Split Revolution”, retrieved on Jun. 30, 2013, available at: <<http://www.freewaregenius.com/winsplit-revolution/>>, Aug. 22, 2007, 4 Pages.
La, “Parallax Gallery”, available at: <<http://webdesignerwall.comtutorials/parallaxgallery/comment-page-1>>, Apr. 25, 2008, 16 pages.
“Live Photo Gallery—Getting Started—from Camera to Panorama”, retrieved on May 5, 2009, available at: <<http:/ /webdotwiz.spaces.live.com/blog/cns!2782760752693233! 1729.entry>>, Sep. 2008, 7 Pages.
Livingston, et al., “Windows 95 Secrets”, 1995, I DG Books Worldwide, 3rd Edition, 1995, pp. 121-127.
“IntelliScreen—New iPhone App Shows Today Screen Type Info in Lock Screen”, retrieved on Nov. 12, 2008, available at: <<http://justanotheriphoneblog.com/wordpress//2008/05/13/intelliscreennew-iphone-app-shows-today-screen-type-info-on-lock-screen/>>, May 13, 2008, 11 pages.
Long, “Gmail Manager 0.6”, retrieved on Sep. 29, 2010, available at: <<https://addons.mozilla.org/en-US/firefox/addon/1320/>>, Jan. 27, 2010, 4 pages.
“Magic mouse”, Retrieved from <<http://web.archive.org/web/20110514134135/http://www.apple.com/magicmouse/>>, Retrieved on May 10, 2011, 3 Pages.
Mann, et al., “Spectrum Analysis of Motion Parallax in a 3D Cluttered Scene and Application to Egomotion”, Journal of the Optical Society of America A, vol. 22, No. 9, Sep. 2005, pp. 1717-1731.
Mantia, “Multitasking: What Does It Mean?”, retrieved on Sep. 23, 2011, available at: <<http://mantia.me/blog/multitasking/>>, 3 pages.
Mao, “Comments of Verizon Wireless Messaging Services, LLC”, Retrieved on May 6, 2009, available at: <<http://www.ntia.doc.gov/osmhome/warnings/comments/verizon.htm>>, Aug. 18, 2000, 5 Pages.
Marie, “MacBook Trackpad Four Fingers Swipe Left/Right to Switch Applications”, MacBook Junkie, retrieved on May 11, 2011, available at: <<http://www.macbookjunkie.com/macbook-trackpadfour-fingers-swipe-left-right-to-switch-applications/>, Nov. 13, 2010, 4 pages.
Matejka, et al., “Patina: Dynamic Heatmaps for Visualizing Application Usage”, In Proceedings: The SIGCHI Conference on Human Factors in Computing Systems, Autodesk Research, Toronto, Ontario, Canada, Apr. 27, 2013, pp. 3227-3236.
Mazo, “How to Switch Applications and Multitask on the Galaxy S3”, retrieved on Feb. 18, 2013, available at: <<http://www.androidcentral.com/how-switch-applications-and-multitask-samsung-galaxys3>>, Jul. 17, 2012, 7 pages.
Mei, et al., “Probabilistic Multimodality Fusion for Event Based Home Photo Clustering”, available at: <<http://ieeexplore.ieee.org//stamp/stamp.jsp?tp=&arnumber=04036960.>>, Dec. 26, 2006, pp. 1757-1760.
Michaluk, “Using the Application Switcher and Closing Apps When Finished to Maximize your BlackBerry Efficiency”, retrieved on Feb. 19, 2013, available at: <<http://crackberry.com/usingapplication-switcher-and-closing-apps-when-finished-maximize-your-blackberryefficiency>>, Aug. 17, 2009, 14 pages.
“Microsoft Windows, The Start Menu (overview)”, Retrieved from <<http://web.archive.org/web/20101219151141/http://windows.microsoft.com/en-US/windows-vista/The-Start-menu-overview>>, Dec. 19, 2010, 5 Pages.
Microsoft.com, “Internet Explorer Window Restrictions,” retrieved on Jun. 28, 2011, available at: <<http://technet.microsoft.com/en-us/library/cc759517(WS.10).aspx>>, 5 pages.
Microsoft.com “Introduction to Windows Touch”, available at: <<<http://download.microsoft.com/download/a/d/f/adf1347d-08dc-41a4-9084-623b1194d4b2/Win7_touch.docx>>, Dec. 18, 2008, pp. 1-7.
Microsoft, “Push Notifications Overview for Windows Phone”, retrieved on Sep. 30, 2010, available at: <<http://msdn.microsoft.com/en-us/library/ff402558%28VS.92%29.aspx>>, Sep. 3, 2010, 1 page.
Microsoft.com, “Notifications”, retrieved on May 10, 2011, available at: <<http://msdn.microsoft.com/enus/library/aa511497.aspx>>, 16 pages.
Microsoft.com, “Snap”, Windows 7 Features—retrieved on Sep. 23, 2011, available at: <<http://windows.microsoft.com/en-US/windows7/products/features/snap>>, 2 pages.
“MIDTB Tip Sheet: Book Courier”, retrieve on Dec. 11, 2008, available at: <<http://www.midtb.org/tipsbookcourier.htm>>, Sep. 26, 2005, 6 Pages.
“Mobile/U 1/Designs/TouchScreen/workingU I”, retrieve on Oct. 26, 2009, available at: <<https:/ /wiki.mozilla.org/Mobile/U 1/Designs/TouchScreen/workingUI>>, 2009, 30 pages.
“Nokia E61 Tips and Tricks for Keyboard Shortcuts”, retrieved on Dec. 17, 2008, available at: <<http://www.mobiletopsoft.com/board/181 O/nokia-e61-tips-and-tricks-for-keyboardshortcuts.html>>, Jan. 27, 2006, 2 Pages.
MobilityFlow, “Touch Shell Free”, retrieved on May 5, 2009, available at: <<http://www.pocketpcfreeware.mobi/download-touch-shell-free.html>>, Feb. 23, 2009, 2 Pages.
“moGo beta v.0.4”, retrieved on Sep. 27, 2010, available at: <<http://forum.xda-developers.com/showthread.php?t=375196>>, Mar. 7, 2008, 10 pages.
MSOffice, “Working with Multiple Windows”, MSOFFICE tutorial!—retrieved on Sep. 23, 2011, available at: <<http://www.msoffice-tutorial.com/working-with-multiple-windows.php>>, 3 pages.
“Multi-touch”, Retrieved from <<http://web.archive.org/web/20090725202338/http://en.wikipedia.org/wiki/Multi-Touch>>, Apr. 17, 2009, 8 Pages.
“My Favorite Gadgets, System Monitor II”, retrieved on Mar. 12, 2013, available at: <<http://www. myfavoritegadgets.info/monitors/System MonitorII/systemmonitorII.html>>, Jun. 8, 2010, 5 pages.
“New Features in WhatsUp Gold v12.0”, retrieved on Jun. 10, 2011, available at: <<http://www.netbright.co.th/?name=product&file=readproduct&id=12>>, 4 pages.
Non Final Office Action Issued in U.S. Appl. No. 12/433,605, dated Jan. 11, 2013, 7 Pages.
Non-final Office Action Issued in U.S. Appl. No. 12/414,434, dated Jan. 17, 2012, 7 Pages.
Non-final Office Action Issued in U.S. Appl. No. 12/978,184, dated Jan. 23, 2013, 7 Pages.
Non-final Office Action Issued in U.S. Appl. No. 12/972,967, dated Jan. 30, 2013, 19 Pages.
Non-final Office Action Issued in U.S. Appl. No. 13/656,574, dated Jan. 31, 2013, 21 Pages.
Non-final Office Action Issued in U.S. Appl. No. 13/224,258, dated Jan. 8, 2013, 35 Pages.
Non-Final Office Action Issued in U.S. Appl. No. 13/657,789, dated Jan. 9, 2013, 38 Pages.
Non Final Office Action Issued in U.S. Appl. No. 12/721,422, dated Oct. 1, 2012, 8 Pages.
Non-Final Office Action Issued in U.S. Appl. No. 12/469,480, dated Oct. 17, 2012, 16 Pages.
Non Final Office Action Issued in U.S. Appl. No. 13/228,707, dated Oct. 25, 2013, 12 Pages.
Non Final Office Action Issued in U.S. Appl. No. 13/229,155, dated Nov. 18, 2013, 14 Pages.
Non-final Office Action Issued in U.S. Appl. No. 12/470,558, dated Nov. 22, 2011, 9 Pages.
Non Final Office Action Issued in U.S. Appl. No. 13/228,876, dated Nov. 22, 2013, 14 Pages.
Non-final Office Action Issued in U.S. Appl. No. 14/089,149, dated Nov. 3, 2014, 8 Pages.
Non-final Office Action Issued in U.S. Appl. No. 12/469,419, dated Nov. 9, 2011, 15 Pages.
Non-Final Office Action Issued in U.S. Appl. No. 12/414,476, dated Nov. 9, 2012, 22 Pages.
Non-final Office Action Issued in U.S. Appl. No. 12/983,106, dated Nov. 9, 2012, 17 Pages.
Non-final Office Action Issued in U.S. Appl. No. 13/655,390, dated Dec. 17, 2012, 12 Pages.
Non-Final Office Action Issued in U.S. Appl. No. 13/492,495, dated Dec. 19, 2012, 6 Pages.
Non-final Office Action Issued in U.S. Appl. No. 13/655,386, dated Dec. 26, 2012, 23 Pages.
Non-final Office Action Issued in U.S. Appl. No. 12/560,081, dated Dec. 7, 2011, 16 Pages.
Non-final Office Action Issued in U.S. Appl. No. 12/484,845, dated Dec. 7, 2011, 16 Pages.
Non-final Office Action Issued in U.S. Appl. No. 12/977,584, dated Dec. 7, 2012, 8 Pages.
Non Final Office Action Issued in U.S. Appl. No. 13/228,888, dated Feb. 10, 2014, 23 Pages.
Non-final Office Action Issued in U.S. Appl. No. 13/118,339, dated Feb. 11, 2013, 15 Pages.
Non-final Office Action Issued in U.S. Appl. No. 13/118,347, dated Feb. 12, 2013, 14 Pages.
Non Final Office Action Issued in U.S. Appl. No. 13/658,694, dated Feb. 13, 2014, 19 Pages.
Non-Final Office Action Issued in U.S. Appl. No. 13/118,314, dated Feb. 27, 2014, 13 Pages.
Non-Final Office Action Issued in U.S. Appl. No. 13/118,204, dated Feb. 28, 2013, 13 Pages.
Non-Final Office Action Issued in U.S. Appl. No. 13/345,383, dated Feb. 28, 2013, 27 Pages.
Non-Final Office Action Issued in U.S. Appl. No. 12/433,667, dated Feb. 3, 2012, 16 Pages.
Non-final Office Action Issued in U.S. Appl. No. 13/656,354, dated Feb. 6, 2013, 10 Pages.
Non-Final Office Action Issued in U.S. Appl. No. 13/196,272, dated Feb. 6, 2013, 10 Pages.
Non-Final Office Action Issued in U.S. Appl. No. 13/656,639, dated Feb. 7, 2013, 29 Pages.
Non-final Office Action Issued in U.S. Appl. No. 13/657,621, dated Feb. 7, 2013, 19 Pages.
Non-Final Office Action Issued in U.S. Appl. No. 13/118,221, dated Mar. 1, 2013, 22 Pages.
Non-Final Office Action Issued in U.S. Appl. No. 13/229,693, dated Mar. 12, 2013, 21 Pages.
Non-Final Office Action Issued in U.S. Appl. No. 12/244,545, dated Mar. 27, 2012, 18 Pages.
Non Final Office Action Issued in U.S. Appl. No. 13/229,556, dated Mar. 28, 2014, 13 Pages.
Non-Final Office Action Issued in U.S. Appl. No. 13/118,181, dated Mar. 4, 2013, 22 Pages.
Non Final Office Action issued in U.S. Appl. No. 13/118,302, dated Mar. 4, 2014, 11 Pages.
Non-Final Office Action Issued in U.S. Appl. No. 13/118,257, dated Mar. 5, 2013, 19 Pages.
Non-Final Office Action Issued in U.S. Appl. No. 12/577,400, dated Apr. 11, 2012, 22 Pages.
Non-Final Office Action Issued in U.S. Appl. No. 12/577,400, dated Apr. 15, 2013, 25 Pages.
Non-Final Office Action Issued in U.S. Appl. No. 13/658,694, dated Apr. 26, 2013, 13 Pages.
Non Final Office Action Issued in U.S. Appl. No. 12/480,969, dated Apr. 4, 2013, 22 Pages.
Non-Final Office Action Issued in U.S. Appl. No. 12/469,419, dated May 23, 2012, 13 Pages.
Non Final Office Action Issued in U.S. Appl. No. 12/469,458, dated May 3, 2013, 21 Pages.
Non-Final Office Action Issued in U.S. Appl. No. 12/414,434, dated May 31, 2012, 7 Pages.
Non-Final Office Action Issued in U.S. Appl. No. 13/118,321, dated Jun. 10, 2013, 32 Pages.
Non-final Office Action Issued in U.S. Appl. No. 11/215,052, dated Jun. 23, 2011, 17 Pages.
Non-final Office Action Issued in U.S. Appl. No. 12/433,605, dated Jun. 24, 2011, 10 Pages.
Non Final Office Action issued in U.S. Appl. No. 13/118,292, dated Jun. 6, 2014, 25 Pages.
Non-final Office Action Issued in U.S. Appl. No. 12/433,667, dated Jun. 7, 2011, 15 Pages.
Non-final Office Action Issued in U.S. Appl. No. 12/469,458, dated Jul. 1, 2011, 15 Pages.
Non-Final Office Action Issued in U.S. Appl. No. 12/413,977, dated Jul. 19, 2011, 17 Pages.
Non-Final Office Action Issued in U.S. Appl. No. 12/413,977, dated Jul. 20, 2012, 18 Pages.
Non-Final Office Action Issued in U.S. Appl. No. 13/073,300, dated Jul. 25, 2013, 13 Pages.
Non-final Office Action Issued in U.S. Appl. No. 12/414,382, dated Jul. 26, 2011, 9 Pages.
Non-final Office Action Issued in U.S. Appl. No. 12/414,458, dated Jul. 6, 2011, 8 Pages.
Non-Final Office Action Issued in U.S. Appl. No. 12/484,799, dated Aug. 11, 2011, 12 Pages.
Non-Final Office Action Issued in U.S. Appl. No. 12/244,545, dated Aug. 17, 2011, 15 Pages.
Non-final Office Action Issued in U.S. Appl. No. 12/414,434, dated Aug. 2, 2011, 6 Pages.
Non Final Office Action Issued in U.S. Appl. No. 12/469,480, dated Aug. 27, 2013, 22 Pages.
Non-final Office Action Issued in U.S. Appl. No. 12/414,455, dated Aug. 29, 2011, 8 Pages.
Non-Final Office Action Issued in U.S. Appl. No. 12/414,476, dated Aug. 3, 2011, 21 Pages.
Non-final Office Action Issued in U.S. Appl. No. 12/484,799, dated Aug. 7, 2012, 13 Pages.
Non-final Office Action Issued in U.S. Appl. No. 12/480,969, dated Aug. 7, 2012, 15 Pages.
Non-Final Office Action Issued in U.S. Appl. No. 11/502,264, dated Sep. 14, 2012, 14 Pages.
Non-final Office Action Issued in U.S. Appl. No. 13/492,495, dated Sep. 17, 2012, 8 Pages.
Non-final Office Action Issued in U.S. Appl. No. 11/305,789, dated Sep. 21, 2009, 5 Pages.
Non-Final Office Action Issued in U.S. Appl. No. 12/469,458, dated Sep. 21, 2012, 14 Pages.
Non-Final Office Action Issued in U.S. Appl. No. 12/469,480, dated Sep. 22, 2011, 14 Pages.
Non-Final Office Action Issued in U.S. Appl. No. 13/196,272, dated Sep. 3, 2013, 5 Pages.
Non-final Office Action Issued in U.S. Appl. No. 11/502,264, dated Sep. 30, 2009, 16 Pages.
Notice of Allowance Issued in U.S. Appl. No. 12/414,455, dated Jan. 4, 2012, 5 Pages.
Notice of Allowance Issued in U.S. Appl. No. 12/977,584, dated Oct. 11, 2013, 2 Pages.
Notice of Allowance Issued in U.S. Appl. No. 12/484,799, dated Oct. 22, 2012, 10 Pages.
Notice of Allowance Issued in U.S. Appl. No. 12/414,458, dated Oct. 31, 2011, 3 Pages.
Notice of Allowance Issued in U.S. Appl. No. 11/305,789, dated Nov. 23, 2009, 8 Pages.
Notice of Allowance Issued in U.S. Appl. No. 12/469,419, dated Nov. 27, 2012, 14 Pages.
Notice of Allowance Issued in U.S. Appl. No. 12/414,458, dated Nov. 29, 2011, 2 Pages.
Notice of Allowance Issued in U.S. Appl. No. 12/978,184, dated Nov. 6, 2013, 7 Pages.
Notice of Allowance Issued in U.S. Appl. No. 13/196,272, dated Nov. 8, 2013, 10 Pages.
Notice of Allowance Issued in U.S. Appl. No. 14/089,149, dated Dec. 5, 2014, 5 Pages.
Notice of Allowance Issued in U.S. Appl. No. 14/089,149, dated Feb. 20, 2015, 2 Pages.
Notice of Allowance Issued in U.S. Appl. No. 11/215,052, dated Mar. 14, 2012, 5 Pages.
Notice of Allowance Issued in U.S. Appl. No. 12/484,845, dated Mar. 16, 2012, 5 Pages.
Notice of Allowance Issued in U.S. Appl. No. 14/089,149, dated Mar. 20, 2015, 2 Pages.
Notice of Allowance Issued in U.S. Appl. No. 12/470,558, dated Apr. 2, 2012, 7 Pages.
Notice of Allowance Issued in U.S. Appl. No. 13/492,495, dated Apr. 26, 2013, 5 Pages.
Notice of Allowance Issued in U.S. Appl. No. 12/414,382, dated Apr. 4, 2012, 4 Pages.
Notice of Allowance Issued in U.S. Appl. No. 13/655,390, dated May 24, 2013, 5 Pages.
Notice of Allowance Issued in U.S. Appl. No. 12/977,584, dated Jun. 19, 2013, 5 Pages.
Notice of Allowance Issued in U.S. Appl. No. 12/433,667, dated Jun. 25, 2013, 15 Pages.
Notice of Allowance Issued in U.S. Appl. No. 12/721,422, dated Jul. 11, 2013, 9 Pages.
Notice of Allowance Issued in U.S. Appl. No. 12/433,667, dated Aug. 1, 2013, 2 Pages.
Notice of Allowance Issued in U.S. Appl. No. 12/414,458, dated Aug. 10, 2011, 6 Pages.
Notice of Allowance Issued in U.S. Appl. No. 12/414,434, dated Aug. 17, 2012, 5 Pages.
Notice of Allowance Issued in U.S. Appl. No. 13/656,639, dated Aug. 19, 2013, 15 Pages.
Notice of Allowance Issued in U.S. Appl. No. 12/978,184, dated Aug. 2, 2013, 5 Pages.
Notice of Allowance Issued in U.S. Appl. No. 12/470,558, dated Aug. 23, 2012, 2 Pages.
“Nokia E61 Tips and Tricks for Keyboard Shortcuts”, Retrieved frm <<http://web.archive.org/web/20070129151327/http://www.mobiletopsoft.com/board/1810/nokia-e61-tips-and-tricks-for-keyboard-shortcuts.html>>, Jan. 27, 2006, 2 Pages.
Nokia.com, “Basics of Your Device: Get Familiar with the Home Screen”, Nokia USA—How to—retrieved on Nov. 20, 2008, available at: <<http://www.nokia.ca/get-support-and-software/product-support/c6-01/how-to#>>, 3 pages.
Nordgren, “Development of a Touch Screen Interface for Scania Interactor”, Master's Thesis in C, available at: <<http://www.cs.umu.se/education/examina/Rapporter/PederNordgren.pdf>>, Computing Science, UMEA University, Apr. 10, 2007, pp. 1-59.
“Notifications”, Retrieved from <<https://msdn.microsoft.com/en-us/library/dn742472.aspx>>, Retrieved on: May 10, 2011, 16 Pages.
NTrig.com, “DuoSenseTM Multi-Touch Gestures”, available at: <<http://www.ntrig.com/Data/Uploads/Misc/DuoSenseMTG_final.pdf>>, Aug. 2008, 4 pages.
Ntwind software, “Switch between Windows of the Same App”, retrieved on Feb. 20, 2013, available at: <<http://www.ntwind.com/software/vistaswitcher/instance-switcher.html>>, May 12, 2012, 2 pages.
Oliver, “Potential iPhone Usability and Interface Improvements”, retrieved on Nov. 12, 2008, available at: <<http://www.appleinsider.com/articles/08/09/18/potential_iphone usability_and_interface_improvements.html>>, AppleInsider, Sep. 18, 2008, 4 pages.
“OmneMon[TM] System Resource Metrics”, retrieved on Jun. 10, 2011, available at: <<http://www.omnesys.com/documents/OmneMonSRM_Brochure.pdf>>, 3 pages.
“ONYX Graphics Announces New ONYX Prepedge Job Preparation Software”, retrieved on May 10, 2011, available at: <<http://www.largeformatreview.com/rip-software/433-onyx-graphicsannounces-new-onyx->>, 2 pages.
Oracle, “Oracle8i Application Developer's Guide—Advanced Queuing Release 8.1.5”, retrieved on May 6, 2006, available at: <<http://www.cs.umbc.edu/help/oracle8/server.815/a68005/03_adq1i.htm>>, Feb. 1999, 29 Pages.
Oracle.com, “Oracle8i Concepts Release 8.1.5”, retrieved on May 6, 2009, available at: <<http://www.cs.umbc.edu/help/oracle8/server.815/a67781/c16queue.htm>>, Feb. 1999, 10 Pages.
Oracle.com, “Oracle Mobile Sales Assistant User Guide for BlackBerry”, available at: <<http://download.oracle.com/docs/cd/E1254701/books/PDF/MobileSalesAsstBB.pdf.>>, Nov. 2008, 20 pages.
“Oracle8i Application Developer's Guide—Advanced Queuing Release 2 (8.1.6)”, retrieved on May 6, 2009, available at: <<http://www.cs.otago.ac.nz/oradocs/appdev.817/a76938/adq01in5.htm>>, Dec. 1999, 8 pages.
Oryl, “Review: Asus P527 Smartphone for North America”, retrieved on Dec. 17, 2008, available at: <<http://www.mobileburn.com/review.jsp?Id=4257>>, Mar. 5, 2008, 1 Page.
Padilla, “Palm Treo 750 Cell Phone Review—Hardware”, retrieved on Dec. 11, 2008, available at: <<http://www.wirelessinfo.com/content/palm-Treo-750-Cell-Phone-Review/Hardware.htm>>, Mar. 17, 2007, 4 Pages.
“Palette Extender 1.0.2”, Retrieved from <<https://web.archive.org/web/20071202034958/http://palette-extender.en.softonic.com/symbian>>, Jan. 21, 2003, 2 pages.
“Parallax Scrolling”, retrieved on May 5, 2009, available at: <<http://en.wikipedia.org/wiki/Parallax_scrolling>>, May 4, 2009, 3 Pages.
Paul, “Hands-on: KDE 4.5 Launches with Tiling, New Notifications”, retrieved on Sep. 29, 2010, available at: <<http://arstechnica.com/open-source/reviews/2010/08/hands-on-kde-45-launches-withtiling-new-notifications.ars>>, Aug. 2010, 3 pages.
Pendharkar, “Fluid Home Screen for Mobile Phones”, Helsinki Metropolia University of Applied Sciences, Master of Engineering, Information Technology, Thesis, available at: <<http://theseus17-kk.lib.helsinki.fi/bitstream/handle/10024/46481/FinalThesis3.pdf?sequence=1/>>, Apr. 12, 2012, 48 pages.
Perry, “Teach Yourself Windows 95 in 24 Hours”, 1997, Sams Publishing, 2nd Edition, 1997, pp. 193-198.
Phonespell.org, “Dial a number”, retrieved on Nov. 20, 2008, available at: <<http://www.phonespell.org/ialhelp.html>>, Nov. 20, 2008, 1 page.
“Push Notifications Overview for Windows Phone”, Retrieved from <<http://msdn.microsoft.com/en-us/library/ff402558(VS_92,printer).aspx>>, Sep. 3, 2010, 1 Page.
QNX.com, “QNX Photon microGUI Windowing System”, available at: <<http://www.qnx.com/developers/docs/6.5.0/index.jsp?topic=%2Fcom.qnx.doc.photon_prog_guide%2Fdragndrop.html>>, Apr. 6, 2014, 22 pages.
Raghaven, et al., “Model Based Estimation and Verification of Mobile Device Performance”, Available at http://alumni.cs.ucsb.edu/-raimisl/emsoft04_ 12.pdf., Sep. 27-29, 2004, 10 Pages.
Rathbone, “Windows Phone 7 Live Tile Schedules—How to Execute Instant Live Tile Updates”, Retrieved from <http://www.diaryofaninja.com/blog/2011/04/03/windowsphone-7-live-tile-schedules-ndash-executing-instant-live-tile-updates> on Mar. 11, 2013, Apr. 3, 2011, 8 pages.
Ray, “Microsoft Re-Tiles Mobile Platform for Windows 7 Era”, retrieved from <http://www.theregister.co.uk/2010/02/15/windows_phone_7 series/> on May 11, 2011, Feb. 15, 2010, 2 pages.
Reed, “Microsoft Demos Windows Mobile 6.1 at CTIA”, Retrieved from: <http://www.networkworld.com/news/2008/040208-ctia-microsoft-windows-mobile.html> on Jul. 18, 2008, 0410212008, 1 page.
“Remapping the Keyboard”, Retrieved from <<http://publib.boulder.ibm.com/infocenter/hodhelp/v9r0/index.jsp?topic=/com.ibm.hod9.doc/help/assignkey.html>>, Jul. 15, 2005, 5 Pages.
Remond, “Mobile Marketing Solutions”, Retrieved from: <http://www.mobilemarketingmagazine.co.uk/mobile_social_networking/> on May 5, 2009., Apr. 28, 2009, 16 pages.
Rice, et al., “A System for Searching Sound Palettes”, Proceedings of the Eleventh Biennial Symposium on Arts and Technology., Available at <http://www.comparisonics.com/FindSoundsPalettePaper.pdf>,Feb. 2008, 6 pages.
Ritchie, “iOS 4 features: iPod touch Wi-Fi stays connected when asleep—iPhone too?”, Retrieved from: <http://www.goip.com/2010/06/ios-4-features-ipod-touch-wi-fistays-connected-when-asleep-%E2%80%94-iphone-too/> on Sep. 30, 2010, Jun. 14, 2010, 2 pages.
Ritscher, “Using Surface APIs in your WPF application—Part 1”, Retrieved from: <http://blog.wpfwonderland.com/2009/06/30/using-surface-apis-in-your-wpf-application/> on Sep. 28, 2010, Jun. 30, 2009, 7 pages.
Roberts, “Touching and Gesturing on the iPhone”, Available at < http://www. site pen. com/blog/2008/0711 O/touch i ng-an d-gestu ring-on-theiphone/comments-pare-1 >, Jul. 10, 2008, 16 pages.
Restriction Requirement Issued in U.S. Appl. No. 13/118,265, dated Feb. 27, 2014, 7 Pages.
Restriction Requirement Issued in U.S. Appl. No. 13/658,694, dated Feb. 6, 2013, 6 Pages.
Restriction Requirement Issued in U.S. Appl. No. 13/118,288, dated Mar. 4, 2014, 7 Pages.
Sandoval, “A development platform and execution environment for mobile applications”, Universidad Aut6noma de Baja California, School of Chemical Sciences and Engineering, Available at <http://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.86.7989&rep=rep1 &type=pdf>,2004, 18 pages.
Schonfeld, “Windows 8 is Gorgeous, but is it More Than Just a Shell? (Video)”, retrieved on Jun. 20, 2011, available at: <<http://techcrunch.com/2011/06/02/windows-8-gorgeous-shell-video/>>, Jun. 2, 2011, 6 pages.
“SeaMonkey Hangs during Drag Operation”, Retrieved From<<http://forums.mozillazine.org/viewtopic.php?f=5&t=1783735>>, Retrieved Date: Sep. 18, 2011, 2 Pages.
Second Office Action Issued in Chinese Patent Application No. 201180071186.4, dated Jan. 20, 2015, 7 Pages.
“SecureMe-Anti-Theft Security Application for S60 3rd”, Retrieved from <<http:/www.killermobile.com/newsite/mobile-software/s60-applications/secureme-%11-anti%11theft-security-application-for-s60-3rd.htm>>, Dec. 15, 2008, 3 Pages.
Shar, “HTC Shows HTC Snap with Snappy Email Feature”, retrieved on May 5, 2009, available at: <<http://www.wirelessandmobilenews.com/smartphones/>>, Apr. 1, 2009, 10 Pages.
Simeone, “A Cross-Device Drag-and-Drop Technique”, In Proceedings of MUM 2013, Dec. 2013, 4 pages.
Singh, et al., “CINEMA: Columbia InterNet Extensible Multimedia Architecture”, Available at <http://www1.cs.columbia.edu/-library/TR-repository/reports/reports-2002/cucs-011-02.pdf>, Sep. 3, 2002, 83 Pages.
Smith, et al., “GroupBar: The TaskBar Evolved”, Proceedings of OZCHI 2003—Available at <http://research.microsoft.com/pubs/64316/ozchi2003-groupbar.pdf>, Nov. 2003, pp. 1-10.
“Snap—Windows 7 Features”, Retrieved from <<http://web.archive.org/web/20110923164136/http://windows.microsoft.com/en-US/windows7/products/features/snap>>, Retrieved on Sep. 23, 2011, 2 Pages.
Softonic.com, “Palette Extender 1.0.2”, retrieved on Nov. 6, 2008, available at: <<http://paletteextender.en.softonic.com/symbian>>, Jan. 21, 2003, 2 pages.
Spradlin, “Switcher Provides an Incredible Gesture-based App Switching Tool”, Retrieved from <http://www.androidpolice.com/2012/07/09/switcher-proof-of-concepthits-the-play-store-providing-an-incredible-gesture-based-app-switching-tool/> on Feb. 18, 2013, Jul. 9, 2012, 7 pages.
Stebih, “Windows 8 Mouse and Keyboard Commands”, retrieved from <http://www.helpdesktv.ca/windows/windows-8-mouse-and-keyboard-commands.html> on Nov. 23, 2011, Sep. 21, 2011, 5 pages.
Steinicke, et al., “Multi-Touching 3D Data: Towards Direct Interaction in Stereoscopic Display Environments coupled with Mobile Devices”, Advanced Visual Interfaces (AVI) Workshop on Designing Multi-Touch Interaction Techniques for Coupled Public, Available at <http://viscg.uni-muenster.de/publications/2008/SHSK08/ppdworkshop.-pdf.>,Jun. 15, 2008, 4 Pages.
The Supplementary European Search Report dated Jul. 31, 2012 for European Patent Application No. 10823883.3, 5 pages.
Supplemental Notice of Allowance received for U.S. Appl. No. 12/978,184, dated Feb. 25, 2014, 2 pages.
Supplemental Notice of Allowance Issued in U.S. Appl. No. 13/655,390, dated Jul. 25, 2013, 2 Pages.
Supplemental Notice of Allowance received for U.S. Appl. No. 12/977,584, dated Sep. 16, 2013, 2 Pages.
Supplemental Notice of Allowance Issued in U.S. Appl. No. 13/655,390, dated Sep. 19, 2013, 2 pages.
Suror, “PocketShield—New Screenlock App for the HTC Diamond and Pro”, Retrieved from: <http://wmpoweruser.com/?tag=htc-touch-diamond> on Jun. 28, 2011, Oct. 23, 2008, 2 pages.
“Switch between Windows of the Same App”, Retrieved From <<http://www.ntwind.com/software/vistaswitcher/instance-switcher.html>>, May 12, 2012, 2 Pages.
“Symbian Applications”, retrieved on May 5, 2009, available at: <<http://symbianfullversion.blogspot.com/2008_12_01_archive.html>>, Jan. 2009, 51 Pages.
“TaskOS”, Retrieved From<<http://www.appbrain.com/app/taskos/com.profete162.TaskOS>>, Retrieved Date: Sep. 18, 2011, 2 Pages.
TechCredo, “Top 3 Task Switchers for Android”, retrieved on May 11, 2011, available at: <<http://www.techcredo.com/android/top-3-task-switchers-for-android>>, Mar. 9, 2011, 5 pages.
Terpstra, “Beta Beat: Grape, a New Way to Manage Your Desktop Clutter”, Retrieved from: Beta Beat: Grape, a New Way to Manage Your Desktop Clutter on Jun. 28, 2011, Apr. 14, 2009, 4 pages.
Symbian O.S., “The Map Screen”, retrieved on Jun. 17, 2011, available at: <<http://www.symbianos.org/whereamiusersguide>>, 3 pages.
“The Start Menu (overview)”, retrieved from <http://web.archive.org/web/20101219151141/http://www.windows.microsoft.com/en-US/windows-vista/The-Start-menu-overview> on Feb. 12, 2013, Dec. 19, 2010, 5 pages.
“ThumbJot”, available at: <<http://thumbjot.com/wp/?page_id=8>>, 2008, 2 pages.
“Tiles and Notifications for Windows Phone”, Retrieved From<<<http://msdn.microsoft.com/en-us/library/windowsphone/develop/jj662933.aspx>>, Mar. 11, 2014, 2 Pages.
“Top 3 Task Switchers for Androd”, Retrieved from <<http://www.techcredo.com/android/top-3-task-switchers-for-android>>, Mar. 9, 2011, 5 Pages.
“Top Android App: Swipepad”, Retrieved from <<https://web.archive.org/web/20110902110536/http://www.bestandroidappsreview.com/2011/01/top-android-app-swipepad-launcher.html>>, Jan. 19, 2011, 4 Pages.
“Touch Shell Free”, Retrieved from <<http://www.pocketpcfreeware.mobi/download-touch-shell-free.html>>, Feb. 23, 2009, 2 Pages.
“User Guide: WireframeSketcher”, Retrieved from <<http://wireframesketcher.com/help/help.html>>, Retrieved on Jun. 17, 2011, 19 Pages.
Vallerio, et al., “Energy-Efficient Graphical User Interface Design”, Retrieved from: <http://www.cc.gatech .ed u/classes/ A Y2007 /cs 7 4 70 fal l/zhong-energy-efficient-userinterface.pdf>, Jun. 10, 2004, pp. 1-13.
Vermeulen, “BlackBerry PlayBook Hands-on”, retrieved from <http://mybroadband.co.za/news/gadgets/20104-BlackBerry-PlayBook-hands-.html> on May 11, 2011, May 8, 2011, 4 pages.
Viticci, “Growl 1.3 to be Released on Mac App Store, Introduce Lion Support and Drop GrowlMail Support”, Retrieved from: <http://www.macstories.net/stories/growl-1-3-to-be-released-on-mac-app-store-introduce-lion-support-and-drop-growlmail-supporU> on Jul. 22, 2011, Jul. 6, 2011, 6 pages.
Vornberger, “Bluetile”, Retrieved from: <http://www.bluetile.org> on Sep. 29, 2010, 5 pages.
Webster, “VsNotepad: An Improved Way to Jot Down on Windows Mobile”, Retrieved from <http://pocketnow.com/index.php/index.php?a=portal_detail&t=news&id=7424>, May 19, 2009, 2 pages.
Wikipedia, “Multi-touch”, retrieved on Apr. 24, 2009, available at: <<http://en.wikipedia.org/wiki/Multitouch#Microsoft_Surface>>, Apr. 17, 2009, 8 pages.
Wilson, “How the iPhone Works”, Retrieved from: <http://electronics.howstuffworks.com/iphone2.htm> on Apr. 24, 2009, Jan. 2007, 9 pages.
Wilson, “Robust Computer Vision-Based Detection of Pinching for One and Two-Handed Gesture Input”, In Proceedings of UIST 2006—Available at <http://research.microsoft.com/enus/um/people/awilson/publications/wilsonuist2006/uist%202006%20taffi.pdf>, Oct. 2006, 4 pages.
“Windows 8 is Gorgeous, but is it More Than Just a Shell? (Video)”, retrieved from <http://techcrunch.com/2011/06/02/windows-8-gorgeous-shell-video/> on Jun. 20, 2011, Jun. 2, 2011, 6 pages.
“Windows Phone 7 (Push Notification)”, retrieved on Jul. 6, 2011, available at: <<http://unknownerror.net/2011-06/windows-phone-7-push-notification-36520>>, 4 pages.
“Windows Phone 7 Live Tiles”, retrieved on May 11, 2011, available at: <<http://www.knowyourmobile.com/microsofUwindowsphone7/startscreen/640737/windows_phone_7live_tiles.html>>, Oct. 20, 2010, 3 pages.
“Winterface Review”, retrieved on Nov. 12, 2008, available at: <<http://www.mytodayscreen.com/winterfacereview/>>, Jul. 9, 2008, 42 pages.
Wireframesketcher.com, “User Guide”, retrieved on Jun. 17, 2011, available at: <<http://wireframesketcher.com/help/help.html>>, 19 pages.
Wobbrock, et al., “User-Defined Gestures for Surface Computing”, CHI 2009, Apr. 4-9, 2009, Boston, MA available at <http://faculty.washington.edu/wobbrock/pubs/chi-09.2.pdf>, Apr. 4, 2009, 10 pages.
“Womma”, retrieved on May 5, 2009, available at: <<http://www.womma.org/blog/links/wom-trends/>>, 2007, 70 Pages.
“Working with Multiple Windows”, Retrieved from <<https://web.archive.org/web/20090427024713/http://www.msoffice-tutorial.com/working-with-multiple-windows.php>>, Retrieved on Sep. 23, 2011, 3 Pages.
Wu, et al., “Achieving a Superior Ownership Experience in Manageability and Quality for Siebel CRM”, available at <http://www.oracle.com/us/products/enterprisemanager/superior-exp-for-siebel-crm-068962.pdf>, Aug. 2008, 25 pages.
Wyatt, “/Flash/the art of parallax scrolling”, .net Magazine, Aug. 1, 2007, pp. 74-76.
Yahoo.com, “YUI 3: ScrollView [beta]”, retrieved on Sep. 28, 2010, available at: <<http://developer.yahoo.com/yui/3/scrollview/>>, 5 pages.
Yang,, et al., “Dual-Surface Input: Augmenting One-Handed Interaction with Coordinated Front and Behind-the-Screen Input”, In Proceedings of the 11th International Conference on Human-Computer Interaction with Mobile Devices and Services, Sep. 15-18, 2009, 10 Pages.
Yang, et al., “Semantic Photo Album Based on MPEG-4 Compatible Application Format”, Retrieved from: <http://ieeexplore.ieee.org/stamp/stamp.jsp?arnumber=04146254.>, 2007, 2 Pages.
“You've Got Mail 1.4 Build”, retrieved on Jan. 6, 2010, available at: <<http://www.fileshome.com/Shows_Animation_Piays_Sound_Automatic_N . . . >>, Jun. 18, 2007, 2 pages.
Related Publications (1)
Number Date Country
20160210027 A1 Jul 2016 US
Continuations (1)
Number Date Country
Parent 13335001 Dec 2011 US
Child 14980321 US