Content-based snap point

Information

  • Patent Grant
  • 9423951
  • Patent Number
    9,423,951
  • Date Filed
    Friday, December 31, 2010
    14 years ago
  • Date Issued
    Tuesday, August 23, 2016
    8 years ago
Abstract
This document describes content-based snap points and techniques that use these snap points. In some embodiments, multiple content-based snap points are used to stop at points in content that are convenient, prevent overshooting of important parts in the content, and/or aid user's in manipulating and consuming the content.
Description
BACKGROUND

Conventional content interfaces permit users to pan through content, such as a news article or a list of files. These interfaces typically operate in one of two ways. A first way pans based on the amount of content displayable to the user—for each user selection, the interface presents a next display worth of content. This can be thought of as a page-by-page progression, though the page is often based on the amount of content displayable to the user at the time rather than the content having been organized into pages independent of the presentation size. Thus, content presented on a small screen, in a small window, or with a low resolution may require many pages to fully represent. Likewise, large bodies of content may require many pages, even for large screens, large windows, or high resolutions. Not surprisingly, panning through content page-by-page is often slow and annoying for users.


A second way enables a user to vary how far the interface pans through the content based on some factor of the user's selection, such as how fast a user flicks her finger across a touch screen, touch pad, or mouse scroller. This second way permits users to move slowly or quickly through content but may stop at inconvenient points in the content, overshoot important parts of the content, or be difficult to use.


SUMMARY

This document describes content-based snap points and techniques that use these snap points. In some embodiments, multiple content-based snap points are used to stop at points in content that are convenient, prevent overshooting of important parts in the content, and aid user's in manipulating and consuming the content.


This summary is provided to introduce simplified concepts for a content-based snap point that are further described below in the Detailed Description. This summary is not intended to identify essential features of the claimed subject matter, nor is it intended for use in determining the scope of the claimed subject matter.





BRIEF DESCRIPTION OF THE DRAWINGS

Embodiments of techniques and apparatuses for a content-based snap point are described with reference to the following drawings. The same numbers are used throughout the drawings to reference like features and components:



FIG. 1 illustrates an example system in which techniques enabling and/or using a content-based snap point can be implemented.



FIG. 2 illustrates an example method that uses a content-based snap point.



FIG. 3 illustrates an example display having an immersive interface presenting content and through which a content-manipulation input is received.



FIG. 4 illustrates a full representation of content shown in FIG. 3 having content-based snap points and threshold ranges.



FIG. 5 illustrates an example method for determining content-based snap points and/or threshold ranges associated with these content-based snap points.



FIG. 6 illustrates an example method for dynamically altering threshold ranges associated with content-based snap points.



FIG. 7 illustrates dynamically altered threshold ranges relative to threshold ranges illustrated in FIG. 4.



FIG. 8 illustrates an example of content having images, primary content-based snap points, and secondary content-based snap points.



FIG. 9 illustrates an example method using a content-based snap point in an application-launching interface.



FIG. 10 illustrates an example application-launching interface presenting a subset of a full set of selectable labels, the full set of selectable labels, and a later presentation of another subset of the full set of selectable labels.



FIG. 11 illustrates an example device in which techniques enabling and/or using a content-based snap point can be implemented.





DETAILED DESCRIPTION
Overview

This document describes content-based snap points. A content-based snap point can permit users to stop at points in content that are convenient, prevent overshooting of important parts in the content, and aid user's in manipulating and consuming the content.


Consider a case where a user wishes to find a particular part of a news article about the United Kingdom's coldest winter in 100 years, namely the part of the article concerning a severe freeze to the Lake District of England. Assume that this article needs 14 pages to fully present based on a presentation size and resolution and that it includes four subtitled sections, one of which is “Lake District Frozen Solid.” As is often the case, the user in this example does not know exactly where in the article the desired portion begins.


In one of the conventional ways to move through content noted above, the user can select to advance through the article page-by-page. Note that this conventional way will not necessarily present a page that begins with the desired Lake District section (the Lake District section will eventually come up, but may begin at the middle or end of a page). Assume that the Lake District section will be displayed starting in the middle of the seventh page. Therefore, to pan through the article page-by-page to find the Lake District section, a user will see the first page and then select to move to a next page, see that page displayed, look for content about the Lake District throughout the page rather than just the top of the page, select to move to a next page, look for content about the Lake District throughout that page, and so forth six times. Thus, following this first conventional way the user selects to move through the article six times, searches visually through seven pages, and then finds the section about the Lake District at the middle of that seventh page. As is readily apparent even with this relatively short news article, this conventional way in which to move through content can be slow and annoying.


Consider the same case for the second conventional way. Here the user again does not know where in the article the Lake District section begins. Assume that the user makes a low-movement input to begin moving through the article (e.g., a small or slow gesture) in response to which the conventional interface presents part of a next page of content. The user looks over the content, does not see anything on the Lake District in the displayed page, and then selects again with another low-movement selection. In response, the conventional interface presents more content but not quite a full page of content. The user again looks through the presented content without finding the Lake District section. Assume that the user then makes a moderate-movement selection, in response to which the conventional interface moves forward about two pages, thereby skipping roughly a page worth of content. At this point the user may be concerned that she has missed the Lake District section and go backwards, or may carefully study the content to see if it is about the Lake District because, even if she is expecting a subtitle, she does not know if she missed it. Following either of these cases, assume the user now makes a large-movement selection in response to which the conventional interface moves forward five pages, skipping the beginning of the Lake District section. At this point the user can back up, carefully study the content to see if she is somewhere in the Lake District section, or move forward again (likely to the end of the article). Or perhaps the user, now frustrated, searches for a search box, enters text of “Lake District,” see multiple options, select on one or more of these, and eventually find the Lake District section. Or the user could decide to use low-movement selections, possibly taking five, ten, or even fifteen selections and accompanying reviews of these pages to finally find the section. As is readily apparent, this second conventional way can be difficult to use, overshoot or undershoot a desired portion, or be quite slow.


In contrast, consider an example of techniques that use content-based snap points to move through content. As noted above, the news article has four sections each of which has a subtitle. Here the techniques assign four content-based snap points—one for each of the four sections. The article begins with a section about London, then a second section about Wales, a third about the Lake District, and a fourth about Scotland. Here assume that the user makes a moderate- or fast-movement selection in response to which the techniques present the section on Wales, the presented page beginning with the section and having a subtitle stating “Wales Expects Poor Crop Due to Prolonged Winter.” The user quickly ascertains that this is not the Lake District section, knowing that the section subtitle indicates otherwise, and also needing only to look at the top portion of the displayed content. The user then selects again and is presented with the Lake District section, entitled at the top of the portion displayed. The user quickly and easily found the desired portion of the content, here with two selections and two relatively short analyses of displayed content.


This is but one example of techniques using content-based snap points—others are described below including those that skip some content-based snap points, have subordinate content-based snap points, use static and dynamic threshold ranges for snap points, and numerous other aspects of the techniques.


Example System



FIG. 1 illustrates an example system 100 in which techniques using a content-based snap point can be embodied. System 100 includes a computing device 102, which is illustrated with six examples: a laptop computer 104, a tablet computer 106, a smart phone 108, a set-top box 110, a desktop computer 112, and a gaming device 114, though other computing devices and systems, such as servers and netbooks, may also be used.


Computing device 102 includes computer processor(s) 116 and computer-readable storage media 118 (media 118). Media 118 includes an operating system 120, windows-based mode module 122, immersive mode module 124, application-launching module 126, content handler 128, and one or more content 130, each having one or more content-based snap points 132. Computing device 102 also includes or has access to one or more displays 134, four examples of which are illustrated in FIG. 1.


Windows-based mode module 122, immersive mode module 124, and application-launching module 126 may present and use content 130 in different ways and through various content interfaces. Content handler 128, however, can use content-based snap points 132 in one or a combination of these three interfaces to aid in handling content 130, as well as other interfaces that present content. These three modules are described in greater detail below.


Windows-based mode module 122 presents at least some of content 130 through windows having frames. These frames may provide controls through which to interact with an application and/or controls enabling a user to move and size the window.


Immersive mode module 124 provides an environment by which a user may view and interact with content 130. In some embodiments, this environment presents content of, and enables interaction with, applications with little or no window frame and/or without a need for a user to manage a window frame's layout or primacy relative to other windows (e.g., which window is active or up front) or manually size or position content. This environment can be, but is not required to be, hosted and/or surfaced without use of a windows-based desktop environment. Thus, in some cases immersive mode module 124 presents an environment that is not a window (even one without a substantial frame) and precludes usage of desktop-like displays (e.g., a taskbar). Further still, in some embodiments this immersive environment is similar to an operating system in that it is not closeable or capable of being un-installed. While not required, in some cases this immersive mode module 124 enables use of all or nearly all of the pixels of a display for presentation of content 130.


Application-launching module 126 provides an application-launching interface having selectable labels associated with various applications. These labels are one example of content 130. In this context, content-based snap points 132 may reside between groups of these labels, for example.


Content-based snap points 132 are associated with various points in each respective content 130. These snap points 132 can be determined by content handler 128, received from the provider of each content 130, and/or received or determined based on information from one or more of operating system 120, windows-based mode module 122, immersive mode module 124, or application-launching module 126.


These snap points 132 are based on the content 130 to which they are associated, such as to logical breaks or dividers in content (e.g., the four sections having subtitles in the Winter article noted above), ways in which the content is displayed (e.g., a snap point set to prevent displaying half of a still image), an organization of content (e.g., labels or files organized into groups), or an expected use of content (e.g., a snap point based on expectation that a user zooming to see the State of Hawaii likely wants to zoom quickly past a screen of blue ocean). Ways in which snap points 132 are received or determined are set forth in greater detail below.


As noted, content 130 can be presented in at least these three ways, though other modes are also permitted. Operating system 120, modules 122, 124, and 126, as well as content handler 128 can be separate from each other or combined or integrated in some form.


Example Methods


Using a Content-Based Snap Point



FIG. 2 depicts a method 200 using a content-based snap point. This method is shown as a set of blocks that specify operations performed but is not necessarily limited to the order shown for performing the operations by the respective blocks. In portions of the following discussion reference may be made to system 100 of FIG. 1, reference to which is made for example only.


Block 202 receives, during presentation of content, a content-manipulation input having a projected stopping point in the content. This content-manipulation input can be intended to zoom in or out of content to increase or decrease resolution (e.g., a map or photo), pan forward or back through content (e.g., an article or list), and rotate content (e.g., an image) for example.


A content-manipulation input can be one of a set of multiple possible inputs, such as a set of five hot keys. Another possibility is an input that can vary, such as a gesture having a variable velocity or inertia, a press-hold-and-release control or button (which varies based on how long the user holds), a flick of a mouse scroller or movement of a mouse selector (either varying based on velocity and/or orientation), to name just a few. These inputs can be received in many ways, such as through a touch screen, touch pad, mouse ball, and mouse scroller to name just a few.


Consider, by way of example, FIG. 3, which illustrates display 300 and immersive interface 302 presenting content 304. Display 300 is one example of display 134, immersive interface 302 is provided by immersive mode module 124, and content 304 is an example of content 130, all of FIG. 1. A fast-movement flick (a gesture having a high velocity) is received at 306. A factor of this gesture 306 is used in calculating a projected stopping point in content 304, here the magnitude of its velocity. Some interfaces behave differently in response to various inputs, such as by using different friction values for gestures, having different hot keys, or moving through content at different speeds during a hold of a press-hold-and-release, to name a few. In each of these cases, however, a projected stopping point in the content can be calculated. Here immersive mode module 124 receives gesture 306 through a touch screen, calculates a projected stopping point, and provides this to content handler 128.


Block 204 determines whether the projected stopping point in the content is within a threshold range of a content-based snap point. Each content-based snap point 132 of content 130 includes a threshold range. This threshold range can be centered around the snap point or not centered, and can be received or determined by content handler 128. In some cases the ranges vary based on various factors, including a current or prior content-manipulation input, to name just two. In the ongoing example, however, a relatively simple case of fixed threshold ranges is contemplated. Other examples consider variable thresholds below.


Continuing the ongoing example, consider FIG. 4, which illustrates a full representation of content 402 of which content 304 of FIG. 3 is only a part. The full content 402 includes four content-based snap points 404, 406, 408, and 410, which are assumed here to have been received from the provider of the news article, here a “News” website. Here assume that snap point 404 is at a first subtitle in the article, snap point 406 is at an advertisement within the article, snap point 408 is at a second subtitle, and snap point 410 at a conclusion section. This example also assumes that the threshold ranges shown 412, 414, 416, and 418 are also received by content handler 128 from the News website and do not vary.


If block 204 determines that the projected stopping point is within a threshold range for a content-based snap point, method 200 proceeds along a “Yes” path to block 206. If not, method 200 proceeds along a “No” path to block 208.


In this example, block 204, through content handler 128, determines that gesture 306 has a projected stopping point within a threshold range of a snap point, here more than halfway through the content, shown at projected stopping point 420 in FIG. 4 and within threshold range 416 associated with content-based snap point 408.


Note here that content handler 128 skips over threshold ranges 412 and 414 for snap points 404 and 406, respectively. In some other cases, however, a threshold range can be used that effectively hard-stops movement at a snap point, such as by setting the range to include all of the content that is past the snap point.


Block 206, responsive to determining that the projected stopping point in the content is within the threshold range of the snap point, causes the content to be presented at the snap point. In this example content handler 128 causes presentation of content of the news article at the snap point by passing snap point 408 as input to immersive mode module 124 of FIG. 1, responsive to which immersive mode module 124 moves through the “Satellite May Crash to Earth” article to a second subtitled section in immersive interface 302 (not shown).


Thus, instead of presenting content based on projected stopping point 420 of FIG. 4, content is presented based on content-based snap point 408. Similarly to the United Kingdom winter weather article noted above, these techniques permit a user to more quickly and easily move through content using these snap points, here to skip a first subtitled section and an advertisement to start at a second subtitled section. If the user did not desire this section, however, a fast or moderate-speed gesture will likely result in presentation of one of the other sections or the advertisement. As is readily apparent, the techniques permit the user to quickly find the various content-based sections of this news article.


In some embodiments, content handler 128 visually indicates that the content is being presented at a snap point and/or that snap points have been passed over. Content handler 128, for example, may indicate this through a temporary slow-down in a scan of content shown in immersive interface 302 at content of snap points 404 and 406 on the way to stopping at snap point 408 or projected stopping point 420. Content handler 128 may also or instead indicate information when stopping at content-based snap point 420, such as with a visual “bounce,” which temporarily shows some content (e.g., a ¼ of a page) that follows the snap point before bouncing back to settle at the content of snap point 408 or shows that there is no content (e.g., a ½ of a page of white space) before bouncing back. Other visual indicators can optionally be used, such as to show that a user is at an end of the content with a hard stop (a visual rattle or shake of the content presented, etc.). These are but a few of the possible indicators that may further enable users to move through content more easily by providing information to the user.


Also in some embodiments, content handler 128 may alter an input to change the projected stopping point so that it equals the snap point. Consider the case where the projected stopping point is within a threshold range and thus, content handler 128 will cause the content to be presented at the snap point instead of the projected stopping point. Content handler 128 may alter the content-manipulation input or how the stopping point is calculated so that presenting the content at the snap point appears natural or expected. One way to do so is to calculate a new factor for an input that would result in a new projected stopping point equaling the snap point. Thus, a new velocity for a received gesture is calculated and passed back to one of operating system 120 or modules 122, 124, or 126 effective to cause the new projected stopping point to equal the desired stopping point—namely the snap point. Another way is to alter an algorithm used to calculate the projected stopping point, such as to alter a friction of a velocity-based gesture, which can also be passed to one of operating system 120 or modules 122, 124, or 126. By so doing, the presentation of content at the snap point appears more natural or expected, rather than looking like an artificial additional movement or jarring stop.


Block 208, responsive to determining that the projected stopping point in the content is not within the threshold range of the snap point, causes the content to be presented at the projected stopping point. Assume, for example, that projected stopping point 420 is instead just outside threshold range 416. In such a case, content handler 128 causes the content to be presented at the projected stopping point. Here assume that content handler 128 does so either by passing back the projected stopping point 420 to immersive mode module 124 or by indicating that no change in presentation is to be made.


The examples described above in relation to method 200 assume threshold ranges that are received from a content provider and do not change. Numerous alternatives are contemplated by the techniques, some examples of which are described below.


Determining a Content-Based Snap Point and/or Threshold Ranges



FIG. 5 depicts a method 500 for determining content-based snap points and/or threshold ranges associated with these content-based snap points. This method is shown as a set of blocks that specify operations performed, but is not necessarily limited to the order shown for performing the operations by the respective blocks. Method 500 may be used with, instead of, or in addition to one or more of the blocks of method 200. In portions of the following discussion reference may be made to system 100 of FIG. 1, reference to which is made for example only.


Block 502 receives or determines information about content. This information can include an organization of or logical breaks or discontinuities in the content, such as white space, subtitles, or images in text-based content, organizational changes in a file list, such as groupings, file types, or names changes, low-value and high-value content, such as ocean around an island for map content, to name just a few.


Block 504 determines a content-based snap point based on the information about the content. Thus, content handler 128 can set content-based snap points at an island-size resolution to extend or stop a zooming input from a world-sized resolution or at each new first-later alphabetical change to a long list of files (e.g., at “A”, “B”, and so forth to “Z”), to name just a few.


This information received may also be used to determine threshold ranges, even in cases where the snap points are received rather than determined. Block 506, which optionally proceeds from either of block 502 or block 504, determines a threshold range for a content-based snap point based on information about the content. The information received may also include information about the snap points as well as the other information noted above, such as the distance (in content) between adjacent snap points. Here content handler 128 determines threshold ranges for each snap point based at least in part on the size of the content and the distance between the snap points.


Consider again the example of FIG. 4, which illustrates four snap points 404, 406, 408, and 410, a distance between the snap points, and a total size of the content. As shown, threshold range 416 is large and has a wider threshold range relative to snap point 406 than snap point 410 because snap point 406 is further away. Whether or not threshold ranges are received or determined, content handler 128 may alter these threshold ranges based on received input(s), which is described below.


Dynamically Altering a Threshold Range



FIG. 6 illustrates a method 600 for dynamically altering threshold ranges associated with content-based snap points. This method is shown as a set of blocks that specify operations performed, but is not necessarily limited to the order shown for performing the operations by the respective blocks. Method 600 may be used with, instead of, or in addition to one or more of the blocks of method 200 and/or 500. In portions of the following discussion reference may be made to system 100 of FIG. 1, reference to which is made for example only.


Block 602 receives one or more content-manipulation inputs. Block 602 may do so similarly to block 202 of FIG. 2, though block 602 may receive numerous inputs some or all of which may be received prior to presentation of the content, such as when the method uses, at least in part, previously received inputs representing a user's history.


Block 604 alters, based on one or more received content-manipulation inputs, a threshold range associated with a content-based snap point.


Consider again the example of FIG. 4. Here assume that threshold ranges 412, 414, 416, and 418 are received from a content provider. Assume also that the current content-manipulation input is gesture 306 of FIG. 3, which has a high velocity. As such, it likely indicates a user's desire to move quickly to find major content-based breaks in the content rather than a desire to move to a particular line or paragraph. Here content handler 128 alters the four threshold ranges by increasing their size based on gesture 306 having a high velocity. By so doing, the current input is more likely to result in stopping at a snap point.


Here content handler 128 expands the thresholds as shown in FIG. 7, thereby requiring the current input to result in snapping to a snap point or hitting an end of the article but not stopping at a projected stopping point (unless it happens to stop at exactly one of those points). The threshold ranges 414, 414, 416, and 418 are shown as contrast to altered threshold ranges 702, 704, 706, and 708. Note that these altered threshold ranges now cover all of the full representation of content 402.


Continuing this example, assume that the altered thresholds resulted in the content being presented at snap point 406 according to blocks 204 and 206 of method 200. Following presentation of the content of snap point 406 (the advertisement mentioned above), assume that another input is received at block 602, here a low-velocity gesture (or some other high-resolution input). In response, content handler 128 at block 604 reduces the thresholds to reflect that the user is likely wishing to move more slowly or precisely through the news article. These reduced threshold ranges are also shown in FIG. 7, at 710, 712, 714, and 716. Note that with these reduced threshold ranges the new input received will likely not land on a snap point, which is sometimes desirable when a user wishes to slowly or carefully manipulate content. Even for larger threshold ranges, a user's input may not cause the content to be snapped to a snap point (e.g., in many cases a projected stopping point will not fall within one of the threshold ranges 414, 414, 416, and 418 illustrated in FIG. 4 for content 402).


Primary and Secondary Content-Based Snap Points


In some embodiments threshold ranges for various snap points may overlap. In such a case the techniques determine which snap point to use. Consider, for example, image content 800 illustrated in FIG. 8, which has over 100 images (represented by small squares) and four primary content-based snap points 802, 804, 806, and 808 each corresponding to organizational breaks in image content 800, and having primary threshold ranges 810, 812, 814, and 816, respectively. Other, secondary content-based snap points are shown with dashed lines (collectively snap points 818) each having same secondary threshold ranges 820. These secondary threshold ranges 820 are effective to prevent portions of images to be “chopped” off or only partially presented. As noted above, any of these snap points or their threshold ranges can received or determined, including dynamically based on one or more inputs.


Consider the case where content presented is shown in window 822 by windows-based mode module 122 of FIG. 1. Assume also that a content-manipulation input is received having a projected stopping point of 824. Content movement progresses horizontally and shown at projected progress arrow 826 with window 822 shown in dashed lines represented the projected stopping point 824 (shown repeated and separate from image content 800 for visual clarity). At this point content handler 128 determines that projected stopping point 824 lands within two threshold ranges, here primary threshold range 814 and secondary threshold range 820-1, respectively associated with primary content-based snap point 806 and secondary content-based snap point 818-1. Content handler 128 determines, based on a hierarchy of threshold ranges, to present window 822 at snap point 806. This difference from the projected stopping point 824 and the actual presentation point of content-based snap point 806 is shown with reverse progress arrow 828. Note that content handler 128 may optionally show content for the projected stopping point 824 temporarily and then show a visual reversal (a “snap back”) to content-based snap point 806.


While not shown, the techniques may determine between numerous hierarchies of content-based snap points, though in this example only primary and secondary snap points are illustrated.


If, however, a projected stopping point is not within one of primary threshold ranges 810, 812, 814, and 816 but is within one of secondary threshold ranges 820 (which is necessarily the case for content 800 due to complete coverage by these secondary threshold ranges), content handler 128 causes presentation at the appropriate secondary content-based snap point.


Content-Based Snap Points in an Application-Launching Interface



FIG. 9 depicts a method 900 using a content-based snap point in an application-launching interface. This method is shown as a set of blocks that specify operations performed but is not necessarily limited to the order shown for performing the operations by the respective blocks. In portions of the following discussion reference may be made to system 100 of FIG. 1, reference to which is made for example only. Method 900 may be used with, instead of, or in addition to one or more of the methods 200, 500, or 600 or blocks thereof. In portions of the following discussion reference may be made to system 100 of FIG. 1, reference to which is made for example only.


Block 902 presents, in an application-launching interface, selectable labels associated with applications, the presented selectable labels a subset of a full set of selectable labels available through the application-launching interface. These selectable labels are another example of content capable of being handled with content-based snap points.


Consider FIG. 10, which illustrates an example application-launching interface 1000 presented by application-launching module 126 of FIG. 1. Interface 1000 presents eight selectable labels 1002 associated with eight applications. These eight labels 1002 are a subset of a full set of selectable labels shown at 1004.


Block 904 receives, through the application-launching interface, a content-manipulation input to present additional selectable labels of the full set of selectable labels, the content-manipulation input having a projected stopping point.


Here assume that the content-manipulation input has a projected stopping point 1006 shown within the full set of selectable labels 1004. As shown, one content-based snap point 1008 is illustrated, here having a threshold range of 1010.


Block 906 determines whether the projected stopping point for the content-manipulation input in the full set of selectable labels is within a threshold range of a content-based snap point. If block 906 determines that the projected stopping point is within the threshold range, method 900 proceeds along a “Yes” path to block 908. If not, method 900 proceeds along a “No” path to block 910.


Block 908 presents additional, selectable labels of the full set of selectable labels associated with and at the content-based snap point. As shown in the illustrated example, projected stopping point 1006 is within threshold range 1010. In response, content handler 128 passes the content-based snap point 1008 to application-launching interface 126, which presents the interface 1000 having six additional selectable labels, shown at 1012 (three small and three large selectable labels shown).


Block 910 presents additional, selectable labels at the projected stopping point if it is determined not to be within a threshold range (e.g., the end of the full set if the projected stopping point is farther than the threshold range 110).


As noted above, content-based snap points can be received or determined in numerous manners. For this example assume that a user selected, through application-launching interface 1000, to group selectable labels shown at 1002 and 1012 (the first 15 labels of the 21 total labels shown at 1004). Based on this grouping selection, content handler 128 and/or application-launching module 126 set content-based snap point 1008 at its location and having associated threshold range 1010. By so doing, a user can easily select to see the remainder of this group of labels on most content-manipulation inputs. A high velocity selection, however, would likely have a projected stopping point outside threshold range 1010, thereby causing content handler 128 to move to the end of the full set of labels 1004. Thus, a content-based snap point can enable a user to determine how an interface responds to content-manipulation inputs. In this example all but high velocity, low resolution, or otherwise large inputs result in seeing the remainder of the labels in the user's group.


A content-based snap point can be moved or adjusted by content handler 128 based on various factors, including changes to content or by user selection. Consider again application-launching interface 1000. Here assume a user installs two new applications. Selectable labels for these can be added to the full set 1004 by application-launching module 126 to the group that currently has fifteen selectable labels (say at the end of these labels but before the last six labels). Content handler 128 may adjust content-based snap point 1008 or it may be set for adjustment automatically based on the group size expanding. Thus, content-based snap points can change over time based on content changing. In this example the new selectable labels will also be shown (which likely pushes off all of the labels shown at 1002 responsive to the content-manipulation input).


The preceding discussion describes methods relating to content-based snap points. Aspects of these methods may be implemented in hardware (e.g., fixed logic circuitry), firmware, software, manual processing, or any combination thereof. A software implementation represents program code that performs specified tasks when executed by a computer processor. The example methods may be described in the general context of computer-executable instructions, which can include software, applications, routines, programs, objects, components, data structures, procedures, modules, functions, and the like. The program code can be stored in one or more computer-readable memory devices, both local and/or remote to a computer processor. The methods may also be practiced in a distributed computing mode by multiple computing devices. Further, the features described herein are platform-independent and can be implemented on a variety of computing platforms having a variety of processors.


These techniques may be embodied on one or more of the entities shown in environment 100 of FIG. 1 and/or example device 1100 described below, which may be further divided, combined, and so on. Thus, environment 100 and/or device 1100 illustrate some of many possible systems or apparatuses capable of employing the described techniques. The entities of environment 100 and/or device 1100 generally represent software, firmware, hardware, whole devices or networks, or a combination thereof. In the case of a software implementation, for instance, the entities (e.g., modules 122, 124, and 126 and content handler 128) represent program code that performs specified tasks when executed on a processor (e.g., processor(s) 116). The program code can be stored in one or more computer-readable memory devices, such as computer-readable storage media 118 or computer-readable media 1114 of FIG. 11. The features and techniques described herein are platform-independent, meaning that they may be implemented on a variety of commercial computing platforms having a variety of processors.


Example Device



FIG. 11 illustrates various components of example device 1100 that can be implemented as any type of client, server, and/or computing device as described with reference to the previous FIGS. 1-10 to implement techniques that use a content-based snap point. In embodiments, device 1100 can be implemented as one or a combination of a wired and/or wireless device, as a form of television client device (e.g., television set-top box, digital video recorder (DVR), etc.), consumer device, computer device, server device, portable computer device, user device, communication device, video processing and/or rendering device, appliance device, gaming device, electronic device, and/or as another type of device. Device 1100 may also be associated with a user (e.g., a person) and/or an entity that operates the device such that a device describes logical devices that include users, software, firmware, and/or a combination of devices.


Device 1100 includes communication devices 1102 that enable wired and/or wireless communication of device data 1104 (e.g., received data, data that is being received, data scheduled for broadcast, data packets of the data, etc.). The device data 1104 or other device content can include configuration settings of the device, media content stored on the device, and/or information associated with a user of the device. Media content stored on device 1100 can include any type of audio, video, and/or image data. Device 1100 includes one or more data inputs 1106 via which any type of data, media content, and/or inputs can be received, such as user-selectable inputs, messages, music, television media content, recorded video content, and any other type of audio, video, and/or image data received from any content and/or data source.


Device 1100 also includes communication interfaces 1108, which can be implemented as any one or more of a serial and/or parallel interface, a wireless interface, any type of network interface, a modem, and as any other type of communication interface. The communication interfaces 1108 provide a connection and/or communication links between device 1100 and a communication network by which other electronic, computing, and communication devices communicate data with device 1100.


Device 1100 includes one or more processors 1110 (e.g., any of microprocessors, controllers, and the like), which process various computer-executable instructions to control the operation of device 1100 and to enable techniques enabling and/or using a content-based snap point. Alternatively or in addition, device 1100 can be implemented with any one or combination of hardware, firmware, or fixed logic circuitry that is implemented in connection with processing and control circuits which are generally identified at 1112. Although not shown, device 1100 can include a system bus or data transfer system that couples the various components within the device. A system bus can include any one or combination of different bus structures, such as a memory bus or memory controller, a peripheral bus, a universal serial bus, and/or a processor or local bus that utilizes any of a variety of bus architectures.


Device 1100 also includes computer-readable storage media 1114, such as one or more memory devices that enable persistent and/or non-transitory data storage (i.e., in contrast to mere signal transmission), examples of which include random access memory (RAM), non-volatile memory (e.g., any one or more of a read-only memory (ROM), flash memory, EPROM, EEPROM, etc.), and a disk storage device. A disk storage device may be implemented as any type of magnetic or optical storage device, such as a hard disk drive, a recordable and/or rewriteable compact disc (CD), any type of a digital versatile disc (DVD), and the like. Device 1100 can also include a mass storage media device 1116.


Computer-readable storage media 1114 provides data storage mechanisms to store the device data 1104, as well as various device applications 1118 and any other types of information and/or data related to operational aspects of device 1100. For example, an operating system 1120 can be maintained as a computer application with the computer-readable storage media 1114 and executed on processors 1110. The device applications 1118 may include a device manager, such as any form of a control application, software application, signal-processing and control module, code that is native to a particular device, a hardware abstraction layer for a particular device, and so on.


The device applications 1118 also include any system components or modules to implement techniques using or enabling a content-based snap point. In this example, the device applications 1118 can include windows-based mode module 122, immersive mode module 124, application-launching module 126, and content handler 128.


CONCLUSION

Although embodiments of techniques and apparatuses enabling and using a content-based snap point have been described in language specific to features and/or methods, it is to be understood that the subject of the appended claims is not necessarily limited to the specific features or methods described. Rather, the specific features and methods are disclosed as example implementations for a content-based snap point.

Claims
  • 1. A computer-implemented method comprising: receiving, during presentation of content, a content-manipulation input having a projected stopping point in the content, the projected stopping point is based on a velocity or an inertia and a friction of the content-manipulation input;responsive to receiving the content-manipulation input, altering a threshold range of a primary snap point based at least in part on a factor of the content-manipulation input, the factor of the content-manipulation input including one or more of the velocity or the inertia and affecting movement through the content, the threshold range of the primary snap point having a first distance before or after the primary snap point, and the primary snap point having a primary hierarchy value;determining, after altering the threshold range of the primary snap point, whether the projected stopping point in the content is within both the threshold range of the primary snap point and a threshold range of a secondary snap point, the secondary snap point having a secondary hierarchy value; andresponsive to determining that the projected stopping point in the content is within both the threshold range of the primary snap point and the threshold range of the secondary snap point, causing the content to be presented at the primary snap point based on the primary snap point having the primary hierarchy value, and altering one or more of the velocity, the inertia, or the friction of the content manipulation input.
  • 2. A computer-implemented method as described in claim 1, wherein the factor of the content-manipulation input includes a variable velocity or a variable inertia on which the projected stopping point is based.
  • 3. A computer-implemented method as described in claim 1, further comprising visually indicating that additional content is or is not available following the primary snap point.
  • 4. A computer-implemented method as described in claim 3, wherein visually indicating indicates that additional content is available by temporally presenting a portion of the additional content prior to causing the content to be presented at the primary snap point.
  • 5. A computer-implemented method as described in claim 1, further comprising receiving, prior to receiving the content-manipulation input, the primary snap point from a provider of the content.
  • 6. A computer-implemented method as described in claim 1, further comprising determining, prior to receiving the content-manipulation input, the primary snap point.
  • 7. A computer-implemented method as described in claim 6, wherein determining the primary snap point is based on a logical break or divider in the content, an organization of the content, or an expected use of the content.
  • 8. A computer-implemented method as described in claim 1, wherein causing the content to be presented at the primary snap point zooms in or out of the content.
  • 9. A computer-implemented method as described in claim 1, wherein causing the content to be presented at the primary snap point rotates the content.
  • 10. A computer-implemented method as described in claim 1, wherein the primary snap point acts as a hard stop, the threshold range of the primary snap point extending to an end of the content effective to cause the content-manipulation input to be within the threshold range.
  • 11. A computer-implemented method as described in claim 1, further comprising: scanning past the secondary snap point prior to causing the content to be presented at the primary snap point or causing the content to be presented at the projected stopping point; andaltering a velocity associated with the scanning while passing the adjacent secondary snap point.
  • 12. A computer-implemented method as described in claim 1, further comprising determining, prior to receiving the content-manipulation input, the secondary snap point.
  • 13. A computer-implemented method as described in claim 1, further comprising receiving, prior to receiving the content-manipulation input, the secondary snap point from a provider of the content.
  • 14. A computer-implemented method comprising: presenting, in an application-launching interface, selectable labels associated with applications, the presented selectable labels a subset of a full set of selectable labels available through the application-launching interface;determining, based on the organization of information contained in the selectable labels associated with applications, one or more snap points being positioned based on the organization of information contained in the selectable labels;determining a threshold range of a primary snap point of the one or more snap points based on a distance between the primary snap point and an adjacent secondary snap point, the primary snap point having a primary hierarchy value and the second snap point having a secondary hierarchy value, the primary snap point based on a grouping within the full set of selectable labels and being changeable responsive to the grouping changing size, the grouping selected through prior user interaction with the application-launching interface;receiving, through the application-launching interface, a content-manipulation input to present additional selectable labels of the full set of selectable labels and having a projected stopping point within the full set of selectable labels;responsive to receiving the content-manipulation input, altering the threshold range of the primary snap point based at least in part on a velocity of the content-manipulation input; andresponsive to determining that the projected stopping point in the full set of selectable labels is within both the threshold range of the primary snap point and a threshold range of the secondary snap point, presenting, in the application-launching interface, additional selectable labels of the full set of selectable labels associated with and at the primary snap point based on the primary snap point having a primary hierarchy value.
  • 15. A computer-implemented method as described in claim 14, wherein the velocity of the content-manipulation input includes a magnitude of a velocity and causing the content to be presented at the primary snap point alters the velocity associated with the content-manipulation input.
  • 16. A computing device comprising: one or more computer processors; andone or more computer-readable storage media having instructions stored thereon that, responsive to execution by the one or more computer processors, perform operations comprising: receiving content;determining, based on the organization of information contained in the content, one or more content-based snap points positioned based on the organization of information contained in the content;determining a threshold range of a content-based snap point of the one or more content-based snap points based on both a distance between the content-based snap point and an adjacent snap point and on a size of information contained in the content associated with the content-based snap point;receiving, during presentation of the content, a content-manipulation input having a projected stopping point in the content determined based on a velocity or an inertia and a friction of the content-manipulation input;responsive to receiving the content-manipulation input, altering the threshold range of the content-based snap point based at least in part on a factor of the content-manipulation input, the factor of the content-manipulation input including one or more of the velocity or the inertia and affecting movement through the content;determining, after altering the threshold range, whether the projected stopping point in the content is within the threshold range of the content-based snap point; andresponsive to determining that the projected stopping point in the content is within the threshold range of the content-based snap point, causing the content to be presented at the content-based snap point and altering one or more of the velocity, the inertia, or the friction of the content-manipulation input; orresponsive to determining that the projected stopping point in the content is not within the threshold range of the content-based snap point, causing the content to be presented at the projected stopping point.
  • 17. A computing device as described in claim 16, wherein the threshold range includes a distance before or after the content-based snap point.
  • 18. A computing device as described in claim 16, wherein the projected stopping point is within the threshold range of the content-based snap point, and wherein the operations further comprise visually indicating whether additional content is or is not available following the content-based snap point.
  • 19. A computing device as described in claim 16, further comprising determining, prior to receiving the content-manipulation input, the content-based snap point.
  • 20. A computing device as described in claim 16, further comprising receiving, prior to receiving the content-manipulation input, the content-based snap point from a provider of the content.
US Referenced Citations (866)
Number Name Date Kind
4823283 Diehm et al. Apr 1989 A
5045997 Watanabe Sep 1991 A
5046001 Barker et al. Sep 1991 A
5189732 Kondo Feb 1993 A
5258748 Jones Nov 1993 A
5297032 Trojan et al. Mar 1994 A
5321750 Nadan Jun 1994 A
5339392 Risberg et al. Aug 1994 A
5432932 Chen et al. Jul 1995 A
5463725 Henckel et al. Oct 1995 A
5485197 Hoarty Jan 1996 A
5495566 Kwatinetz Feb 1996 A
5506951 Ishikawa Apr 1996 A
5510808 Cina et al. Apr 1996 A
5515495 Ikemoto May 1996 A
5574836 Broemmelsiek Nov 1996 A
5598523 Fujita Jan 1997 A
5611060 Belfiore et al. Mar 1997 A
5623613 Rowe et al. Apr 1997 A
5640176 Mundt et al. Jun 1997 A
5650827 Tsumori et al. Jul 1997 A
5657049 Ludolph et al. Aug 1997 A
5659693 Hansen Aug 1997 A
5675329 Barker Oct 1997 A
5680562 Conrad et al. Oct 1997 A
5687331 Volk et al. Nov 1997 A
5712995 Cohn Jan 1998 A
5740389 Li et al. Apr 1998 A
5771042 Santos-Gomez Jun 1998 A
5793415 Gregory et al. Aug 1998 A
5819284 Farber et al. Oct 1998 A
5844547 Minakuchi et al. Dec 1998 A
5847706 Kingsley Dec 1998 A
5860073 Ferrel et al. Jan 1999 A
5905492 Straub et al. May 1999 A
5914720 Maples et al. Jun 1999 A
5940076 Sommers et al. Aug 1999 A
5959621 Nawaz et al. Sep 1999 A
5963204 Ikeda et al. Oct 1999 A
6008809 Brooks Dec 1999 A
6008816 Eisler Dec 1999 A
6009519 Jones et al. Dec 1999 A
6011542 Durrani et al. Jan 2000 A
6028600 Rosin et al. Feb 2000 A
6057839 Advani et al. May 2000 A
6064383 Skelly May 2000 A
6104418 Tanaka et al. Aug 2000 A
6108003 Hall, Jr. et al. Aug 2000 A
6111585 Choi Aug 2000 A
6115040 Bladow et al. Sep 2000 A
6163749 McDonough et al. Dec 2000 A
6166736 Hugh Dec 2000 A
6188405 Czerwinski et al. Feb 2001 B1
6211921 Cherian et al. Apr 2001 B1
6212564 Harter et al. Apr 2001 B1
6216141 Straub et al. Apr 2001 B1
6266098 Cove et al. Jul 2001 B1
6278448 Brown et al. Aug 2001 B1
6281940 Sciammarella Aug 2001 B1
6311058 Wecker et al. Oct 2001 B1
6313854 Gibson Nov 2001 B1
6369837 Schirmer Apr 2002 B1
6385630 Ejerhed May 2002 B1
6389386 Hetherington et al. May 2002 B1
6396963 Shaffer May 2002 B2
6411307 Rosin et al. Jun 2002 B1
6424338 Andersone Jul 2002 B1
6426753 Migdal Jul 2002 B1
6433789 Rosman Aug 2002 B1
6448987 Easty et al. Sep 2002 B1
6449638 Wecker et al. Sep 2002 B1
6456334 Duhault Sep 2002 B1
6489977 Sone Dec 2002 B2
6493002 Christensen Dec 2002 B1
6505243 Lortz Jan 2003 B1
6507643 Groner Jan 2003 B1
6510144 Dommety et al. Jan 2003 B1
6510466 Cox et al. Jan 2003 B1
6510553 Hazra Jan 2003 B1
6538635 Ringot Mar 2003 B1
6570597 Seki et al. May 2003 B1
6577323 Jamieson et al. Jun 2003 B1
6577350 Proehl et al. Jun 2003 B1
6591244 Jim et al. Jul 2003 B2
6597374 Baker et al. Jul 2003 B1
6628309 Dodson et al. Sep 2003 B1
6636246 Gallo et al. Oct 2003 B1
6662023 Helle Dec 2003 B1
6675387 Boucher et al. Jan 2004 B1
6690387 Zimmerman et al. Feb 2004 B2
6697825 Underwood et al. Feb 2004 B1
6707449 Hinckley et al. Mar 2004 B2
6710771 Yamaguchi et al. Mar 2004 B1
6721958 Dureau Apr 2004 B1
6724403 Santoro et al. Apr 2004 B1
6784925 Tomat Aug 2004 B1
6798421 Baldwin Sep 2004 B2
6801203 Hussain Oct 2004 B1
6807558 Hassett et al. Oct 2004 B1
6832355 Duperrouzel et al. Dec 2004 B1
6857104 Cahn Feb 2005 B1
6865297 Loui Mar 2005 B2
6873329 Cohen et al. Mar 2005 B2
6876312 Yu Apr 2005 B2
6885974 Holle Apr 2005 B2
6904597 Jin Jun 2005 B2
6920445 Bae Jul 2005 B2
6938101 Hayes et al. Aug 2005 B2
6961731 Holbrook Nov 2005 B2
6971067 Karson et al. Nov 2005 B1
6972776 Davis et al. Dec 2005 B2
6975306 Hinckley et al. Dec 2005 B2
6976210 Silva et al. Dec 2005 B1
6978303 McCreesh et al. Dec 2005 B1
6983310 Rouse Jan 2006 B2
6987991 Nelson Jan 2006 B2
7013041 Miyamoto Mar 2006 B2
7017119 Johnston et al. Mar 2006 B1
7019757 Brown et al. Mar 2006 B2
7028264 Santoro et al. Apr 2006 B2
7032187 Keely, Jr. et al. Apr 2006 B2
7036090 Nguyen Apr 2006 B1
7036091 Nguyen Apr 2006 B1
7042460 Hussain et al. May 2006 B2
7051291 Sciammarella et al. May 2006 B2
7058955 Porkka Jun 2006 B2
7065385 Jarrad et al. Jun 2006 B2
7065386 Smethers Jun 2006 B1
7075535 Aguera y Arcas Jul 2006 B2
7089507 Lection et al. Aug 2006 B2
7091998 Miller-Smith Aug 2006 B2
7093201 Duarte Aug 2006 B2
7106349 Baar et al. Sep 2006 B2
7111044 Lee Sep 2006 B2
7133707 Rak Nov 2006 B1
7133859 Wong Nov 2006 B1
7139800 Bellotti et al. Nov 2006 B2
7146573 Brown et al. Dec 2006 B2
7155729 Andrew et al. Dec 2006 B1
7158123 Myers et al. Jan 2007 B2
7158135 Santodomingo et al. Jan 2007 B2
7178111 Glein et al. Feb 2007 B2
7180527 Sakai et al. Feb 2007 B2
7194506 White et al. Mar 2007 B1
7210099 Rohrabaugh et al. Apr 2007 B2
7216588 Suess May 2007 B2
7249326 Stoakley et al. Jul 2007 B2
7262775 Calkins et al. Aug 2007 B2
7263668 Lentz Aug 2007 B1
7277924 Wichmann et al. Oct 2007 B1
7280097 Chen Oct 2007 B2
7283620 Adamczyk Oct 2007 B2
7289806 Morris et al. Oct 2007 B2
7293244 Randall Nov 2007 B2
7296184 Derks et al. Nov 2007 B2
7296242 Agata et al. Nov 2007 B2
7310100 Hussain Dec 2007 B2
7333092 Zadesky et al. Feb 2008 B2
7333120 Venolia Feb 2008 B2
7336263 Valikangas Feb 2008 B2
7369647 Gao et al. May 2008 B2
7376907 Santoro et al. May 2008 B2
7386807 Cummins et al. Jun 2008 B2
7388578 Tao Jun 2008 B2
7403191 Sinclair Jul 2008 B2
7408538 Hinckley et al. Aug 2008 B2
7409646 Vedbrat et al. Aug 2008 B2
7412663 Lindsay et al. Aug 2008 B2
7433920 Blagsvedt et al. Oct 2008 B2
7447520 Scott Nov 2008 B2
7461151 Colson et al. Dec 2008 B2
7469380 Wessling et al. Dec 2008 B2
7469381 Ording Dec 2008 B2
7478326 Holecek et al. Jan 2009 B2
7479949 Jobs Jan 2009 B2
7480870 Anzures Jan 2009 B2
7483418 Maurer Jan 2009 B2
7487467 Kawahara et al. Feb 2009 B1
7496830 Rubin Feb 2009 B2
7500175 Colle et al. Mar 2009 B2
7512966 Lyons, Jr. et al. Mar 2009 B2
7577918 Lindsay Aug 2009 B2
7581034 Polivy et al. Aug 2009 B2
7593995 He et al. Sep 2009 B1
7595810 Louch Sep 2009 B2
7599790 Rasmussen et al. Oct 2009 B2
7600189 Fujisawa Oct 2009 B2
7600234 Dobrowski et al. Oct 2009 B2
7606714 Williams et al. Oct 2009 B2
7607106 Ernst et al. Oct 2009 B2
7610563 Nelson et al. Oct 2009 B2
7619615 Donoghue Nov 2009 B1
7640518 Forlines et al. Dec 2009 B2
7653883 Hotelling et al. Jan 2010 B2
7657849 Chaudhri et al. Feb 2010 B2
7663607 Hotelling et al. Feb 2010 B2
7664067 Pointer Feb 2010 B2
7669140 Matthews et al. Feb 2010 B2
7671756 Herz et al. Mar 2010 B2
7694221 Fortes Apr 2010 B2
7702683 Kirshenbaum Apr 2010 B1
7755674 Kaminaga Jul 2010 B2
7792925 Werner et al. Sep 2010 B1
7834861 Lee Nov 2010 B2
7844915 Platzer et al. Nov 2010 B2
7877707 Westerman et al. Jan 2011 B2
7880728 De Los Reyes et al. Feb 2011 B2
7889180 Byun et al. Feb 2011 B2
7895309 Belali et al. Feb 2011 B2
7903115 Platzer et al. Mar 2011 B2
7924271 Christie et al. Apr 2011 B2
7933632 Flynt et al. Apr 2011 B2
7962281 Rasmussen et al. Jun 2011 B2
7983718 Roka Jul 2011 B1
7987431 Santoro et al. Jul 2011 B2
8006276 Nakagawa et al. Aug 2011 B2
8028239 Al-Hilali et al. Sep 2011 B1
8065629 Ragan Nov 2011 B1
8086275 Wykes Dec 2011 B2
8108781 Laansoo et al. Jan 2012 B2
8131808 Aoki et al. Mar 2012 B2
8150924 Buchheit et al. Apr 2012 B2
8171431 Grossman et al. May 2012 B2
8175653 Smuga et al. May 2012 B2
8176438 Zaman et al. May 2012 B2
8209623 Barletta et al. Jun 2012 B2
8225193 Kleinschnitz et al. Jul 2012 B1
8238876 Teng Aug 2012 B2
8245152 Brunner et al. Aug 2012 B2
8250494 Butcher Aug 2012 B2
8255473 Eren et al. Aug 2012 B2
8255812 Parparita et al. Aug 2012 B1
8269736 Wilairat Sep 2012 B2
8279241 Fong Oct 2012 B2
8307279 Fioravanti et al. Nov 2012 B1
8384726 Grabowski et al. Feb 2013 B1
8429565 Agarawala et al. Apr 2013 B2
8448083 Migos et al. May 2013 B1
8473870 Hinckley et al. Jun 2013 B2
8493510 Bryan et al. Jul 2013 B2
8525808 Buening Sep 2013 B1
8539384 Hinckley et al. Sep 2013 B2
8548431 Teng et al. Oct 2013 B2
8560959 Zaman et al. Oct 2013 B2
8589815 Fong et al. Nov 2013 B2
8612874 Zaman et al. Dec 2013 B2
8624933 Marr et al. Jan 2014 B2
8627227 Matthews et al. Jan 2014 B2
8665272 Fitzmaurice et al. Mar 2014 B2
8669950 Forstall et al. Mar 2014 B2
8687023 Markiewicz et al. Apr 2014 B2
8689123 Zaman et al. Apr 2014 B2
8706515 Cobbs et al. Apr 2014 B2
8830270 Zaman et al. Sep 2014 B2
8893033 Donahue et al. Nov 2014 B2
8922575 Garside et al. Dec 2014 B2
8933952 Zaman et al. Jan 2015 B2
8935631 Leonard et al. Jan 2015 B2
8990733 Deutsch et al. Mar 2015 B2
9015606 Zaman et al. Apr 2015 B2
9052820 Jarrett et al. Jun 2015 B2
9104307 Jarrett et al. Aug 2015 B2
9104440 Jarrett et al. Aug 2015 B2
9141262 Nan et al. Sep 2015 B2
9146670 Zaman et al. Sep 2015 B2
9158445 Wong et al. Oct 2015 B2
9213468 Zaman et al. Dec 2015 B2
9229918 Zaman et al. Jan 2016 B2
9244802 Yalovsky et al. Jan 2016 B2
20010022621 Squibbs Sep 2001 A1
20020000963 Yoshida et al. Jan 2002 A1
20020018051 Singh Feb 2002 A1
20020026524 Dharap Feb 2002 A1
20020035607 Checkoway Mar 2002 A1
20020054117 van Dantzich et al. May 2002 A1
20020060701 Naughton et al. May 2002 A1
20020070961 Xu et al. Jun 2002 A1
20020077156 Smethers Jun 2002 A1
20020083025 Robarts et al. Jun 2002 A1
20020091755 Narin Jul 2002 A1
20020097264 Dutta et al. Jul 2002 A1
20020105531 Niemi Aug 2002 A1
20020115476 Padawer et al. Aug 2002 A1
20020128036 Yach et al. Sep 2002 A1
20020129061 Swart et al. Sep 2002 A1
20020138248 Corston-Oliver et al. Sep 2002 A1
20020142762 Chmaytelli et al. Oct 2002 A1
20020145631 Arbab et al. Oct 2002 A1
20020149622 Uesaki et al. Oct 2002 A1
20020152305 Jackson et al. Oct 2002 A1
20020154176 Barksdale et al. Oct 2002 A1
20020161634 Kaars Oct 2002 A1
20020165923 Prince Nov 2002 A1
20020186251 Himmel et al. Dec 2002 A1
20020194385 Linder et al. Dec 2002 A1
20030003899 Tashiro et al. Jan 2003 A1
20030008686 Park et al. Jan 2003 A1
20030011643 Nishihata Jan 2003 A1
20030020671 Santoro et al. Jan 2003 A1
20030040300 Bodic Feb 2003 A1
20030046396 Richter et al. Mar 2003 A1
20030052900 Card et al. Mar 2003 A1
20030073414 Capps Apr 2003 A1
20030096604 Vollandt May 2003 A1
20030105827 Tan et al. Jun 2003 A1
20030135582 Allen et al. Jul 2003 A1
20030187996 Cardina et al. Oct 2003 A1
20030222907 Heikes et al. Dec 2003 A1
20030225846 Heikes et al. Dec 2003 A1
20030231327 Ashey et al. Dec 2003 A1
20040066414 Czerwinski et al. Apr 2004 A1
20040068543 Seifert Apr 2004 A1
20040078299 Down-Logan Apr 2004 A1
20040111673 Bowman et al. Jun 2004 A1
20040128359 Horvitz et al. Jul 2004 A1
20040155908 Wagner Aug 2004 A1
20040185883 Rukman Sep 2004 A1
20040212586 Denny Oct 2004 A1
20040217954 O'Gorman et al. Nov 2004 A1
20040217980 Radburn et al. Nov 2004 A1
20040237048 Tojo et al. Nov 2004 A1
20040250217 Tojo et al. Dec 2004 A1
20040266491 Howard et al. Dec 2004 A1
20050005241 Hunleth et al. Jan 2005 A1
20050028208 Ellis Feb 2005 A1
20050044058 Matthews et al. Feb 2005 A1
20050050462 Whittle et al. Mar 2005 A1
20050054384 Pasquale et al. Mar 2005 A1
20050058353 Matsubara Mar 2005 A1
20050060647 Doan et al. Mar 2005 A1
20050060663 Arkeketa et al. Mar 2005 A1
20050060665 Rekimoto Mar 2005 A1
20050079896 Kokko et al. Apr 2005 A1
20050081155 Martin et al. Apr 2005 A1
20050085215 Kokko Apr 2005 A1
20050085272 Anderson et al. Apr 2005 A1
20050108655 Andrea et al. May 2005 A1
20050114788 Fabritius May 2005 A1
20050120306 Klassen et al. Jun 2005 A1
20050125736 Ferri et al. Jun 2005 A1
20050143138 Lee et al. Jun 2005 A1
20050149879 Jobs et al. Jul 2005 A1
20050156947 Sakai et al. Jul 2005 A1
20050182798 Todd et al. Aug 2005 A1
20050183021 Allen et al. Aug 2005 A1
20050184999 Daioku Aug 2005 A1
20050198159 Kirsch Sep 2005 A1
20050198584 Matthews et al. Sep 2005 A1
20050200762 Barletta et al. Sep 2005 A1
20050207734 Howell et al. Sep 2005 A1
20050216300 Appelman et al. Sep 2005 A1
20050223057 Buchheit et al. Oct 2005 A1
20050223069 Cooperman et al. Oct 2005 A1
20050232166 Nierhaus Oct 2005 A1
20050250547 Salman et al. Nov 2005 A1
20050268237 Crane et al. Dec 2005 A1
20050273614 Ahuja Dec 2005 A1
20050280719 Kim Dec 2005 A1
20060004685 Pyhalammi et al. Jan 2006 A1
20060010394 Chaudhri et al. Jan 2006 A1
20060015736 Callas et al. Jan 2006 A1
20060015812 Cunningham Jan 2006 A1
20060026013 Kraft Feb 2006 A1
20060036425 Le Cocq et al. Feb 2006 A1
20060048073 Jarrett et al. Mar 2006 A1
20060048101 Krassovsky et al. Mar 2006 A1
20060059430 Bells Mar 2006 A1
20060061597 Hui Mar 2006 A1
20060070005 Gilbert et al. Mar 2006 A1
20060074735 Shukla et al. Apr 2006 A1
20060074771 Kim Apr 2006 A1
20060075360 Bixler Apr 2006 A1
20060103623 Davis May 2006 A1
20060107231 Matthews et al. May 2006 A1
20060112354 Park et al. May 2006 A1
20060129543 Bates et al. Jun 2006 A1
20060135220 Kim et al. Jun 2006 A1
20060136773 Kespohl et al. Jun 2006 A1
20060152803 Provitola Jul 2006 A1
20060156228 Gallo et al. Jul 2006 A1
20060161863 Gallo Jul 2006 A1
20060172724 Linkert et al. Aug 2006 A1
20060173911 Levin et al. Aug 2006 A1
20060176403 Gritton et al. Aug 2006 A1
20060184901 Dietz Aug 2006 A1
20060190833 SanGiovanni et al. Aug 2006 A1
20060199598 Lee et al. Sep 2006 A1
20060212806 Griffin et al. Sep 2006 A1
20060218234 Deng et al. Sep 2006 A1
20060218501 Wilson et al. Sep 2006 A1
20060224993 Wong et al. Oct 2006 A1
20060227153 Anwar et al. Oct 2006 A1
20060236264 Cain et al. Oct 2006 A1
20060246955 Nirhamo Nov 2006 A1
20060253685 Wong et al. Nov 2006 A1
20060253801 Okaro et al. Nov 2006 A1
20060259870 Hewitt et al. Nov 2006 A1
20060259873 Mister Nov 2006 A1
20060262134 Hamiter et al. Nov 2006 A1
20060268100 Karukka et al. Nov 2006 A1
20060271520 Ragan Nov 2006 A1
20060281448 Plestid et al. Dec 2006 A1
20060288280 Makela Dec 2006 A1
20060293088 Kokubo Dec 2006 A1
20060294063 Ali et al. Dec 2006 A1
20060294396 Witman Dec 2006 A1
20070005716 LaVasseur et al. Jan 2007 A1
20070006094 Canfield et al. Jan 2007 A1
20070011610 Sethi et al. Jan 2007 A1
20070015532 Deelman Jan 2007 A1
20070024646 Saarinen Feb 2007 A1
20070028267 Ostojic et al. Feb 2007 A1
20070035513 Sherrard et al. Feb 2007 A1
20070038567 Allaire et al. Feb 2007 A1
20070044039 Amadio et al. Feb 2007 A1
20070050724 Lee et al. Mar 2007 A1
20070054679 Cho et al. Mar 2007 A1
20070055770 Karmakar et al. Mar 2007 A1
20070061488 Alagappan et al. Mar 2007 A1
20070061714 Stuple et al. Mar 2007 A1
20070063995 Bailey et al. Mar 2007 A1
20070067272 Flynt Mar 2007 A1
20070067737 Zielinski et al. Mar 2007 A1
20070073718 Ramer Mar 2007 A1
20070076013 Campbell Apr 2007 A1
20070080954 Griffin Apr 2007 A1
20070082707 Flynt et al. Apr 2007 A1
20070082708 Griffin Apr 2007 A1
20070083746 Fallon et al. Apr 2007 A1
20070083821 Garbow et al. Apr 2007 A1
20070094045 Cobbs et al. Apr 2007 A1
20070106635 Frieden et al. May 2007 A1
20070120835 Sato May 2007 A1
20070127638 Doulton Jun 2007 A1
20070143705 Peters Jun 2007 A1
20070152961 Dunton et al. Jul 2007 A1
20070157089 Van Os et al. Jul 2007 A1
20070157105 Owens et al. Jul 2007 A1
20070171192 Seo et al. Jul 2007 A1
20070180381 Rice Aug 2007 A1
20070182595 Ghasabian Aug 2007 A1
20070182999 Anthony et al. Aug 2007 A1
20070185847 Budzik et al. Aug 2007 A1
20070192707 Maeda et al. Aug 2007 A1
20070192726 Kim et al. Aug 2007 A1
20070192730 Simila et al. Aug 2007 A1
20070192733 Horiuchi Aug 2007 A1
20070192739 Hunleth et al. Aug 2007 A1
20070197196 Shenfield et al. Aug 2007 A1
20070198420 Goldstein Aug 2007 A1
20070208840 Mcconville et al. Sep 2007 A1
20070211034 Griffin et al. Sep 2007 A1
20070214429 Lyudovyk et al. Sep 2007 A1
20070216651 Patel Sep 2007 A1
20070216661 Chen et al. Sep 2007 A1
20070222769 Otsuka et al. Sep 2007 A1
20070225022 Satake Sep 2007 A1
20070233654 Karlson Oct 2007 A1
20070236468 Tuli Oct 2007 A1
20070238488 Scott Oct 2007 A1
20070247435 Benko et al. Oct 2007 A1
20070250583 Hardy et al. Oct 2007 A1
20070250787 Kawahara et al. Oct 2007 A1
20070253758 Suess Nov 2007 A1
20070256029 Maxwell Nov 2007 A1
20070257891 Esenther et al. Nov 2007 A1
20070257933 Klassen Nov 2007 A1
20070260674 Shenfield Nov 2007 A1
20070262964 Zotov et al. Nov 2007 A1
20070263843 Foxenland Nov 2007 A1
20070273663 Park et al. Nov 2007 A1
20070273668 Park et al. Nov 2007 A1
20070280457 Aberethy Dec 2007 A1
20070281747 Pletikosa Dec 2007 A1
20080005668 Mavinkurve Jan 2008 A1
20080028294 Sell et al. Jan 2008 A1
20080032681 West Feb 2008 A1
20080034318 Louch et al. Feb 2008 A1
20080036743 Westerman Feb 2008 A1
20080040692 Sunday et al. Feb 2008 A1
20080048986 Khoo Feb 2008 A1
20080052370 Snyder Feb 2008 A1
20080057910 Thoresson et al. Mar 2008 A1
20080057926 Forstall et al. Mar 2008 A1
20080065607 Weber Mar 2008 A1
20080072173 Brunner et al. Mar 2008 A1
20080076472 Hyatt Mar 2008 A1
20080082911 Sorotokin et al. Apr 2008 A1
20080082934 Kocienda et al. Apr 2008 A1
20080085700 Arora Apr 2008 A1
20080092054 Bhumkar et al. Apr 2008 A1
20080094368 Ording et al. Apr 2008 A1
20080095100 Cleveland et al. Apr 2008 A1
20080102863 Hardy May 2008 A1
20080104544 Collins et al. May 2008 A1
20080107057 Kannan et al. May 2008 A1
20080113656 Lee et al. May 2008 A1
20080114535 Nesbitt May 2008 A1
20080122796 Jobs May 2008 A1
20080132252 Altman et al. Jun 2008 A1
20080138030 Bryan et al. Jun 2008 A1
20080141153 Samson et al. Jun 2008 A1
20080153551 Baek et al. Jun 2008 A1
20080155425 Murthy et al. Jun 2008 A1
20080158189 Kim Jul 2008 A1
20080162651 Madnani Jul 2008 A1
20080163104 Haug Jul 2008 A1
20080165132 Weiss Jul 2008 A1
20080165136 Christie et al. Jul 2008 A1
20080165141 Christie Jul 2008 A1
20080165163 Bathiche Jul 2008 A1
20080165210 Platzer et al. Jul 2008 A1
20080167058 Lee et al. Jul 2008 A1
20080168349 Lamiraux et al. Jul 2008 A1
20080168379 Forstall et al. Jul 2008 A1
20080168382 Louch et al. Jul 2008 A1
20080168402 Blumenberg Jul 2008 A1
20080168403 Westerman et al. Jul 2008 A1
20080172609 Rytivaara Jul 2008 A1
20080174570 Jobs et al. Jul 2008 A1
20080180399 Cheng Jul 2008 A1
20080182628 Lee et al. Jul 2008 A1
20080184112 Chiang et al. Jul 2008 A1
20080189653 Taylor et al. Aug 2008 A1
20080189658 Jeong et al. Aug 2008 A1
20080192056 Robertson et al. Aug 2008 A1
20080198141 Lee et al. Aug 2008 A1
20080200142 Abdel-Kader et al. Aug 2008 A1
20080208973 Hayashi Aug 2008 A1
20080215475 Ramer et al. Sep 2008 A1
20080222273 Lakshmanan Sep 2008 A1
20080222545 Lemay et al. Sep 2008 A1
20080222547 Wong et al. Sep 2008 A1
20080222560 Harrison Sep 2008 A1
20080222569 Champion Sep 2008 A1
20080225014 Kim Sep 2008 A1
20080242362 Duarte Oct 2008 A1
20080259042 Thorn Oct 2008 A1
20080261513 Shin et al. Oct 2008 A1
20080261660 Huh et al. Oct 2008 A1
20080263457 Kim et al. Oct 2008 A1
20080270558 Ma Oct 2008 A1
20080284798 Weybrew et al. Nov 2008 A1
20080297475 Woolf et al. Dec 2008 A1
20080299999 Lockhart et al. Dec 2008 A1
20080301046 Martinez Dec 2008 A1
20080301575 Fermon Dec 2008 A1
20080307351 Louch et al. Dec 2008 A1
20080309626 Westerman et al. Dec 2008 A1
20080313540 Dirks et al. Dec 2008 A1
20080316177 Tseng Dec 2008 A1
20080317240 Chang et al. Dec 2008 A1
20080320413 Oshiro Dec 2008 A1
20090007009 Luneau et al. Jan 2009 A1
20090007017 Anzures et al. Jan 2009 A1
20090012952 Fredriksson Jan 2009 A1
20090029736 Kim et al. Jan 2009 A1
20090031243 Kano et al. Jan 2009 A1
20090031247 Walter et al. Jan 2009 A1
20090037469 Kirsch Feb 2009 A1
20090037846 Spalink et al. Feb 2009 A1
20090051671 Konstas Feb 2009 A1
20090055749 Chatterjee et al. Feb 2009 A1
20090058821 Chaudhri et al. Mar 2009 A1
20090061837 Chaudhri et al. Mar 2009 A1
20090061948 Lee et al. Mar 2009 A1
20090064055 Chaudhri Mar 2009 A1
20090070673 Barkan et al. Mar 2009 A1
20090077649 Lockhart Mar 2009 A1
20090079740 Fitzmaurice et al. Mar 2009 A1
20090083656 Dokhon Mar 2009 A1
20090085851 Lim Apr 2009 A1
20090085878 Heubel Apr 2009 A1
20090089215 Newton Apr 2009 A1
20090089459 Jeyaseelan et al. Apr 2009 A1
20090089704 Makela Apr 2009 A1
20090094562 Jeong et al. Apr 2009 A1
20090103515 Pointer Apr 2009 A1
20090106696 Duarte Apr 2009 A1
20090109243 Kraft Apr 2009 A1
20090117942 Boningue et al. May 2009 A1
20090125844 Weir et al. May 2009 A1
20090132942 Santoro et al. May 2009 A1
20090140061 Schultz et al. Jun 2009 A1
20090140986 Karkkainen et al. Jun 2009 A1
20090144642 Crystal Jun 2009 A1
20090144652 Wiley Jun 2009 A1
20090144653 Ubillos Jun 2009 A1
20090144753 Morris Jun 2009 A1
20090146962 Ahonen et al. Jun 2009 A1
20090150618 Allen, Jr. et al. Jun 2009 A1
20090153492 Popp Jun 2009 A1
20090160809 Yang Jun 2009 A1
20090163182 Gatti et al. Jun 2009 A1
20090164888 Phan Jun 2009 A1
20090164928 Brown et al. Jun 2009 A1
20090164936 Kawaguchi Jun 2009 A1
20090171920 Wade et al. Jul 2009 A1
20090178007 Matas et al. Jul 2009 A1
20090182788 Chung et al. Jul 2009 A1
20090184939 Wohlstadter et al. Jul 2009 A1
20090189868 Joo et al. Jul 2009 A1
20090192942 Cottrille et al. Jul 2009 A1
20090193358 Mernyk et al. Jul 2009 A1
20090199122 Deutsch et al. Aug 2009 A1
20090199128 Matthews et al. Aug 2009 A1
20090199130 Tsern et al. Aug 2009 A1
20090205041 Michalske Aug 2009 A1
20090215504 Lando Aug 2009 A1
20090225038 Bolsinga et al. Sep 2009 A1
20090228825 Van Os et al. Sep 2009 A1
20090228841 Hildreth Sep 2009 A1
20090235200 Deutsch et al. Sep 2009 A1
20090235203 Iizuka Sep 2009 A1
20090248421 Michaelis et al. Oct 2009 A1
20090249257 Bove et al. Oct 2009 A1
20090265662 Bamford Oct 2009 A1
20090271778 Mandyam et al. Oct 2009 A1
20090284482 Chin Nov 2009 A1
20090284657 Roberts et al. Nov 2009 A1
20090288044 Matthews et al. Nov 2009 A1
20090292989 Matthews et al. Nov 2009 A1
20090293007 Duarte et al. Nov 2009 A1
20090293013 O'Shaugnessy et al. Nov 2009 A1
20090298547 Kim et al. Dec 2009 A1
20090303231 Robinet et al. Dec 2009 A1
20090305732 Marcellino et al. Dec 2009 A1
20090307105 Lemay et al. Dec 2009 A1
20090307589 Inose et al. Dec 2009 A1
20090307623 Agarawala et al. Dec 2009 A1
20090313584 Kerr et al. Dec 2009 A1
20090315839 Wilson et al. Dec 2009 A1
20090315847 Fujii Dec 2009 A1
20090322760 Kwiatkowski Dec 2009 A1
20090327969 Estrada Dec 2009 A1
20100008490 Gharachorloo et al. Jan 2010 A1
20100010934 Barry et al. Jan 2010 A1
20100013782 Liu et al. Jan 2010 A1
20100020025 Lemort et al. Jan 2010 A1
20100020091 Rasmussen et al. Jan 2010 A1
20100030578 Siddique et al. Feb 2010 A1
20100031186 Tseng et al. Feb 2010 A1
20100042911 Wormald et al. Feb 2010 A1
20100050076 Roth Feb 2010 A1
20100057566 Itzhak Mar 2010 A1
20100058248 Park Mar 2010 A1
20100062811 Park et al. Mar 2010 A1
20100066698 Seo Mar 2010 A1
20100070931 Nichols Mar 2010 A1
20100073160 Gilmour et al. Mar 2010 A1
20100073380 Kaplan et al. Mar 2010 A1
20100075628 Ye Mar 2010 A1
20100077058 Messer Mar 2010 A1
20100077310 Karachale et al. Mar 2010 A1
20100077330 Kaplan et al. Mar 2010 A1
20100079392 Chiang et al. Apr 2010 A1
20100079413 Kawashima et al. Apr 2010 A1
20100081475 Chiang et al. Apr 2010 A1
20100086022 Hunleth et al. Apr 2010 A1
20100087169 Lin Apr 2010 A1
20100087173 Lin Apr 2010 A1
20100088635 Louch Apr 2010 A1
20100088641 Choi Apr 2010 A1
20100100839 Tseng et al. Apr 2010 A1
20100102998 Fux Apr 2010 A1
20100103118 Townsend et al. Apr 2010 A1
20100103124 Kruzeniski Apr 2010 A1
20100105370 Kruzeniski Apr 2010 A1
20100105424 Smuga Apr 2010 A1
20100105438 Wykes Apr 2010 A1
20100105439 Friedman Apr 2010 A1
20100105440 Kruzeniski Apr 2010 A1
20100105441 Voss Apr 2010 A1
20100106915 Krishnaprasad et al. Apr 2010 A1
20100107067 Vaisanen Apr 2010 A1
20100107068 Butcher Apr 2010 A1
20100107100 Schneekloth Apr 2010 A1
20100121705 Ramer et al. May 2010 A1
20100122110 Ordogh May 2010 A1
20100123737 Williamson et al. May 2010 A1
20100138767 Wang et al. Jun 2010 A1
20100145675 Lloyd et al. Jun 2010 A1
20100146384 Peev et al. Jun 2010 A1
20100146437 Woodcock et al. Jun 2010 A1
20100159966 Friedman Jun 2010 A1
20100159994 Stallings et al. Jun 2010 A1
20100159995 Stallings et al. Jun 2010 A1
20100162180 Dunnam et al. Jun 2010 A1
20100167699 Sigmund et al. Jul 2010 A1
20100169766 Duarte et al. Jul 2010 A1
20100169772 Stallings et al. Jul 2010 A1
20100169819 Bestle et al. Jul 2010 A1
20100175018 Petschnigg et al. Jul 2010 A1
20100175029 Williams Jul 2010 A1
20100180233 Kruzeniski Jul 2010 A1
20100185932 Coffman et al. Jul 2010 A1
20100216491 Winkler et al. Aug 2010 A1
20100223569 Vuong et al. Sep 2010 A1
20100248688 Teng Sep 2010 A1
20100248689 Teng Sep 2010 A1
20100248741 Setlur et al. Sep 2010 A1
20100248787 Smuga Sep 2010 A1
20100248788 Yook et al. Sep 2010 A1
20100251153 SanGiovanni et al. Sep 2010 A1
20100251167 DeLuca et al. Sep 2010 A1
20100265196 Lee et al. Oct 2010 A1
20100281402 Staikos et al. Nov 2010 A1
20100281409 Rainisto et al. Nov 2010 A1
20100281481 Rainisto et al. Nov 2010 A1
20100283743 Coddington et al. Nov 2010 A1
20100289806 Lao et al. Nov 2010 A1
20100293056 Flynt et al. Nov 2010 A1
20100293501 Russ et al. Nov 2010 A1
20100295789 Shin et al. Nov 2010 A1
20100295795 Wilairat Nov 2010 A1
20100298034 Shin et al. Nov 2010 A1
20100302172 Wilairat Dec 2010 A1
20100302176 Nikula et al. Dec 2010 A1
20100302278 Shaffer et al. Dec 2010 A1
20100302712 Zednicek et al. Dec 2010 A1
20100311470 Seo et al. Dec 2010 A1
20100313165 Louch et al. Dec 2010 A1
20100321403 Inadome Dec 2010 A1
20100328431 Kim et al. Dec 2010 A1
20100329642 Kam et al. Dec 2010 A1
20100333008 Taylor Dec 2010 A1
20110004839 Cha et al. Jan 2011 A1
20110004845 Ciabarra Jan 2011 A1
20110018806 Yano Jan 2011 A1
20110029598 Arnold et al. Feb 2011 A1
20110029904 Smith et al. Feb 2011 A1
20110029927 Lietzke et al. Feb 2011 A1
20110029934 Locker et al. Feb 2011 A1
20110032365 Yett Feb 2011 A1
20110035702 Williams et al. Feb 2011 A1
20110043527 Ording et al. Feb 2011 A1
20110055773 Agarawala et al. Mar 2011 A1
20110055775 Saito Mar 2011 A1
20110074699 Marr et al. Mar 2011 A1
20110074710 Weeldreyer et al. Mar 2011 A1
20110074719 Yeh et al. Mar 2011 A1
20110078624 Missig et al. Mar 2011 A1
20110087988 Ray et al. Apr 2011 A1
20110093778 Kim et al. Apr 2011 A1
20110093815 Gobeil Apr 2011 A1
20110093816 Chang et al. Apr 2011 A1
20110093821 Wigdor et al. Apr 2011 A1
20110107272 Aguilar May 2011 A1
20110113337 Liu et al. May 2011 A1
20110113486 Hunt et al. May 2011 A1
20110119586 Blinnikka et al. May 2011 A1
20110124376 Kim et al. May 2011 A1
20110126156 Krishnaraj et al. May 2011 A1
20110138313 Decker et al. Jun 2011 A1
20110154235 Min et al. Jun 2011 A1
20110157027 Rissa Jun 2011 A1
20110161845 Stallings et al. Jun 2011 A1
20110163968 Hogan Jul 2011 A1
20110167403 French et al. Jul 2011 A1
20110173556 Czerwinski et al. Jul 2011 A1
20110173568 Royal, Jr. et al. Jul 2011 A1
20110173569 Howes et al. Jul 2011 A1
20110175930 Hwang et al. Jul 2011 A1
20110181617 Tsuda et al. Jul 2011 A1
20110202834 Mandryk et al. Aug 2011 A1
20110202837 Fong et al. Aug 2011 A1
20110202866 Huang et al. Aug 2011 A1
20110209039 Hinckley et al. Aug 2011 A1
20110209089 Hinckley et al. Aug 2011 A1
20110209100 Hinckley et al. Aug 2011 A1
20110209101 Hinckley et al. Aug 2011 A1
20110209102 Hinckley et al. Aug 2011 A1
20110209103 Hinckley et al. Aug 2011 A1
20110209104 Hinckley et al. Aug 2011 A1
20110225547 Fong et al. Sep 2011 A1
20110231796 Vigil Sep 2011 A1
20110252346 Chaudhri Oct 2011 A1
20110252380 Chaudhri Oct 2011 A1
20110258563 Lincke Oct 2011 A1
20110276864 Oules Nov 2011 A1
20110283188 Farrenkopf et al. Nov 2011 A1
20110316884 Giambalvo et al. Dec 2011 A1
20120005584 Seago et al. Jan 2012 A1
20120009903 Schultz et al. Jan 2012 A1
20120017162 Khokhlov Jan 2012 A1
20120028687 Wykes Feb 2012 A1
20120050185 Davydov et al. Mar 2012 A1
20120050332 Nikara et al. Mar 2012 A1
20120062604 Lobo Mar 2012 A1
20120081310 Chrock et al. Apr 2012 A1
20120089950 Tseng Apr 2012 A1
20120102433 Falkenburg Apr 2012 A1
20120124469 Nakajima et al. May 2012 A1
20120151397 Oberstein et al. Jun 2012 A1
20120159395 Deutsch et al. Jun 2012 A1
20120159402 Nurmi et al. Jun 2012 A1
20120162266 Douglas et al. Jun 2012 A1
20120167008 Zaman Jun 2012 A1
20120167011 Zaman Jun 2012 A1
20120169593 Mak et al. Jul 2012 A1
20120174029 Bastide et al. Jul 2012 A1
20120176401 Hayward et al. Jul 2012 A1
20120179992 Smuga Jul 2012 A1
20120210265 Delia et al. Aug 2012 A1
20120212495 Butcher Aug 2012 A1
20120216139 Ording et al. Aug 2012 A1
20120226971 Tocchini et al. Sep 2012 A1
20120233571 Wever et al. Sep 2012 A1
20120236035 Kimura Sep 2012 A1
20120244841 Teng Sep 2012 A1
20120254780 Mouton Oct 2012 A1
20120265644 Roa et al. Oct 2012 A1
20120290962 Zielinski et al. Nov 2012 A1
20120299968 Wong et al. Nov 2012 A1
20120304068 Zaman et al. Nov 2012 A1
20120304092 Jarrett et al. Nov 2012 A1
20120304108 Jarrett et al. Nov 2012 A1
20120304113 Patten et al. Nov 2012 A1
20120304114 Wong et al. Nov 2012 A1
20120304116 Donahue et al. Nov 2012 A1
20120304117 Donahue et al. Nov 2012 A1
20120304118 Donahue et al. Nov 2012 A1
20120311485 Caliendo, Jr. et al. Dec 2012 A1
20120323992 Brobst et al. Dec 2012 A1
20130031508 Kodosky et al. Jan 2013 A1
20130033525 Markiewicz Feb 2013 A1
20130042203 Wong et al. Feb 2013 A1
20130042206 Zaman et al. Feb 2013 A1
20130044141 Markiewicz et al. Feb 2013 A1
20130047079 Kroeger et al. Feb 2013 A1
20130047105 Jarrett Feb 2013 A1
20130047117 Deutsch Feb 2013 A1
20130057587 Leonard et al. Mar 2013 A1
20130057588 Leonard Mar 2013 A1
20130063442 Zaman Mar 2013 A1
20130063443 Garside Mar 2013 A1
20130063465 Zaman Mar 2013 A1
20130063490 Zaman Mar 2013 A1
20130067381 Yalovsky Mar 2013 A1
20130067390 Kwiatkowski Mar 2013 A1
20130067391 Pittappilly Mar 2013 A1
20130067398 Pittappilly Mar 2013 A1
20130067399 Elliott Mar 2013 A1
20130067412 Leonard Mar 2013 A1
20130067420 Pittappilly Mar 2013 A1
20130093757 Cornell Apr 2013 A1
20130111396 Brid May 2013 A1
20130169649 Bates Jul 2013 A1
20130176316 Bates Jul 2013 A1
20140033099 Treitman et al. Jan 2014 A1
20140082552 Zaman et al. Mar 2014 A1
20140098108 Fong et al. Apr 2014 A1
20140109008 Zaman et al. Apr 2014 A1
20150046829 Donahue et al. Feb 2015 A1
20150058763 Leonard et al. Feb 2015 A1
20150113476 Deutsch et al. Apr 2015 A1
20150186376 Garside et al. Jul 2015 A1
20150193403 Zaman et al. Jul 2015 A1
20150317062 Jarrett et al. Nov 2015 A1
20150378554 Jan et al. Dec 2015 A1
20150378594 Zaman et al. Dec 2015 A1
20160041710 Zaman et al. Feb 2016 A1
20160041711 Zaman et al. Feb 2016 A1
20160041730 Zaman et al. Feb 2016 A1
20160070357 Huang et al. Mar 2016 A1
20160110090 Patten et al. Apr 2016 A1
Foreign Referenced Citations (105)
Number Date Country
1326564 Dec 2001 CN
1591305 Mar 2005 CN
1734440 Feb 2006 CN
1845054 Oct 2006 CN
1902575 Jan 2007 CN
1904823 Jan 2007 CN
1930568 Mar 2007 CN
101036104 Sep 2007 CN
101114303 Jan 2008 CN
101233477 Jul 2008 CN
101233504 Jul 2008 CN
101578577 Nov 2009 CN
101809531 Aug 2010 CN
101819498 Sep 2010 CN
101981522 Feb 2011 CN
102004603 Apr 2011 CN
102033710 Apr 2011 CN
102144213 Aug 2011 CN
102197702 Sep 2011 CN
102460370 May 2012 CN
0583060 Feb 1994 EP
1353505 Oct 2003 EP
1752868 Feb 2007 EP
1939718 Jul 2008 EP
2172836 Apr 2009 EP
2262193 Dec 2010 EP
2659347 Nov 2013 EP
2350991 Dec 2000 GB
2003513350 Apr 2003 JP
2004227393 Aug 2004 JP
2004357257 Dec 2004 JP
2005527888 Sep 2005 JP
2006268849 Oct 2006 JP
2006293989 Oct 2006 JP
2006323672 Nov 2006 JP
2007058740 Mar 2007 JP
2007516496 Jun 2007 JP
2007195186 Aug 2007 JP
2008508600 Mar 2008 JP
2008527540 Jul 2008 JP
2008276584 Nov 2008 JP
2009266192 Nov 2009 JP
2010039761 Feb 2010 JP
2010073099 Apr 2010 JP
2011048835 Mar 2011 JP
2011070525 Apr 2011 JP
2011516936 May 2011 JP
2011128029 Jun 2011 JP
2012527684 Nov 2012 JP
2012256147 Dec 2012 JP
200303655 Feb 2003 KR
20060019198 Mar 2006 KR
1020070036114 Apr 2007 KR
20070093084 Sep 2007 KR
1020070098337 Oct 2007 KR
20070120368 Dec 2007 KR
1020080025951 Mar 2008 KR
1020080041809 May 2008 KR
1020080076390 Aug 2008 KR
100854333 Sep 2008 KR
1020080084156 Sep 2008 KR
1020080113913 Dec 2008 KR
1020090041635 Apr 2009 KR
20100010072 Feb 2010 KR
20100048375 May 2010 KR
20100056369 May 2010 KR
1020100056369 May 2010 KR
2409833 Apr 2009 RU
2363033 Jul 2009 RU
2412463 Jan 2010 RU
201023026 Jun 2010 TW
201037592 Oct 2010 TW
I333157 Nov 2010 TW
WO-9926127 May 1999 WO
WO-0129976 Apr 2001 WO
WO03075547 Sep 2003 WO
WO-2005026931 Mar 2005 WO
WO-2005027506 Mar 2005 WO
WO-2006019639 Feb 2006 WO
WO-2006074267 Jul 2006 WO
WO-2007065019 Jun 2007 WO
WO-2007121557 Nov 2007 WO
WO-2007134623 Nov 2007 WO
WO-2008030608 Mar 2008 WO
WO-2008031871 Mar 2008 WO
WO-2008035831 Mar 2008 WO
WO-2009000043 Dec 2008 WO
WO-2009012398 Jan 2009 WO
WO-2009049331 Apr 2009 WO
WO-2009158310 Dec 2009 WO
WO-2010024969 Mar 2010 WO
WO-2010036660 Apr 2010 WO
WO-2010041826 Apr 2010 WO
WO-2010048229 Apr 2010 WO
WO-2010048448 Apr 2010 WO
WO-2010048519 Apr 2010 WO
WO-2010110613 Sep 2010 WO
WO-2010117643 Oct 2010 WO
WO-2010119356 Oct 2010 WO
WO-2010125451 Nov 2010 WO
WO-2010134718 Nov 2010 WO
WO-2010135155 Nov 2010 WO
WO-2011041885 Apr 2011 WO
WO-2012088485 Jun 2012 WO
WO-2012166188 Dec 2012 WO
Non-Patent Literature Citations (498)
Entry
Adobe Acrobat 8 Standard User Guide, Dec. 31, 2007, Apple Inc., pp. 36 and 39.
“Application User Model IDs”, Retrieved from: <http://msdn.microsoft.com/en-us/library/dd378459(VS.85).aspx> on Sep. 28, 2010, (2010),6 pages.
“Email Notification for Microsoft Outlook and Outlook Express”, Retrieved from: <http://www.contextmadic.com/express-notification/> on Sep. 29, 2010, (Jul. 21, 2004),3 pages.
“Enhanced IBM Power Systems Software and PowerVM Restructuring”, IBM United States Announcement 208-082, dated Apr. 8, 2008, available at <http://www.ibm.com/common/ssi/rep—ca/2/897/ENUS208-082/ENUS208082.PDF>,(Apr. 8, 2008),pp. 1-19.
“eXtreme Energy Conservation: Advanced Power-Saving Software for Wireless Devices”, White Paper, Freescale Semiconductor, Inc., Document No. XTMENRGYCNSVWP, Rev #0, available at <http://www.freescale.com/files/32bit/doc/white—paper/XTMENRGYCNSVWP.pdf>,(Feb. 2006), 15 pages.
“Final Office Action”, U.S. Appl. No. 11/305,789, (Apr. 1, 2009), 10 pages.
“Final Office Action”, U.S. Appl. No. 11/502,264, (Feb. 4, 2010), 15 pages.
“Final Office Action”, U.S. Appl. No. 11/502,264, (Apr. 3, 2009), 9 pages.
“How Do I Cancel a “Drag” Motion on an Android Seekbar?”, retrieved from <http://stackoverflow.com/questions/2917969/how-do-i-cancel-a-drag-motion-on-an-android-seekbar> on Jun. 20, 2011,(May 28, 2010), 1 page.
“iPad User Guide”, retrieved from <http://cyndidannerkuhn.info/CDK/iPads—Resources—files/iPad—User—Guide.pdf> on Jun. 17, 2011, 154 pages.
“Magic mouse”, Retrieved from: <http://www.apple.com/magicmouse/> on May 10, 2011,3 pages.
“moGo beta v.0.4”, Retrieved from: <http://forum.xda-developers.com/showthread.php?t=375196> on Sep. 27, 2010, (Mar. 7, 2008), 10 pages.
“New Features in WhatsUp Gold v12.0”, retrieved from <http://www.netbright.co.th/?name=product&file=readproduct&id=12> on Jun. 10, 2011,4 pages.
“Non-Final Office Action”, U.S. Appl. No. 11/305,789, (Sep. 21, 2009),5 pages.
“Non-Final Office Action”, U.S. Appl. No. 11/502,264, (Sep. 30, 2009), 15 pages.
“Notice of Allowance”, U.S. Appl. No. 11/305,789, (Nov. 23, 2009),8 pages.
“Notifications”, retrieved from <http://msdn.microsoft.com/en-us/library/aa511497.aspx> on May 10, 2011, 16 pages.
“OmneMon™ System Resource Metrics”, retrieved from <http://www.omnesys.com/documents/OmneMonSRM—Brochure.pdf> on Jun. 10, 2011,3 pages.
“ONYX Graphics Announces New ONYX Prepedge Job Preparation Software”, retrieved from <http://www.largeformatreview.com/rip-software/433-onyx-graphics-announces-new-onyx-> on May 10, 2011,2 pages.
“Push Notifications Overview for Windows Phone”, Retrieved from: <http://msdn.microsoft.com/en-us/library/ff402558%28VS.92%29.aspx> on Sep. 30, 10 (Sep. 3, 2010), 1 page.
“The Map Screen”, retrieved from <http://www.symbianos.org/whereamiusersguide> on Jun. 17, 2011,3 pages.
“User Guide”, retrieved from <http://wireframesketcher.com/help/help.html> on Jun. 17, 2011, 19 pages.
“Windows 8 is Gorgeous, But is it More Than Just a Shell? (Video)”, retrieved from <http://techcrunch.com/2011/06/02/windows-8-gorgeous-shell-video/> on Jun. 20, 2011,(Jun. 2, 2011),6 pages.
“Windows Phone 7 Live Tiles”, Retrieved from: <http://www.knowyourmobile.com/microsoft/windowsphone7/startscreen/640737/windows—phone—7—live—tiles.html> on May 11, 2011,(Oct. 20, 2010),3 pages.
Bates, John “A Framework to Support Large-Scale”, University of Cambridge Computer Laboratory, Available at <http://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.48.1690&rep=rep1&type=pdf>,(1996),8 pages.
Bjork, Staffan et al., “Redefining the Focus and Context of Focus+Context Visualizations”, In Proceedings of INFOVIS 2000, Available at <http://www.johan.redstrom.se/papers/redefining.pdf>,(Oct. 2000),9 pages.
Buring, Thorsten “User Interaction with Scatterplots on Small Screens—A Comparative Evaluation of Geometric-Semantic Zoom and Fisheye Distortion”, IEEE Transactions on Visualization and Computer Graphics, vol. 12, Issue 5, Available at <http://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.134.4568&rep=rep1&type=pdf>,(Sep. 2006),pp. 829-836.
Carrera, Enrique V., et al., “Conserving Disk Energy in Network Servers”, available at <http://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.6.8301&rep=rep1&type=ps>,(Nov. 2002), 15 pages.
Cawley, Christian “How to Customise Your Windows Phone 7”, Retrieved from: <http://www.brighthub.com/mobile/windows-mobile-platform/articles/95213.aspx> on May 10, 2011,(Nov. 12, 2010),3 pages.
Cawley, Christian “Windows Phone 7 Customization Tips and Tricks”, retrieved from <http://www.brighthub.com/mobile/windows-mobile-platform/articles/95213.aspx> on Jun. 20, 2011,(May 16, 2011),2 pages.
Cohen, Michael F., et al., “Wang Tiles for Image and Texture Generation”, In Proceedings of SIGGRAPH 2003, Available <http://research.microsoft.com/en-us/um/people/cohen/WangFinal.pdf>,(2003),8 pages.
Delimarsky, Den “Sending Tile Push Notifications on Windows Phone 7”, retrieved from <http://mobile.dzone.com/articles/sending-tile-push> on May 10, 2011,(Aug. 25, 2010),2 pages.
Denoue, Laurent et al., “WebNC: Efficient Sharing of Web Applications”, In Proceedings of WWW 2009, Available at <http://www.fxpal.com/publications/FXPAL-PR-09-495.pdf>,(2009),2 pages.
Dunsmuir, Dustin “Selective Semantic Zoom of a Document Collection”, Available at <http://www.cs.ubc.ca/˜tmm/courses/533/projects/dustin/proposal.pdf>,(Oct. 30, 2009),pp. 1-9.
Fisher, Bill “Cool Discussion of Push Notifications—Toast and Tile—on Windows Phone”, Retrieved from: <http://www.windowsphoneexpert.com/Connection/forums/p/4153/18399.aspx> on Sep. 29, 2010 (May 3, 2010),3 pages.
Janecek, Paul et al., “An Evaluation of Semantic Fisheye Views for Opportunistic Search in an Annotated Image Collection”, Available at <http://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.67.3084&rep=rep1&ty pe=pdf>,(Feb. 15, 2005),pp. 1-15.
Long, Todd “Gmail Manager 0.6”, Retrieved from: <https://addons.mozilla.org/en-US/firefox/addon/1320/> on Sep. 29, 2010, (Jan. 27, 2010),4 pages.
Paul, Ryan “Hands-on: KDE 4.5 Launches with Tiling, New Notifications”, Retrieved from: <http://arstechnica.com/open-source/reviews/2010/08/hands-on-kde-45-launches-with-tiling-new-notifications.ars> on Sep. 29, 2010, (Aug. 2010),3 pages.
Ray, Bill “Microsoft Re-Tiles Mobile Platform for Windows 7 Era”, retrieved from <http://www.theregister.co.uk/2010/02/15/windows—phone—7—series/> on May 11, 2011,(Feb. 15, 2010),2 pages.
Ritchie, Rene “iOS 4 features: iPod touch Wi-Fi stays connected when asleep—iPhone too?”, Retrieved from: <http://www.goip.com/2010/06/ios-4-features-ipod-touch-wi-fi-stays-connected-when-asleep-%E2%80%94-iphone-too/> on Sep, 30, 2010, (Jun. 14, 2010),2 pages.
Sandoval, Guillermo L., “A development platform and execution environment for mobile applications”, Universidad Autónoma de Baja California, School of Chemical Sciences and Engineering, Available at <http://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.86.7989&rep=rep1&type=pdf>,(2004),18 pages.
Smith, Greg et al., “GroupBar: The TaskBar Evolved”, Proceedings of OZCHI 2003, Available at <http://research.microsoft.com/pubs/64316/ozchi2003-groupbar.pdf>,(Nov. 2003),pp. 1-10.
Vornberger, Jan “Bluetile”, Retrieved from: <http://www.bluetile.org> on Sep. 29, 2010, 5 pages.
Wu, Chung et al., “Achieving a Superior Ownership Experience in Manageability and Quality for Siebel CRM”, available at <http://wvwv.oracle.com/us/products/enterprise-manager/superior-exp-for-siebel-crm-068962.pdf>,(Aug. 2008),25 pages.
Wilson, Andrew D., “Robust Computer Vision-Based Detection of Pinching for One and Two-Handed Gesture Input”, In Proceedings of UIST 2006, Available at <http://research.microsoft.com/en-us/um/people/awilson/publications/wilsonuist2006/uist%202006%20taffi.pdf>,(Oct. 2006),4 pages.
“PCT Search Report and Written Opinion”, Application No. PCT/US2011/055511, (Apr. 24, 2012), 9 pages.
“PCT Search Report and Written Opinion”, Application No. PCT/US2011/055520, (May 9, 2012), 8 pages.
“PCT Search Report and Written Opinion”, Application No. PCT/US2011/055521, (May 15, 2012), 9 pages.
“PCT Search Report and Written Opinion”, Application No. PCT/US2011/055522, (May 15, 2012), 9 pages.
Ritscher, Walt., “Using Surface APIs in your WPF application—Part 1”, Retrieved at << http://blog.wpfwonderland.com/2009/06/30/using-surface-apis-in-your-wpf-application/>>, Jun. 30, 2009, pp. 7.
Davis, Ashley., “A WPF Custom Control for Zooming and Panning”, Retrieved at << http://www.codeproject.com/KB/WPF/zoomandpancontrol.aspx >>, Jun. 29, 2010, pp. 21.
Bowes, et al., “Transparency for Item Highlighting”, Retrieved at << http://torch.cs.dal.ca/˜dearman/pubs/GI2003-bowes,dearman,perkins-paper.pdf >>, 2003, pp. 2.
“How do I use Categories with my Weblog?”, Retrieved at << http://tpsupport.mtcs.sixapart.com/tp/us-tp1/how—do—i—use—categories—with—my—weblog.html >>, Sep. 16, 2009, pp. 3.
“Class ScrollView”, Retrieved at << http://www.blackberry.com/developers/docs/6.0.0api/net/rim/device/api/ui/ScrollView.html >>, Retrieved Date: Sep. 28, 2010, pp. 13.
“YUI 3: ScrollView [beta]”, Retrieved at << http://developer.yahoo.com/yui/3/scrollview/ >>, Retrieved Date: Sep. 28, 2010, pp. 5.
“GnomeCanvas”, Retrieved at << http://library.gnome.org/devel/libgnomecanvas/unstable/GnomeCanvas—html >>, Retrieved Date: Sep. 28, 2010, pp. 11.
“Advisory Action”, U.S. Appl. No. 12/414,382, (Jan. 20, 2012), 3 pages.
“Advisory Action”, U.S. Appl. No. 12/433,605, (Apr. 5, 2012), 3 pages.
“Alltel Adds Dedicated Search Key to Phones” Retrieved from: <http://www.phonescoop.com/news/item.php?n=2159> on Nov. 26, 2008., (Apr. 12, 2007), 2 Pages.
“Android 2.3 Users Guide”, AUG-2.3-103, Android mobile technology platform 2.3,(Dec. 13, 2010), 380 pages.
“Apple iPhone—8GB AT&T”, Retrieved from: <http://nytimes.com.com/smartphones/apple-iphone-8gb-at/4515-6452—7-32309245.html> on Nov. 20, 2008, (Jun. 29, 2007), 11 pages.
“Ask Web Hosting”, Retrieved from: <http://www.askwebhosting.com/story/18501/HTC—FUZE—From—ATandampT—Fuses—Fun—and—Function—With—the—One-Touch—Power—of—TouchFLO—3D.html> on May 5, 2009., (Nov. 11, 2008), 3 pages.
“Basics of Your Device: Get Familiar with the Home Screen”, Nokia USA—How to, retrieved from <http://www.nokia.ca/get-support-and-software/product-support/c6-01/how-to#> on May 11, 2011, 3 pages.
“Blackberry office tools: Qwerty Convert”, Retrieved from: <http://blackberrysoftwarelist.net/blackberry/download-software/blackberry-office/qwerty—convert.aspx> on Nov. 20, 2008, (Nov. 20, 2008), 1 page.
“Calc4M”, Retrieved from: <http://www.hellebo.com/Calc4M.html> on Dec. 11, 2008, (Sep. 10, 2008), 4 Pages.
“Content-Centric E-Mail Message Analysis in Litigation Document Reviews”, Retrieved from: <http://www.busmanagement.com/article/Issue-14/Data-Management/Content-Centric-E-Mail-Message-Analysis-in-Litigation-Document-Reviews/> on May 6, 2009, (2009), 5 Pages.
“Dial a number”, Retrieved from: <http://www.phonespell.org/ialhelp.html> on Nov. 20, 2008, 1 page.
“DuoSense™ Multi-Touch Gestures”, Retrieved from: <http://www.n-trig.com/Data/Uploads/Misc/DuoSenseMTG—final.pdf>, (Jul. 2008), 4 pages.
“Elecont Quick Desktop 1.0.43”, Retrieved from: <http://handheld.softpedia.com/get/System-Utilities/Launcher-Applications/Elecont-Quick-Desktop-72131.shtml> on May 5, 2009., (Mar. 13, 2009), 2 pages.
“Exclusive: Windows Mobile 7 to Focus on Touch and Motion Gestures”, Retrieved from: <http://anti-linux.blogspot.com/2008/08/exclusive-windows-mobile-7-to-focus-on.html> on May 6, 2009, (Aug. 1, 2008), 14 pages.
“Extended European Search Report”, European Patent Application No. 09818253.8, (Apr. 10, 2012), 7 pages.
“Final Office Action”, U.S. Appl. No. 12/244,545, (Dec. 7, 2011), 16 pages.
“Final Office Action”, U.S. Appl. No. 12/244,545, (Sep. 7, 2012), 23 pages.
“Final Office Action”, U.S. Appl. No. 12/413,977, (Nov. 17, 2011), 16 pages.
“Final Office Action”, U.S. Appl. No. 12/414,382, (Dec. 23, 2011), 7 pages.
“Final Office Action”, U.S. Appl. No. 12/414,476, (Dec. 1, 2011), 20 pages.
“Final Office Action”, U.S. Appl. No. 12/433,605, (Feb. 3, 2012), 11 pages.
“Final Office Action”, U.S. Appl. No. 12/433,667, (Sep. 13, 2011), 17 pages.
“Final Office Action”, U.S. Appl. No. 12/469,458, (Nov. 17, 2011), 15 pages.
“Final Office Action”, U.S. Appl. No. 12/469,480, (Feb. 9, 2012), 17 pages.
“Final Office Action”, U.S. Appl. No. 12/484,799, (Apr. 30, 2012), 13 pages.
“Final Office Action”, U.S. Appl. No. 12/560,081, (Mar. 14, 2012), 16 pages.
“Freeware.mobi”, Retrieved from: <http://www.palmfreeware.mobi/download-palette.html> Nov. 6, 2008, (Oct. 9, 2001), 2 pages.
“How do you dial 1-800-FLOWERS”, Retrieved from: <http://blogs.msdn.com/windowsmobile/archive/2007/02/06/how-do-you-dial-1-800-flowers.aspx> on Nov. 20, 2008, (Feb. 6, 2007), 24 pages.
“HTC Shows HTC Snap with Snappy Email Feature”, Retrieved from: <http://www.wirelessmobilenews.com/smartphones/ on May 5, 2009>, (May 4, 2009), 10 Pages.
“IntelliScreen—New iPhone App Shows Today Screen Type Info in Lock Screen”, Retrieved from: <http://justanotheriphoneblog.com/wordpress//2008/05/13/intelliscreen-new-iphone-app-shows-today-screen-type-info-on-lock-screen/> on Nov. 12, 2008, (May 13, 2008), 11 pages.
“International Search Report and Written Opinion”, International Application No. PCT/US2011/055514, (May 22, 2012), 8 pages.
“International Search Report”, Application No. PCT/US2010/028553, Application Filing Date: Mar. 24, 2010,(Nov. 9, 2010), 9 pages.
“Internet Explorer Window Restrictions”, Retrieved from: http://technet.microsoft.com/en-us/library/cc759517(WS.10).aspx on Jun. 28, 2011, Microsoft TechNet, 5 pages.
“Introduction to Windows Touch”, Retrieved from: <http://download.microsoft.com/download/a/d/f/adf1347d-08dc-41a4-9084-623b1194d4b2/Win7—touch.docx>, (Dec. 18, 2008), pp. 1-7.
“iPod touch User Guide for iPhone OS 3.0 Software”, Apple Inc.,(2009), 153 pages.
“Keyboard (5)”, Retrieved from: <http://landru.uwaterloo.ca/cgi-bin.man.cgi?section=5&topic=keyboard> on Dec. 11, 2008., (Aug. 11, 1997), 8 Pages.
“Keyboard Shortcuts”, Retrieved from: <http://www.pctoday.com/editorial/article.asp?article=articles%2F2005%2Ft0311%2F26t11%2F26t11.asp> on Aug. 3, 2009., (Nov. 2005), 5 pages.
“Kiosk Browser Chrome Customization Firefox 2.x”, Retrieved from: <http://stlouis-shopper.com/cgi-bin/mozdev-wiki/,pl?ChromeCustomization> on Oct. 22, 2008 Making a new chrome for the kiosk browser, Kiosk Project Kiosk Browser Chrome Customization Firefox-2.x, (Aug. 16, 2007), 2 pages.
“Live Photo Gallery—Getting Started—from Camera to Panorama”, Retrieved from: <http://webdotwiz.spaces.live.com/blog/cns!2782760752B93233!1729.entry> on May 5, 2009., (Sep. 2008), 7 Pages.
“MIDTB Tip Sheet: Book Courier”, Retrieved from: <http://www.midtb.org/tipsbookcourier.htm> on Dec. 11, 2008., (Sep. 26, 2005),6 Pages.
“Mobile/UI/Designs/TouchScreen”, Retrieved from: <https://wiki.mozilla.org/Mobile/UI/Desiqns/TouchScreen> on May 6, 2009., (Feb. 3, 2009), 15 Pages.
“Multi-touch”, Retrieved from <http://en.wikipedia.org/wiki/Multi-touch#Microsoft—Surface> on Apr. 24, 2009, (Apr. 17, 2009), 8 pages.
“Nokia E61 Tips and Tricks for Keyboard Shortcuts”, Retrieved from: <http://www.mobiletopsoft.com/board/1810/nokia-e61-tips-and-tricks-for-keyboard-shortcuts.html> on Dec. 17, 2008., (Jan. 27, 2006), 2 Pages.
“Non-Final Office Action”, U.S. Appl. No. 11/215,052, (Jun. 23, 2011), 17 pages.
“Non-Final Office Action”, U.S. Appl. No. 11/502,264, (Sep. 14, 2012), 14 pages.
“Non-Final Office Action”, U.S. Appl. No. 12/244,545, (Mar. 27, 2012), 18 pages.
“Non-Final Office Action”, U.S. Appl. No. 12/244,545, (Aug. 17, 2011), 15 pages.
“Non-Final Office Action”, U.S. Appl. No. 12/413,977, (Jul. 19, 2011), 17 pages.
“Non-Final Office Action”, U.S. Appl. No. 12/413,977, (Jul. 20, 2012), 18 pages.
“Non-Final Office Action”, U.S. Appl. No. 12/414,382, (Jul. 26, 2011), 9 pages.
“Non-Final Office Action”, U.S. Appl. No. 12/414,434, (Jan. 17, 2012), 7 pages.
“Non-Final Office Action”, U.S. Appl. No. 12/414,434, (May 31, 2012), 7 pages.
“Non-Final Office Action”, U.S. Appl. No. 12/414,434, (Aug. 2, 2011), 6 pages.
“Non-Final Office Action”, U.S. Appl. No. 12/414,455, (Aug. 29, 2011), 8 pages.
“Non-Final Office Action”, U.S. Appl. No. 12/414,458, (Jul. 6, 2011), 8 pages.
“Non-Final Office Action”, U.S. Appl. No. 12/414,476, (Nov. 9, 2012), 22 pages.
“Non-Final Office Action”, U.S. Appl. No. 12/414,476, (Aug. 3, 2011), 21 pages.
“Non-Final Office Action”, U.S. Appl. No. 12/433,605, (Jun. 24, 2011), 10 pages.
“Non-Final Office Action”, U.S. Appl. No. 12/433,667, (Jun. 7, 2011), 15 pages.
“Non-Final Office Action”, U.S. Appl. No. 12/433,667, (Feb. 3, 2012), 16 pages.
“Non-Final Office Action”, U.S. Appl. No. 12/469,419, (Nov. 9, 2011), 15 pages.
“Non-Final Office Action”, U.S. Appl. No. 12/469,419, (May 23, 2012),1 3 pages.
“Non-Final Office Action”, U.S. Appl. No. 12/469,458, (Jul. 1, 2011), 15 pages.
“Non-Final Office Action”, U.S. Appl. No. 12/469,458, (Sep. 21, 2012), 14 pages.
“Non-Final Office Action”, U.S. Appl. No. 12/469,480, (Oct. 17, 2012), 16 pages.
“Non-Final Office Action”, U.S. Appl. No. 12/469,480, (Sep. 22, 2011), 14 pages.
“Non-Final Office Action”, U.S. Appl. No. 12/470,558, (Nov. 22, 2011), 9 pages.
“Non-Final Office Action”, U.S. Appl. No. 12/480,969, (Aug. 7, 2012), 15 pages.
“Non-Final Office Action”, U.S. Appl. No. 12/484,799, (Aug. 11, 2011), 12 pages.
“Non-Final Office Action”, U.S. Appl. No. 12/484,799, (Aug. 7, 2012), 13 pages.
“Non-Final Office Action”, U.S. Appl. No. 12/484,845, (Dec. 7, 2011), 16 pages.
“Non-Final Office Action”, U.S. Appl. No. 12/560,081, (Dec. 7, 2011), 16 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/492,495, (Sep. 17, 2012), 8 pages.
“Notice of Allowance”, U.S. Appl. No. 11/215,052, (Mar. 14, 2012), 5 pages.
“Notice of Allowance”, U.S. Appl. No. 12/414,382, (Apr. 4, 2012), 4 pages.
“Notice of Allowance”, U.S. Appl. No. 12/414,434, (Aug. 17, 2012), 4 pages.
“Notice of Allowance”, U.S. Appl. No. 12/414,455, (Jan. 4, 2012), 4 pages.
“Notice of Allowance”, U.S. Appl. No. 12/414,458, (Oct. 31, 2011) 2 pages.
“Notice of Allowance”, U.S. Appl. No. 12/414,458, (Nov. 29, 2011), 2 pages.
“Notice of Allowance”, U.S. Appl. No. 12/414,458, (Aug. 10, 2011), 6 pages.
“Notice of Allowance”, U.S. Appl. No. 12/470,558, (Apr. 2, 2012), 7 pages.
“Notice of Allowance”, U.S. Appl. No. 12/470,558, (Aug. 23, 2012), 2 pages.
“Notice of Allowance”, U.S. Appl. No. 12/484,799, (Oct. 22, 2012), 10 pages.
“Notice of Allowance”, U.S. Appl. No. 12/484,845, (Mar. 16, 2012), 5 pages.
“Oracle8i Application Developers Guide—Advanced Queuing Release 2 (8.1.6)”, Retrieved from: http://www.cs.otago.ac.nz/oradocs/appdev.817/a76938/adq0lin5.htm on May 6, 2009., (Dec. 1999), 8 pages.
“Oracle8i Application Developers Guide—Advanced Quering” Retrieved from: http://www.cs.umbc.edu/help/oracle8/server.815/a68005/03—adq1i.htm on May 6, 2009., (Feb. 1999), 29 Pages.
“Oracle8i Concepts Release 8.1.5”, Retrieved from: http://www.cs.umbc.edu/help/oracle8/server.815/a67781/c16queue.htm on May 6, 2009., (Feb. 1999), 10 Pages.
“Palette Extender 1.0.2”, Retrieved from: <http://palette-extender.en.softonic.com/symbian> on Nov. 6, 2008, (Jan. 21, 2003), 2 pages.
“Parallax Scrolling”, Retrieved from: <http://en.wikipedia.org/wiki/Parallax—scrolling> on May 5, 2009., (May 4, 2009), 3 Pages.
“PCT Search Report and Written Opinion”, Application No. PCT/US2009/061382, (May 26, 2010), 10 pages.
“PCT Search Report and Written Opinion”, Application No. PCT/US2009/061735, (Jun. 7, 2010), 11 pages.
“PCT Search Report and Written Opinion”, Application No. PCT/US2010/028699, (Oct. 4, 2010), 10 pages.
“PCT Search Report and Written Opinion”, Application No. PCT/US2010/028555, (Oct. 12, 2010), 10 pages.
“PCT Search Report and Written Opinion”, Application No. PCT/US2010/034772, (Dec. 29, 2010), 12 pages.
“PCT Search Report and Written Opinion”, Application No. PCT/US2011/055523, (May 10, 2012), 9 pages.
“PCT Search Report and Written Opinion”, Application No. PCT/US2011/055513, (Mar. 27, 2012), 8 pages.
“PCT Search Report and Written Opinion”, Application No. PCT/US2011/055512, (May 24, 2012), 8 pages.
“PCT Search Report and Written Opinion”, Application No. PCT/US2011/055524, (Jun. 1, 2012), 8 pages.
“PCT Search Report and Written Opinion”, Application No. PCT/US2011/065702, (Aug. 29, 2012), 8 pages.
“PCT Search Report and Written Opinion”, Application No. PCT/US2011/055712, (Sep. 21, 2012), 9 pages.
“PCT Search Report and Written Opinion”, Application No. PCT/US2011/055736, (Sep. 17, 2012), 8 pages.
“PCT Search Report and Written Opinion”, Application No. PCT/US2011/055496, (Sep. 12, 2012), 9 pages.
“PCT Search Report and Written Opinion”, Application No. PCT/US2011/067073, (Sep. 17, 2012), 8 pages.
“PCT Search Report and Written Opinion”, Application No. PCT/US2011/055746, (Sep. 27, 2012), 9 pages.
“PCT Search Report and Written Opinion”, Application No. PCT/US2011/055725, (Sep. 27, 2012), 10 pages.
“PCT Search Report and Written Opinion”, Application No. PCT/US2011/055478, (Sep. 27, 2012), 9 pages.
“PCT Search Report and Written Opinion”, Application No. PCT/US2011/055493, (Sep. 26, 212), 9 pages.
“PCT Search Report and Written Opinion”, PCT Application No. PCT/US2010/038730, (Jan. 19, 2011), 8 pages.
“PCT Search Report”, Application Serial No. PCT/US2009/061864, (May 14, 2010), 10 pages.
“Remapping the Keyboard”, Retrieved from: <http://publib.boulder.ibm.com/infocenter/hodhelp/v9r0/index.jsp?topic=/com.ibm.hod9.doc/help/assignkey.html> on Dec. 11, 2008., (Jul. 15, 2005), 5 Pages.
“SecureMe-Anti-Theft Security Application for S60 3rd”, Retrieved from: <http:/www.killermobile.com/newsite/mobile-software/s60-applications/secureme-%11-anti%11theft-security-application-for-s60-3rd.htm> on Jun. 28, 2011, (Dec. 15, 2008), 3 pages.
“Snap”, Windows 7 Features, retrieved from <http://windows.microsoft.com/en-US/windows7/products/features/snap> on Sep. 23, 2011, 2 pages.
“Symbian Applications”, Retrieved from: <http://symbianfullversion.blogspot.com/2008—12—01—archive.html> on May 5, 2009., (Jan. 2009), 51 Pages.
“Top 3 Task Switchers for Androd”, TechCredo, retrieved from <http://www.techcredo.com/android/top-3-task-switchers-for-android> on May 11, 2011,(Mar. 9, 2011), 5 pages.
“Top Android App: Swipepad”, Best Android Apps Review, retrieved from <http://www.bestandroidappsreview.com/2011/01/top-android-app-swipepad-launcher.html> on May 11, 2011, 4 pages.
“Touch Shell Free”, Retrieved from: <http://www.pocketpcfreeware.mobi/download-touch-shell-free.html> on May 5, 2009., (Feb. 23, 2009), 2 Pages.
“Windows Phone 7 (Push Notification)”, retrieved from <http://unknownerror.net/2011-06/windows-phone-7-push-notification-36520> on Jul. 6, 2011, 4 pages.
“Winterface Review”, Retrieved from: <http://www.mytodayscreen.com/winterface-review/> on Nov. 12, 2008, (Jul. 9, 2008), 42 pages.
“Womma”, Retrieved from: <http://www.womma.org/blog/links/wom-trends/> on May 5, 2009., (2007), 70 Pages.
“Working with Multiple Windows”, MSOFFICE tutorial!, retrieved from <http://www.msoffice-tutorial.com/working-with-multiple-windows.php> on Sep. 23, 2011, 3 pages.
Beiber, Gerald et al., “Screen Coverage: A Pen-Interaction Problem for PDA's and Touch Screen Computers”, In Proceedings of ICWMC 2007,(Mar. 2007), 6 pages.
Damien, “7 Ways to Supercharge Multitasking in Android”, retrieved from <http://maketecheasier.com/7-ways-to-supercharge-multitasking-in-android/2011/01/22/> on May 11, 2011,(Jan. 22, 2011), 5 pages.
Dolcourt, Jessica “Webware”, Retrieved from: <http://news.cnet.com/webware/?categoryId=2010> on May 5, 2009., (May 5, 2009), 13 Pages.
Gade, Lisa “Samsung Alias u740”, Retrieved from: <http://www.mobiletechreview.com/phones/Samsung-U740.htm> on Nov. 20, 2008, (Mar. 14, 2007), 6 pages.
Gao, Rui “A General Logging Service for Symbian based Mobile Phones”, Retrieved from: <http://www.nada.kth.se/utbildning/grukth/exjobb/rapportlistor/2007/rapporter07/gao—rui—07132.pdf.> on Jul. 17, 2008, (Feb. 2007), pp. 1-42.
Ha, Rick et al., “SIMKEYS: An Efficient Keypad Configuration for Mobile Communications”, Retrieved from: <http://ieeexplore.ieee.org/stamp/stamp.jsp?arnumber=01362557.> on Dec. 17, 2008, (Nov. 2004), 7 Pages.
Harrison, Richard “Symbian OS C++ for Mobile Phones vol. 3” Retrieved from: <http://—www.amazon.co.uk/Symbian-OS-Mobile-Phones-Press/dp/productdescription/0470066415> on Oct. 23, 2008, Symbian Press,(Jun. 16, 2003), 4 pages.
Hickey, Andrew R., “Google Android has Landed; T-Mobile, HTC Unveil G1”, Retrieved from: <http://www.crn.com/retail/210603348> on Nov. 26, 2008., (Sep. 23, 2008), 4 pages.
Kcholi, Avi “Windows CE .NET Interprocess Communication”, Retrieved from http://msdn.microsoft.com/en-us/library/ms836784.aspx on Jul. 17, 2008., (Jan. 2004), 15 Pages.
La, Nick “Parallax Gallery”, Available at <http://webdesignerwall.comtutorials/parallax-gallery/comment-page-1>,(Apr. 25, 2008), 16 pages.
Mann, Richard et al., “Spectrum Analysis of Motion Parallax in a 3D Cluttered Scene and Application to Egomotion”, Journal of the Optical Society of America A, vol. 22, No. 9, Available at <http://www.cs.uwaterloo.ca/˜mannr/snow/josa-mann-langer.pdf>,(Sep. 2005), pp. 1717-1731.
Mantia, Louie “Multitasking: What Does It Mean?”, retrieved from <http://mantia.me/blog/multitasking/> on Sep. 23, 2011, 3 pages.
Mao, Jeng “Comments of Verizon Wireless Messaging Services, LLC” Retrieved from: http://www.ntia.doc.gov/osmhome/warnings/comments/verizon.htm on May 6, 2009., (Aug. 18, 2000), 5 Pages.
Marie, Angelina “MacBook Trackpad Four Fingers Swipe Left/Right to Switch Applications”, MacBook Junkie, retrieved from <http://www.macbookjunkie.com/macbook-trackpad-four-fingers-swipe-left-right-to-switch-applications/> on May 11, 2011,(Nov. 13, 2010), 4 pages.
Mei, Tao et al., “Probabilistic Multimodality Fusion for Event Based Home Photo Clustering”, Retrieved from: <http://ieeexplore.ieee.org//stamp/stamp.jsp?tp=arnumber=04036960.>, (Dec. 26, 2006), pp. 1757-1760.
Nordgren, Peder “Development of a Touch Screen Interface for Scania Interactor”, Master's Thesis in Computing Science, UMEA University, Available at <http://www.cs.umu.se/education/examina/Rapporter/PederNordgren.pdf>,(Apr. 10, 2007), pp. 1-59.
Oliver, Sam “Potential iPhone Usability and Interface Improvements”, Retrieved from: <http://www.appleinsider.com/articles/08/09/18/potential—iphone—usability—and—interface—improvements.html> on Nov. 12, 2008, AppleInsider,(Sep. 18, 2008), 4 pages.
Oryl, Michael “Review: Asus P527 Smartphone for North America”, Retrieved from: <http://www.mobileburn.com/review.jsp?Id=4257> on Dec. 17, 2008., (Mar. 5, 2008), 1 Page.
Padilla, Alfredo “Palm Treo 750 Cell Phone Review—Hardware”, Retrieved from: <http://www.wirelessinfo.com/content/palm-Treo-750-Cell-Phone-Review/Hardware.htm> on Dec. 11, 2008., (Mar. 17, 2007), 4 Pages.
Raghaven, Gopal et al., “Model Based Estimation and Verification of Mobile Device Performance”, Available at http://alumni.cs.ucsb.edu/˜raimisl/emsoft04—12.pdf.,(Sep. 27-29, 2004), 10 Pages.
Reed, Brad “Microsoft Demos Windows Mobile 6.1 at CTIA”, Retrieved from: <http://www.networkworld.com/news/2008/040208-ctia-microsoft-windows-mobile.html> on Jul. 18, 2008 (Apr. 2, 2008), 1 page.
Remond, Mickael “Mobile Marketing Solutions”, Retrieved from: <http://www.mobilemarketingmagazine.co.uk/mobile—social—networking/> on May 5, 2009., (Apr. 28, 2009), 16 Pages.
Rice, Stephen V., et al., “A System for Searching Sound Palettes”, Proceedings of the Eleventh Biennial Symposium on Arts and Technology,, Available at <http://www.comparisonics.com/FindSoundsPalettePaper.pdf>,(Feb. 2008), 6 pages.
Roberts, Neil “Touching and Gesturing on the iPhone”, Available at <http://www.sitepen.com/blog/2008/07/10/touching-and-gesturing-on-the-iphone/comments-pare-1>,(Jul. 10, 2008), 16 pages.
Singh, Kundan et al., “CINEMA: Columbia InterNet Extensible Multimedia Architecture”, Available at <http://www1.cs.columbia.edu/˜library/TR-repository/reports/reports-2002/cucs-011-02.pdf>,(Sep. 3, 2002), 83 Pages.
Steinicke, Frank et al., “Multi-Touching 3D Data: Towards Direct Interaction in Stereoscopic Display Environments coupled with Mobile Devices”, Advanced Visual Interfaces (AVI) Workshop on Designing Multi-Touch Interaction Techniques for Coupled Public, Available at <http://viscg.uni-muenster.de/publications/2008/SHSK08/ppd-workshop.-pdf.>,(Jun. 15, 2008), 4 Pages.
Suror, “PocketShield—New Screenlock App for the HTC Diamond and Pro”, Retrieved from: <http://wmpoweruser.com/?tag=htc-touch-diamond> on Jun. 28, 2011, (Oct. 23, 2008), 2 pages.
Terpstra, Brett “Beta Beat: Grape, a New Way to Manage Your Desktop Clutter”, Retrieved from: Beta Beat: Grape, a New Way to Manage Your Desktop Clutter on Jun. 28, 2011, (Apr. 14, 2009), 4 pages.
Vallerio, Keith S., et al., “Energy-Efficient Graphical User Interface Design”, Retrieved from: <http://www.cc.gatech.edu/classes/AY2007/cs7470—fall/zhong-energy-efficient-user-interface.pdf>, (Jun. 10, 2004), pp. 1-13.
Vermeulen, Jan “BlackBerry PlayBook Hands-on”, retrieved from <http://mybroadband.co.za/news/gadgets/20104-BlackBerry-PlayBook-hands-.html> on May 11, 2011,(May 8, 2011), 4 pages.
Viticci, Federico “Growl 1.3 to Be Released on Mac App Store, Introduce Lion Support and Drop GrowlMail Support”, Retrieved from: <http://www.macstories.net/stories/growl-1-3-to-be-released-on-mac-app-store-introduce-lion-support-and-drop-growlmail-support/> on Jul. 22, 2011,(Jul. 6, 2011), 6 pages.
Wilson, Tracy V., “How the iPhone Works”, Retrieved from: <http://electronics.howstuffworks.com/iphone2.htm> on Apr. 24, 2009, (Jan. 2007), 9 pages.
Wobbrock, Jacob O., et al., “User-Defined Gestures for Surface Computing”, CHI 2009, Apr. 4-9, 2009, Boston, MA, available at <http://faculty.washington.edu/wobbrock/pubs/chi-09.2.pdf>,(Apr. 4, 2009), 10 pages.
Wyatt, Paul “/Flash/the art of parallax scrolling”, .net Magazine,(Aug. 1, 2007), pp. 74-76.
Yang, Seungji et al., “Semantic Photo Album Based on MPEG-4 Compatible Application Format”, Retrieved from: <http://ieeexplore.ieee.org/stamp/stamp.jsp?arnumber=04146254.>, (2007), 2 Pages.
“Final Office Action”, U.S. Appl. No. 11/502,264, (Mar. 29, 2013),16 pages.
“My Favorite Gadgets, System Monitor II”, Retrieved from <http://www.myfavoritegadgets.info/monitors/SystemMonitorII/systemmonitorII.html> on Mar. 12, 2013, (Jun. 8, 2010), 5 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/118,204, (Feb. 28, 2013), 13 pages.
“Non-Final Office Action”, U.S. Appl. No. No. 13/118,257, (Mar. 5, 2013), 19 pages.
“Non-Final Office Action”, U.S. Appl. No. No. 13/118,339, (Feb. 11, 2013), 15 pages.
“Non-Final Office Action”, U.S. Appl. No. No. 13/118,347, (Feb. 12, 2013), 14 pages.
“Non-Final Office Action”, U.S. Appl. No. No. 13/229,693, (Mar. 12, 2013), 21 pages.
“Notice of Allowance”, U.S. Appl. No. 13/492,495, (Apr. 26, 2013), 5 pages.
“PCT Search Report and Written Opinion”, Application No. PCT/US2012/047091, (Dec. 27, 2012), 15 pages.
Bruzzese, J. P., “Using Windows 7, Managing and Monitoring Windows 7—Chapter 11”, Que Publishing, (May 5, 2010), 33 pages.
Gralla, Preston “Windows XP Hacks, Chapter 13—Hardware Hacks”, O'Reilly Publishing, (Feb. 23, 2005), 25 pages.
Horowitz, Michael “Installing and Tweaking Process Explorer part 2”, Retrieved <http://web.archive.org/web/20110510093838/http://blogs.computerworld.com/16165/installing—and—tweaking—process—explorer—part—2> on Mar. 12, 2013, (May 23, 2010), 7 pages.
“Non-Final Office Action”, U.S. Appl. No. 12/972,967, (Jan. 30, 2013), 19 pages.
“Non-Final Office Action”, U.S. Appl. No. 12/977,584, (Dec. 7, 2012), 8 pages.
“Non-Final Office Action”, U.S. Appl. No. 12/978,184, (Jan. 23, 2013), 7 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/196,272, (Feb. 6, 2013), 10 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/224,258, (Jan. 8, 2013), 35 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/492,495, (Dec. 19, 2012), 6 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/655,386, (Dec. 26, 2012), 23 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/656,354, (Feb. 6, 2013), 10 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/656,574, (Jan. 31, 2013), 21 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/657,621, (Feb. 7, 2013), 19 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/657,646, (Jan. 3, 2013), 13 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/657,789, (Jan. 9, 2013), 38 pages.
“PCT Search Report and Written Opinion”, Application No. PCT/US2011/067075, (Dec. 12, 2012), 10 pages.
Livingston, et al., “Windows 95 Secrets”, 1995, I DG Books Worldwide, 3rd Edition, (1995), pp. 121-127.
Perry, Greg “Teach Yourself Windows 95 in 24 Hours”, 1997, Sams Publishing, 2nd Edition, (pp. 193-198), 1997.
“Final Office Action”, U.S. Appl. No. 13/655,386, (Jun. 6, 2013), 34 pages.
“Final Office Action”, U.S. Appl. No. 13/656,354, (Jun. 17, 2013), 14 pages.
“Final Office Action”, U.S. Appl. No. 13/657,646, (May 6, 2013), 12 pages.
“Final Office Action”, U.S. Appl. No. 13/657,789, (Jun. 21, 2013), 35 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/073,300, (Jul. 25, 2013), 13 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/118,321, (Jun. 10, 2013), 32 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/118,333, (Jul. 5, 2013), 18 pages.
“Notice of Allowance”, U.S. Appl. No. 12/977,584, (Jun. 19, 2013), 5 pages.
“Notice of Allowance”, U.S. Appl. No. 13/655,390, (May 24, 2013), 5 pages.
“Supplemental Notice of Allowance”, U.S. Appl. No. 13/655,390, (Jul. 25, 2013), 2 pages.
Kurdi, Samer “Acer GridVista: snap your windows to pre-defined sections on your screen(s)”, Retrieved from <http://www.freewaredenius.com/acer-gridvista-snap-your-windows-to-pre-defined-sections-of-your-screens/> on Jun. 30, 2013, (Jan. 19, 2010), 6 pages.
Kurdi, Samer “WinSplit Revolution”, Retrieved from <http://www.freewaregenius.com/winsplit-revolution/> on Jun. 30, 2013, (Aug. 22, 2007), 4 Pages.
“Final Office Action”, U.S. Appl. No. 12/721,422, (Mar. 7, 2013), 10 pages.
“Final Office Action”, U.S. Appl. No. 13/118,339, (Aug. 22, 2013), 21 pages.
“Final Office Action”, U.S. Appl. No. 13/118,347, (Aug. 15, 2013), 25 pages.
“Final Office Action”, U.S. Appl. No. 13/656,574, (Aug. 23, 2013), 20 pages.
“Non-Final Office Action”, U.S. Appl. No. 12/721,422, (Oct. 1, 2012), 7 pages.
“Notice of Allowance”, U.S. Appl. No. 12/721,422, (Jul. 11, 2013), 9 pages.
“Notice of Allowance”, U.S. Appl. No. 12/978,184, (Aug. 2, 2013), 5 pages.
“You've Got Mail 1.4 Build”, retrieved from <http://www.fileshome.com/Shows—Animation—Plays—Sound—Automatic—N . . . > on Jan. 6, 2010 (Jun. 18, 2007), 2 pages.
Farrugia, Michael et al., “Cell Phone Mini Challenge: Node-Link Animation Award Animating Multivariate Dynamic Social Networks”, IEEE Symposium on Visual Analytics Science and Technology, Columbus, OH, USA, Oct. 21-23, 2008, (Oct. 21, 2008), 2 pages.
Keranen, Jaakko “OpenGL-based User Interface Toolkit for Symbian Mobile Devices”, Master of Science Thesis, Tamere University of Technology, Department of Information Technology, (Apr. 6, 2005), 88 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/228,888, Feb. 10, 2014, 21 pages.
“Foreign Office Action”, CN Application No. 201110437572.2, Dec. 3, 2013, 7 pages.
“Foreign Office Action”, CN Application No. 201110429183.5, Jan. 6, 2014, 10 Pages.
“Foreign Office Action”, CN Application No. 201110437542.1, Jan. 6, 2014, 10 Pages.
“Foreign Office Action”, CN Application No. 201110454251.3, Dec. 27, 2013, 12 Pages.
“Restriction Requirement”, U.S. Appl. No. 13/118,265, Feb. 27, 2014, 7 pages.
“Restriction Requirement”, U.S. Appl. No. 13/118,288, Mar. 4, 2014, 7 pages.
“Supplemental Notice of Allowance”, U.S. Appl. No. 12/978,184, Feb. 25, 2014, 2 pages.
“Final Office Action”, U.S. Appl. No. 13/073,300, Apr. 1, 2014, 14 pages.
“Final Office Action”, U.S. Appl. No. 13/118,333, Apr. 23, 2014, 22 pages.
“Final Office Action”, U.S. Appl. No. 13/228,707, May 21, 2014, 16 pages.
“Final Office Action”, U.S. Appl. No. 13/229,155, Jun. 12, 2014, 15 pages.
“Gestures Programming”, Retrieved from <http://doc.qt.digia.com/4.6/gestures-overview.html> on May 28, 2014, 2010, 3 pages.
“Image Gestures Example”, Retrieved from <http://doc.qt.digia.com/4.6/gestures-imagegestures.html> on May 28, 2014, 2010, 3 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/118,265, Jun. 10, 2014, 13 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/118,292, Jun. 6, 2014, 25 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/228,931, Apr. 7, 2014, 21 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/228,945, Apr. 14, 2014, 22 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/229,556, Mar. 28, 2014, 12 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/229,709, Apr. 7, 2014, 12 pages.
“Normalizing Text: A Java Tutorial by Oracle”, Retrieved from: <http://docs.oracle.com/javase/tutorial/i18n/text/normalizerapi.html> on Apr. 8, 2014, Nov. 11, 2006, 3 pages.
“Notice of Allowance”, U.S. Appl. No. 13/655,386, Apr. 25, 2014, 6 pages.
“QPinchGesture Class Reference”, Retrieved from <http://doc.qt.digia.com/4.6/qpinchgesture.html> on May 28, 2014, 2010, 6 pages.
Anson, “Pining for Windows Phone 7 controls? We got ya covered! [Announcing the first release of the Silverlight for Windows Phone Toolkit!]”, Retrieved from <http://blogs.msdn.com/b/delay/archive/2010/09/16/pining-for-windows-phone-7-controls-we-got-ya-covered-announcing-the-first-release-of-the-silverlight-for-windows-phone-toolkit.aspx> on May 30, 2014, Sep. 16, 2010, 17 pages.
“Final Office Action”, U.S. Appl. No. 13/224,258, Jul. 18, 2014, 39 pages.
“Final Office Action”, U.S. Appl. No. 13/228,876, Jul. 18, 2014, 15 pages.
“First Examination Report”, NZ Application No. 618269, May 20, 2014, 2 pages.
“First Examination Report”, NZ Application No. 618284, May 20, 2014, 2 pages.
“Foreign Office Action”, CN Application No. 201180071186.4, Jun. 13, 2014, 12 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/118,288, Jul. 2, 2014, 17 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/229,693, Jun. 20, 2014, 25 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/229,702, Jul. 3, 2014, 28 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/657,621, Jul. 18, 2014, 22 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/657,646, Aug. 12, 2014, 14 pages.
“Notice of Allowance”, U.S. Appl. No. 13/118,204, Jul. 8, 2014, 8 pages.
“Notice of Allowance”, U.S. Appl. No. 13/657,789, Aug. 4, 2014, 16 pages.
“Final Office Action”, U.S. Appl. No. 13/657,621, Sep. 10, 2013, 18 pages.
“Final Office Action”, U.S. Appl. No. 12/972,967, Oct. 11, 2013, 21 pages.
“Final Office Action”, U.S. Appl. No. 13/229,693, Sep. 4, 2013, 23 pages.
“Final Office Action”, U.S. Appl. No. 13/118,204, Nov. 21, 2013, 24 pages.
“Final Office Action”, U.S. Appl. No. 13/118,321, Dec. 19, 2013, 30 pages.
“Final Office Action”, U.S. Appl. No. 13/224,258, Sep. 11, 2013, 37 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/228,707, Oct. 25, 2013, 12 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/655,390, Dec. 17, 2012, 12 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/229,155, Nov. 18, 2013, 14 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/228,876, Nov. 22, 2013, 14 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/196,272, Sep. 3, 2013, 5 pages.
“Notice of Allowance”, U.S. Appl. No. 12/978,184, Nov. 6, 2013, 5 pages.
“Notice of Allowance”, U.S. Appl. No. 13/196,272, Nov. 8, 2013, 8 pages.
“Supplemental Notice of Allowance”, U.S. Appl. No. 12/977,584, Sep. 16, 2013, 2 pages.
“Supplemental Notice of Allowance”, U.S. Appl. No. 13/655,390, Sep. 19, 2013, 2 pages.
“Supplemental Notice of Allowance”, U.S. Appl. No. 12/977,584, Oct. 11, 2013, 2 pages.
“Final Office Action”, U.S. Appl. No. 13/228,876, Jun. 19, 2015, 20 pages.
“Final Office Action”, U.S. Appl. No. 13/229,155, Jun. 4, 2015, 16 pages.
“Foreign Office Action”, CL Application No. 3368-2013, May 7, 2015, 7 pages.
“Foreign Office Action”, CL Application No. 3370-2013, May 7, 2015, 6 pages.
“Foreign Office Action”, CN Application No. 201180071186.4, Jun. 2, 2015, 7 pages.
“Notice of Allowance”, U.S. Appl. No. 14/109,779, Jun. 17, 2015, 4 pages.
“Authoritative Dictionary of IEEE Standards Terms, 7th ed.”, Definitions—processor, memory, and storage, 2000, 3 pages.
“Corrected Notice of Allowance”, U.S. Appl. No. 13/118,333, Jun. 1, 2015, 4 pages.
“Extended European Search Report”, EP Application No. 11871863.4, May 11, 2015, 8 pages.
“Extended European Search Report”, EP Application No. 11871917.8, May 11, 2015, 9 pages.
“Extended European Search Report”, EP Application No. 11872137.2, Apr. 9, 2015, 12 pages.
“Final Office Action”, U.S. Appl. No. 13/228,707, Jun. 1, 2015, 20 pages.
“Foreign Notice of Allowance”, CN Application No. 201110429183.5, Mar. 9, 2015, 4 Pages.
“Foreign Office Action”, CN Application No. 201210331158.8, May 11, 2015, 7 pages.
“Foreign Office Action”, CN Application No. 201210331670.2, Mar. 25, 2015, 14 pages.
“Foreign Office Action”, CN Application No. 201210331686.3, Mar. 3, 2015, 14 pages.
“Foreign Office Action”, CO Application No. 13300256, Apr. 11, 2015, 8 Pages.
“Foreign Office Action”, CO Application No. 13300265, Apr. 21, 2015, 11 Pages.
“Foreign Office Action”, PH Application No. PH/1/2013/502367, Apr. 24, 2015, 3 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/073,300, Apr. 30, 2015, 15 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/229,693, Apr. 23, 2015, 28 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/550,432, Apr. 27, 2015, 15 pages.
“Notice of Allowance”, U.S. Appl. No. 13/118,333, May 4, 2015, 16 pages.
“Notice of Allowance”, U.S. Appl. No. 13/229,702, Apr. 29, 2015, 7 pages.
“Notice of Allowance”, U.S. Appl. No. 13/345,383, May 18, 2015, 5 pages.
“Rename a file”, Retrieved from <http://windows.microsoft.com/en-us/windows7/rename-a-file> on Apr. 27, 2015, Aug. 24, 2009, 1 page.
Bederson,“Implementing a Zooming User Interface: Experience Building Pad++”, Software Practice & Experience, Wiley & Sons, Aug. 1998, 34 pages.
Bederson,“Jazz: An Extensible Zoomable User Interface Graphics Toolkit in Java”, Proceedings of the 2000 ACM SIGCPR Conference, Apr. 2000, 11 pages.
Jetter,“Materializing the Query with Facet-Streams—A Hybrid Surface for Collaborative Search on Tabletops”, May 7, 2011, 10 pages.
“Corrected Notice of Allowance”, U.S. Appl. No. 13/656,574, Jan. 13, 2015, 4 pages.
“Corrected Notice of Allowance”, U.S. Appl. No. 13/656,574, Dec. 12, 2014, 4 pages.
“Extended European Search Report”, EP Application No. 11866553.8, Dec. 9, 2014, 7 pages.
“Extended European Search Report”, EP Application No. 11866579.3, Oct. 9, 2014, 7 pages.
“Extended European Search Report”, EP Application No. 11867033.0, Nov. 27, 2014, 8 pages.
“Final Office Action”, U.S. Appl. No. 13/118,265, Nov. 6, 2014, 19 pages.
“Final Office Action”, U.S. Appl. No. 13/118,288, Nov. 7, 2014, 14 pages.
“Final Office Action”, U.S. Appl. No. 13/118,292, Dec. 5, 2014, 30 pages.
“Final Office Action”, U.S. Appl. No. 13/228,888, Oct. 24, 2014, 29 pages.
“Final Office Action”, U.S. Appl. No. 13/228,931, Dec. 19, 2014, 27 pages.
“Final Office Action”, U.S. Appl. No. 13/228,945, Oct. 23, 2014, 24 pages.
“Final Office Action”, U.S. Appl. No. 13/229,693, Jan. 6, 2015, 24 pages.
“Final Office Action”, U.S. Appl. No. 13/229,702, Jan. 15, 2015, 24 pages.
“Final Office Action”, U.S. Appl. No. 13/345,383, Jul. 25, 2014, 26 pages.
“Final Office Action”, U.S. Appl. No. 13/657,621, Jan. 15, 2015, 22 pages.
“Foreign Notice of Acceptance”, NZ Application No. 618269, Oct. 31, 2014, 1 Page.
“Foreign Office Action”, CN Application No. 201110429183.5, Aug. 21, 2014, 13 Pages.
“Foreign Office Action”, CN Application No. 201110454251.3, Oct. 21, 2014, 13 Pages.
“Foreign Office Action”, CN Application No. 201180071196.8, Sep. 11, 2014, 13 pages.
“Foreign Office Action”, CN Application No. 201210317470.1, Jun. 5, 2014, 12 pages.
“Foreign Office Action”, CN Application No. 201210331158.8, Sep. 3, 2014, 16 Pages.
“Foreign Office Action”, CN Application No. 201210331686.3, Jun. 13, 2014, 13 pages.
“Foreign Office Action”, CO Application No. 13300256, Sep. 24, 2014, 8 Pages.
“Foreign Office Action”, CO Application No. 13300265, Sep. 24, 2014, 10 Pages.
“Guidelines for Panning”, Retrieved From: <http://msdn.microsoft.com/en-in/library/windows/apps/hh465310.aspx> Aug. 19, 2014, Dec. 9, 2012, 5 Pages.
“MS-Content-Zoom-Snap-Points Property”, Retrieved From: <http://msdn.microsoft.com/en-us/library/windows/apps/hh441259.aspx> Aug. 22, 2014, 2 Pages.
“-MS-Scroll-Snap-Type Property”, Retrieved From: <http://msdn.microsoft.com/en-in/library/windows/apps/hh466057.aspx> Aug. 19, 2014, 1 Page.
“Non-Final Office Action”, U.S. Appl. No. 12/972,967, Nov. 6, 2014, 17 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/118,321, Nov. 12, 2014, 33 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/118,339, Dec. 10, 2014, 16 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/118,347, Dec. 3, 2014, 28 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/228,707, Nov. 13, 2014, 20 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/228,876, Jan. 29, 2015, 18 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/229,155, Jan. 7, 2015, 14 pages.
“Non-Final Office Action”, U.S. Appl. No. 14/089,149, Nov. 3, 2014, 8 pages.
“Non-Final Office Action”, U.S. Appl. No. 14/109,779, Nov. 21, 2014, 9 pages.
“Notice of Allowance”, U.S. Appl. No. 13/229,556, Sep. 2, 2014, 12 pages.
“Notice of Allowance”, U.S. Appl. No. 13/229,709, Sep. 2, 2014, 4 pages.
“Notice of Allowance”, U.S. Appl. No. 13/656,574, Sep. 23, 2014, 7 pages.
“Notice of Allowance”, U.S. Appl. No. 14/089,149, Dec. 5, 2014, 4 pages.
“Supplemental Notice of Allowance”, U.S. Appl. No. 13/118,204, Sep. 10, 2014, 2 pages.
“Supplemental Notice of Allowance”, U.S. Appl. No. 13/118,204, Oct. 2, 2014, 2 pages.
“Supplemental Notice of Allowance”, U.S. Appl. No. 13/118,204, Oct. 23, 2014, 3 pages.
Camick, “Wrap Layout”, Java Tips Weblog, retrieved from <https://tips4java.wordpress.com/2008/11/06/wrap-layout/> on Jan. 6, 2015, Nov. 6, 2008, 28 pages.
Foley, “The JavaScript Behind Touch-Friendly Sliders”, Retrieved From: <http://css-tricks.com/the-javascript-behind-touch-friendly-sliders/> Aug. 19, 2014, Jun. 13, 2013, 14 Pages.
Pointet, “Scrollsnap”, Retrieved From: <http://benoit.pointet.info/stuff/jquery-scrollsnap-plugin/> Aug. 19, 2014, Jun. 29, 2013, 3 Pages.
Rakow, et al., “CSS Scroll Snap Points Module Level 1”, Retrieved From: <http://dev.w3.org/csswg/css-snappoints/> Aug. 19, 2014, Mar. 5, 2014, 18 Pages.
Rossi, et al., “Enabling New Interoperable Panning Experiences Through the CSS Scrolling Snap Points Specification”, Retrieved From: <http://blogs.msdn.com/b/ie/archive/2013/10/22/enabling-new-interoperable-panning-experiences-through-the-css-scrolling-snap-points-specification.aspx> Aug. 22, 2014, Oct. 22, 2013, 4 Pages.
“Setting Native-Like Scrolling Offsets in CSS with Scrolling Snap Points”, Retrieved From: <http://generatedcontent.org/post/66817675443/setting-native-like-scrolling-offsets-in-css-with> Aug. 19, 2014, Nov. 18, 2013, 9 pages.
Webmonkey “HTML Cheatsheet”, Retrieved From: <http://www.webmonkey.com/2010/02/html—cheatsheet> on Nov. 7, 2014, Feb. 15, 2010, 5 pages.
“Corrected Notice of Allowance”, U.S. Appl. No. 14/089,149, Feb. 20, 2015, 2 pages.
“Foreign Office Action”, CN Application No. 201180071186.4, Jan. 20, 2015, 8 pages.
“Foreign Office Action”, CN Application No. 201210331188.9, Oct. 10, 2014, 12 pages.
“Foreign Office Action”, CN Application No. 201210331564.4, Dec. 3, 2014, 12 pages.
“Foreign Office Action”, CN Application No. 201210331584.1, Nov. 3, 2014, 14 pages.
“Notice of Allowance”, U.S. Appl. No. 13/657,646, Feb. 6, 2015, 15 pages.
“Supplemental Notice of Allowance”, U.S. Appl. No. 13/656,574, Feb. 25, 2015, 4 pages.
“Extended European Search Report”, EP Application No. 11866699.9, Jan. 15, 2015, 6 pages.
“Extended European Search Report”, EP Application No. 11866772.4, Jan. 15, 2015, 6 pages.
“Final Office Action”, U.S. Appl. No. 13/118,321, Apr. 2, 2015, 30 pages.
“Foreign Notice of Allowance”, CN Application No. 201110437572.2, Mar. 3, 2015, 3 Pages.
“Foreign Office Action”, CN Application No. 201110437542.1, Jan. 28, 2015, 7 pages.
“Foreign Office Action”, CN Application No. 201110437542.1, Aug. 20, 2014, 8 pages.
“Foreign Office Action”, CN Application No. 201110454251.3, Feb. 28, 2015, 9 Pages.
“Foreign Office Action”, CN Application No. 201210317470.1, Feb. 11, 2015, 9 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/345,383, Jan. 29, 2015, 5 pages.
“Notice of Allowance”, U.S. Appl. No. 13/118,339, Mar. 31, 2015, 7 pages.
“Notice of Allowance”, U.S. Appl. No. 13/118,347, Apr. 1, 2015, 7 pages.
“Supplemental Notice of Allowance”, U.S. Appl. No. 14/089,149, Mar. 20, 2015, 2 pages.
“Corrected Notice of Allowance”, U.S. Appl. No. 14/659,442, Nov. 27, 2015, 2 pages.
“Extended European Search Report”, EP Application No. 11872164.6, Sep. 18, 2015, 8 pages.
“Final Office Action”, U.S. Appl. No. 14/059,163, Dec. 3, 2015, 6 pages.
“Foreign Office Action”, CL Application No. 3370-2013, Oct. 29, 2015, 6 pages.
“Foreign Office Action”, JP Application No. 2014-512824, Nov. 6, 2015, 5 pages.
“Foreign Office Action”, JP Application No. 2014-529684, Nov. 4, 2015, 10 pages.
“Foreign Office Action”, JP Application No. 2014-529685, Nov. 4, 2015, 5 pages.
“Foreign Office Action”, JP Application No. 2014-529687, Oct. 20, 2015, 8 pages.
“Foreign Office Action”, RU Application No. 2013152630, Oct. 26, 2015, 5 pages.
“Foreign Office Action”, RU Application No. 2013152635, Oct. 28, 2015, 7 pages.
“Foreign Office Action”, RU Application No. 2014108844, Oct. 27, 2015, 6 pages.
“Foreign Office Action”, RU Application No. 2014108997, Oct. 14, 2015, 6 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/118,292, Dec. 7, 2015, 32 pages.
Sakumi,“Browse/Creation Tool for Hierarchy Contents with Concept Pad System: Zoomable User Interface”, In Lecture Note/Software Science 23 Interactive System and Software VIII, Dec. 20, 1999, 8 pages.
Stein,“Growing Beautiful Code in BioPerl”, In Beautiful Code—Leading Programmers Explain How They Think, Oreilly, 2007, 32 pages.
Toshiyuki,“User Interaction Technique for Mobile Terminal (First Part)—Display Interaction Technique—Information Processing”, In Information Processing Society of Japan, vol. 48, No. 6, Jun. 15, 2007, 12 pages.
“Final Office Action”, JP Application No. 2014-529685, Apr. 5, 2016, 6 pages.
“Foreign Notice of Allowance”, CN Application No. 201210317470.1, Feb. 23, 2016, 4 pages.
“Foreign Notice of Allowance”, CN Application No. 201210331584.1, Feb. 14, 2016, 4 pages.
“Foreign Notice of Allowance”, JP Application No. 2014-512819, Jan. 26, 2016, 4 pages.
“Foreign Office Action”, EP Application No. 11866579.3, Apr. 5, 2016, 4 pages.
“Foreign Office Action”, EP Application No. 11866699.9, Feb. 3, 2016, 4 pages.
“iOS Human Interface Guidelines”, Apple Inc, Retrieved at<<https://itunes.apple.com/in/book/ios-human-interface-guidelines/id877942287?mt=11>>, Mar. 23, 2011, 2 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/228,707, Apr. 19, 2016, 26 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/228,931, Mar. 24, 2016, 22 pages.
“Non-Final Office Action”, U.S. Appl. No. 14/522,420, Apr. 22, 2016, 21 pages.
“Notice of Allowance”, U.S. Appl. No. 14/059,163, Apr. 25, 2016, 7 pages.
Budiu,“Usability of iPad Apps and Websites”, Retrieved at<<https://tdougher.expressions.syr.edu/wrt307fall12/files/2012/08/ipad-usability—report—1st-edition-1.pdf>>, Dec. 31, 2010, 98 pages.
“Corrected Notice of Allowance”, U.S. Appl. No. 13/118,333, Aug. 4, 2015, 4 pages.
“Corrected Notice of Allowance”, U.S. Appl. No. 14/109,779, Jul. 21, 2015, 2 pages.
“Extended European Search Report”, EP Application No. 11872072.1, Jul. 28, 2015, 7 pages.
“Final Office Action”, U.S. Appl. No. 13/073,300, Sep. 4, 2015, 17 pages.
“Final Office Action”, U.S. Appl. No. 13/224,258, Aug. 13, 2015, 39 pages.
“Foreign Notice of Allowance”, CN Application No. 201110437542.1, Aug. 3, 2015, 4 Pages.
“Further Examination Report”, NZ Application No. 618284, Jul. 13, 2015, 2 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/118,265, Aug. 20, 2015, 18 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/118,288, Aug. 24, 2015, 19 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/118,321, Sep. 8, 2015, 34 pages.
“Non-Final Office Action”, U.S. Appl. No. 14/059,163, Aug. 10, 2015, 5 pages.
“Non-Final Office Action”, U.S. Appl. No. 14/659,442, Jul. 7, 2015, 7 pages.
“Notice of Allowance”, U.S. Appl. No. 14/659,442, Aug. 19, 2015, 4 pages.
Gladisch,“MultiStates: Monitoring Databases With Acoustic and Intuitive Perspective Wall Interaction”, In Proceedings of 2nd International Workshop on Design & Evaluation of e-Government Applications and Services, Aug. 24, 2009, 7 pages.
Johnson,“Create an Awesome Zooming Web Page With jQuery”, Retrieved from <http://designshack.net/articles/javascript/create-an-awesome-zooming-web-page-with-jquery/> on Aug. 20, 2015, May 25, 2011, 11 pages.
“Final Office Action”, U.S. Appl. No. 13/657,621, Jan. 21, 2016, 21 pages.
“Foreign Office Action”, CN Application No. 201180071183.0, Jan. 28, 2016, 18 pages.
“Foreign Office Action”, EP Application No. 11866772.4, Feb. 2, 2016, 5 pages.
“Foreign Office Action”, EP Application No. 1866699.9, Feb. 3, 2016, 4 pages.
“International Search Report and Written Opinion”, Application No. PCT/US2015/048751, Nov. 27, 2015, 12 pages.
“Notice of Allowance”, U.S. Appl. No. 13/073,300, Feb. 12, 2016, 9 pages.
“Corrected Notice of Allowance”, U.S. Appl. No. 13/118,333, Sep. 14, 2015, 4 pages.
“Corrected Notice of Allowance”, U.S. Appl. No. 14/109,779, Oct. 19, 2015, 2 pages.
“Corrected Notice of Allowance”, U.S. Appl. No. 14/659,442, Nov. 10, 2015, 2 pages.
“Extended European Search Report”, Application No. 11871739.6, Oct. 5, 2015, 13 pages.
“Final Office Action”, U.S. Appl. No. 13/550,432, Sep. 14, 2015, 18 pages.
“Foreign Notice of Allowance”, CN Application No. 201210331158.8, Sep. 8, 2015, 3 Pages.
“Foreign Office Action”, CN Application No. 201110454251.3, Sep. 30, 2015, 10 Pages.
“Foreign Office Action”, CN Application No. 201210317470.1, Aug. 24, 2015, 9 pages.
“Foreign Office Action”, CN Application No. 201210331564.4, Sep. 2, 2015, 14 pages.
“Foreign Office Action”, CN Application No. 201210331584.1, Aug. 19, 2015, 7 pages.
“Foreign Office Action”, CN Application No. 201210331686.3, Sep. 1, 2015, 11 pages.
“Foreign Office Action”, EP Application No. 11866579.3, Sep. 3, 2015, 4 pages.
“Foreign Office Action”, JP Application No. 2014-512819, Sep. 29, 2015, 4 pages.
“Foreign Office Action”, JP Application No. 2014-528371, Sep. 29, 2015, 16 pages.
“Foreign Office Action”, JP Application No. 2014-529671, Sep. 29, 2015, 16 pages.
“Foreign Office Action”, JP Application No. 2014-529672, Oct. 6, 2015, 7 pages.
“Foreign Office Action”, RU Application No. 2014107906, Oct. 6, 2015, 6 pages.
“Foreign Office Action”, RU Application No. 2014108874, Oct. 6, 2015, 6 pages.
“Introduction to Windows 7”, https://www.google.com/url?q=http://www.wright.edu/sites/default/files/page/attachements/windows7.pdf, Jul. 1, 2010, 13 pages.
“Notice of Allowance”, U.S. Appl. No. 13/229,693, Sep. 14, 2015, 13 pages.
“Samsung Galaxy 2 TouchWiz 4.0 Tour (Homescreen, Settings, etc.)”, https://www.youtube.com/watch?v=oXBbiy0Adiw, May 9, 2011, 3 pages.
“Search Report”, TW Application No. 100136568, Sep. 14, 2015, 2 pages.
D,“Android Apps in Depth—02—Go Launcher EX”, https://www.youtube.com/watch?v=u5LISE8BU—E, Mar. 6, 2011, 3 pages.
D,“Android HTC EVO 4G Tutorials 01—Getting Started, Home Screens”, https://www.youtube.com/watch?v=fwvt-rsCMA8, Feb. 3, 2011, 3 pages.
Demers,“In Depth: GO Launcher EX—Droid Life”, http://www.droid-life.com/2011/01/26/in-depth-go-launcher-ex/, Jan. 26, 2011, 8 pages.
Takahiro,“Let's Manage the Installed Applications”, In Easy to Use Mini, Exhaustion Utilization of iPhone, vol. 1, p. 25 Memo, Mar. 10, 2011, 6 pages.
“Foreign Notice of Allowance”, CN Application No. 201210331670.2, Apr. 26, 2016, 4 pages.
“Foreign Notice of Allowance”, RU Application No. 2014107906, Apr. 25, 2016, 18 pages.
“Foreign Office Action”, CN Application No. 201110454251.3, May 27, 2016, 6 Pages.
“Foreign Office Action”, CN Application No. 201210085754.2, Apr. 5, 2016, 16 pages.
“Foreign Office Action”, JP Application No. 2014-512824, Jun. 7, 2016, 6 pages.
“Foreign Office Action”, PH Application No. 1-2013-502367, Jun. 15, 2016, 2 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/228,876, May 12, 2016, 20 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/550,432, Jun. 16, 2016, 18 pages.
“Non-Final Office Action”, U.S. Appl. No. 14/850,347, Jun. 16, 2016, 26 pages.
“Foreign Notice of Allowance”, RU Application No. 2014108997, May 26, 2016, 17 pages.
“Foreign Notice of Allowance”, JP Application No. 2014-529687, May 31, 2016, 4 pages.
“Foreign Office Action”, JP Application No. 2014-528371, Jun. 7, 2016, 5 pages.
“Foreign Office Action”, CN Application No. 201210331670.2, Dec. 1, 2015, 14 pages.
Related Publications (1)
Number Date Country
20120174005 A1 Jul 2012 US