Classifying the intent of user input

Information

  • Patent Grant
  • 9946307
  • Patent Number
    9,946,307
  • Date Filed
    Monday, May 14, 2012
    11 years ago
  • Date Issued
    Tuesday, April 17, 2018
    6 years ago
Abstract
Different types of user inputs can be input by a user via a keyboard of an input device. These different types of user inputs include, for example, key strikes, multi-touch interactions, single finger motions, and/or mouse clicks. Touch information regarding the pressure applied to the keys of a pressure sensitive keyboard over time (or the contact area of the user input for other types of keyboards over time) is used to classify the intent of the user input as one of the various types of user inputs.
Description
BACKGROUND

Computing devices oftentimes have multiple different types of input mechanisms. These input mechanisms can include, for example, keyboards, virtual keyboards, mice, track pads, and so forth. Although these different types of mechanisms provide multiple input options for the user, they are not without their problems. One such problem is that given the number and different types of input mechanisms, it can be difficult for a user to manage the different input mechanisms, particularly in a mobile setting.


SUMMARY

Classifying the intent of user input techniques are described.


In one or more implementations, touch information regarding a user input to an input device is obtained. Based on this touch information, an intent of the user input is classified as being either a key strike or one or more other types of input.


In one or more implementations, an input device is configured to provide an output that indicates touch information regarding a user input. This touch information is usable by one or more modules to determine a user intent in providing the user input.


This Summary is provided to introduce a selection of concepts in a simplified form that are further described below in the Detailed Description. This Summary is not intended to identify key features or essential features of the claimed subject matter, nor is it intended to be used as an aid in determining the scope of the claimed subject matter.





BRIEF DESCRIPTION OF THE DRAWINGS

The detailed description is described with reference to the accompanying figures. In the figures, the left-most digit(s) of a reference number identifies the figure in which the reference number first appears. The use of the same reference numbers in different instances in the description and the figures may indicate similar or identical items. Entities represented in the figures may be indicative of one or more entities and thus reference may be made interchangeably to single or plural forms of the entities in the discussion.



FIG. 1 is an illustration of an example input device implementing the techniques described herein.



FIG. 2 is an illustration of an example computing device employing the techniques described herein.



FIG. 3 is an illustration of the computing device of FIG. 2 displaying a virtual keyboard.



FIGS. 4A and 4B illustrate an example input device with example user inputs.



FIG. 5 is an illustration of a system in an example implementation that is operable to employ the techniques described herein.



FIG. 6 depicts an example of a graph of pressure at a particular location over time.



FIG. 7 depicts another example of a graph of pressure at a particular location over time.



FIG. 8 is a flowchart illustrating an example process for implementing the techniques described herein in accordance with one or more embodiments.



FIG. 9 is a flowchart illustrating another example process for implementing the techniques described herein in accordance with one or more embodiments.



FIG. 10 illustrates an example system including various components of an example device that can be implemented as any type of computing device as described with reference to FIGS. 1-9 to implement embodiments of the techniques described herein.





DETAILED DESCRIPTION

Overview


Classifying the intent of user input techniques are described. Different types of user inputs can be provided by a user via a keyboard of an input device. These different types of user inputs include, for example, key strikes, multi-touch interactions, and/or mouse clicks. Force information regarding the pressure applied to a pressure sensitive keyboard, as well as one or more locations where that pressure is applied, is used to classify the intent of the user input as one of the various types of user inputs. Contact information regarding the contact area of the user input for other types of keyboards (e.g., capacitive and/or resistive systems), as well as one or more locations where that contact area occurs, may similarly be used to classify the intent of the user input as one of the various types of user inputs.


In the following discussion, an example environment is first described that may employ the techniques described herein. Example procedures are then described which may be performed in the example environment as well as other environments. Consequently, performance of the example procedures is not limited to the example environment and the example environment is not limited to performance of the example procedures.


Example Environment and Procedures



FIG. 1 is an illustration of an example input device 100 implementing the techniques described herein. In the illustrated example, the input device 100 is configured as a keyboard having a QWERTY arrangement of keys although other arrangements of keys are also contemplated. Further, other non-conventional configurations are also contemplated, such as a game controller, a remote control device, a configuration to mimic a musical instrument, and so forth. Thus, the input device 100 and keys incorporated by the input device 100 may assume a variety of different configurations to support a variety of different functionality.


The input device 100 is a multi-use device, supporting various types of user inputs. A user input can have various different intents, such as a multi-touch interaction, a key strike, a mouse click, and so forth. The input device 100 supports these different types of user inputs in a common input area, such as the keyboard. For example, the user may touch the keys “q”, “w”, and “e” on the keyboard, and the intent of those touches may be determined to be key strikes selecting the letters “q”, “w”, and “e”. The user may also swipe his or her finger across the keys “q”, “w”, and “e” on the keyboard, and the intent of that swipe may be determined to be a multi-touch interaction or single finger motion. This determination of the intent of the user input is discussed in more detail below.


The input device 100 may be communicatively coupled to a computing device 102. The input device 100 may be physically separate from the computing device 102 and communicate with the computing device 102 via any of a variety of conventional communication mechanisms. For example, the input device 100 may communicate with the computing device via a wireless connection, via a wired connection, via communication contacts of the devices 100 and 102 in contact with one another, and so forth.


The computing device 102 may range from full resource devices with substantial memory and processor resources to a low-resource device with limited memory and/or processing resources. The computing device 102 may also relate to software that causes the computing device 102 to perform one or more operations. In various implementations, the computing device 102 may assume a variety of different configurations, such as for computer 104, mobile 106, and television 108 uses. Each of these configurations includes devices that may have generally different constructs and capabilities, and thus the computing device 102 may be configured according to one or more of the different device classes.


For instance, the computing device 102 may be implemented as the computer 104 class of a device that includes a personal computer, desktop computer, a multi-screen computer, laptop computer, netbook, and so on. The computing device 102 may also be implemented as the mobile 106 class of device that includes mobile devices, such as a mobile phone, portable music player, portable gaming device, a tablet computer, a multi-screen computer, and so on. The computing device 102 may also be implemented as the television 108 class of device that includes devices having or connected to generally larger screens in casual viewing environments. These devices include televisions, set-top boxes, gaming consoles, and so on.


The computing device 102 also includes an input/output module 110. The input/output module 110 is representative of functionality relating to processing of inputs and rendering outputs of the computing device 102. A variety of different inputs may be processed by the input/output module 110, such as inputs relating to functions that correspond to keys of the input device 100, inputs that are multi-touch interactions or single finger motions recognized through the input device 100 and that cause operations to be performed that correspond to the multi-touch interactions or single finger motions, and so forth. Thus, the input/output module 110 may support a variety of different input techniques by recognizing and leveraging a division between types of inputs including key presses, multi-touch interactions, single finger motions, and so on.


Alternatively, the input device 100 may be included as part of a computing device. FIG. 2 is an illustration of an example computing device 202 employing the techniques described herein. The computing device 202 may assume a variety of different configurations, analogous to computing device 102 of FIG. 1. For example, the computing device 202 may be configured for mobile use, such as a mobile phone, a tablet computer, and so on. The computing device 202 may also include an input/output module 204, analogous to input/output module 110 of FIG. 1, and may also relate to software that causes the computing device 202 to perform one or more operations.


The computing device 202 includes a display device 206 via which, in a display mode, various data and information can be displayed. The display device 206 may use a variety of display technologies. These display technologies may include, for example, liquid crystal (LCD) display technologies, light-emitting diode (LED) display technologies, organic light-emitting diode (OLED) display technologies, plasma display technologies, and so forth. Although examples of display technologies are discussed herein, other display technologies are also contemplated.


The display device 206 can be a touchscreen display, and various user inputs can be provided by the user touching the display device 206. The computing device 202 also supports a keyboard mode in which a virtual keyboard is displayed. FIG. 3 is an illustration of computing device 202 displaying a virtual keyboard 302. Virtual keyboard 302 is a multi-use device, supporting various types of user inputs analogous to the keyboard of input device 100 of FIG. 1. However, rather than being a physically separate device, the keyboard 302 is a virtual keyboard that is part of the computing device 202. Thus, the display device 206 also serves as the input device in computing device 202.


User inputs to the input device (e.g., the input device 100 of FIG. 1 and/or the display device 206 of FIG. 2) are sensed by an input sensing component of the input device. The input device can use a variety of different input sensing technologies. These input sensing technologies may include pressure sensitive systems that sense pressure or force. These input sensing technologies may also include capacitive systems and/or resistive systems that sense touch. These input sensing technologies may also include optical based images that sense reflection or disruption of light from objects touching (or close to) the surface of the display device, such as Sensor in Pixel (SIP) systems, Infrared systems, optical imaging systems, and so forth. Other types of input sensing technologies can also be used, such as surface acoustic wave systems, acoustic pulse recognition systems, dispersive signal systems, and so forth. Although examples of input sensing technologies are discussed herein, other input sensing technologies are also contemplated.


In response to a user touching the keyboard at a particular time, the input sensing component determines a location that was touched at that particular time. Various modules of the input sensing component may use different algorithms or techniques to identify the location that was touched, and these algorithms or techniques can vary based on the particular input sensing technology being used. For pressure sensitive input sensing components, these various modules also identify the amount of pressure applied at the location that was touched. For other types of input sensing components, these various modules also identify the area that was touched (the contact area).


The determination of a location that is touched can be determined at various frequencies, such as 1000 times per second although other sampling frequencies are contemplated. It should be noted that the frequency at which the determination of a location of a touch is made is sufficient to identify whether the user input conforms to the characteristics of different user intents. For example, a sampling frequency of 1000 times per second may be sufficient to determine whether the user inputs conform to the characteristics of a key strike intent, whereas lower sampling frequencies (such as 100 times per second) may not be sufficient to identify whether the user inputs conform to the characteristics of a key strike intent. Thus, in contrast to the input sensing components discussed herein, many input sensing components that determine a location that is touched at a low frequency would be unable to determine whether the user input conforms to the characteristics of some intents.


The input sensing component detects or senses the touch of an object, such as a finger of a user's hand, a stylus, a pen, and so forth. The discussions of the classifying the intent of user input techniques herein refer to the user input being provided by a user's finger (and thus the touch is also referred to as a finger impact), although this user input can alternatively be provided by a stylus or other object controlled by the user.



FIGS. 4A and 4B illustrate an example input device 400 with example user inputs. Input device 400 can be, for example, an input device 100 of FIG. 1 and/or a display device 206 of FIG. 3. The input device 400 is a multi-use device, supporting various types of user inputs. These multiple types of user inputs are supported concurrently. For example, the input device 400 can include a keyboard and the user can provide different types of inputs on that keyboard, with the techniques discussed herein being used to automatically determine the intent of that user input.


Thus, the touching of the input device 400 can have different user intents, such as a key strike, a mouse click, a single finger motion, or a multi-touch interaction, although other intents or types of user input are also contemplated. A key strike refers to user selection of a particular key on the keypad, with different locations of the keyboard being associated with different keys. A mouse click refers to the pressing of a button typically associated with a mouse or button of a track pad. A multi-touch interaction refers to a user touching multiple locations (e.g., with multiple ones of the user's fingers) concurrently to initiate one or functions of a computing device. The multi-touch interaction can also include a motion or path taken by one or more of the user's fingers. For example, a multi-touch interaction may be sliding of the user's fingers in one or more particular directions, the user's fingers tracing one or more particular characters or symbol, and so forth. A single finger motion refers to motion or path taken by the user's finger to move a cursor, pointer, or other object (e.g., an icon, file, etc. being dragged), or to initiate one or more functions of the computing device. A single finger motion intent and a multi-touch interaction intent can also be referred to as movement intents because they typically (although not always for multi-touch interactions) include movement by the user's finger.



FIG. 4A illustrates an example user input the intent of which is a key strike of the letter “d” or a mouse click. Whether the intent of the user input was a key strike or a mouse click can be determined in different manners, such as based on the characteristics of a key strike and the characteristics of a mouse click as discussed below.



FIG. 4B illustrates an example user input the intent of which is a single finger motion moving from left to right. The ending position of the user's finger is illustrated using a dashed outline of a hand. Whether the intent of the user input was a single finger motion can be determined in different manners, such as based on the distance the touch travels and/or a velocity of the touch, as discussed in more detail below.


As illustrated in FIGS. 4A and 4B, both user inputs begin by touching a location of the keyboard corresponding to the letter “d”. However, the intent of the user (e.g., a key strike or a multi-touch interaction) can be determined using the techniques discussed herein. The user need not press any button or select any menu option to indicate his or her intent (e.g., whether the user desires to input key strikes or multi-touch interactions). Rather, the user can simply provide his or her input and the techniques discussed herein automatically determine the intent of that input.


The touch, as well as one or more locations of the touch as sensed by one or more sensors as discussed above, is used to classify the intent of a user input, such as classifying the user input as a key strike or other type of input. The user input refers to the touching by the user of one or more portions of the input device. The touch may be sensed using a dense array of pressure sensors that are sampled at a sufficient frequency to characterize the touch, as discussed above. The touch may alternatively be sensed using a capacitive sensor, with the impact being sensed indirectly by noting the rapid change in coupling as the finger conforms to the surface and then rebounds. Regardless of the type of sensors use, the location of the touch can be readily determined based on (e.g., by interpolating) data from the sensors in the region where the touch occurred.


The touch can have associated force information (in situations in which the input sensing technologies used include pressure sensitive systems) or contact information (in situations in which other input sensing technologies are used). The force information refers to the pressure applied by the user when touching the keyboard. The contact information refers to the area that is touched (the portions of the touchpad or touchscreen that were touched by the user's finger or other object, the amount of light reflected by the user's finger or other object, etc.) by the user when touching the keyboard.


In the following discussions, the classifying the intent of user input techniques are discussed with reference to force information and a user input device that is a pressure sensitive device. However, the classifying the intent of user input techniques can alternatively be implemented using other types of sensors as discussed above. In such situations, rather than using force information to classify the intent of the user input, the contact information is used to classify the intent of the user input. When using the contact information, the contact area is treated analogously to the pressure discussed below.



FIG. 5 is an illustration of a system 500 in an example implementation that is operable to employ the techniques described herein. The system 500 includes a touch information collection module 502 and an input classification module 504. System 500 may be implemented, for example, in the input device 100 and/or the computing device 102 of FIG. 1, or the computing device 202 of FIGS. 2 and 3. Thus, for example, module 502 may be implemented in input device 100 and module 504 may be implemented in computing device 102, both modules 502 and 504 may be implemented in input device 100, and so forth.


Touch information collection module 502 obtains an indication of the amount of pressure applied by a user's finger over time (or alternatively contact information over time, as discussed above), as well as the locations of the touch over time, as sensed by the input sensing component as discussed above. Module 502 obtains (e.g., receives or generates) touch information 506 regarding user inputs to the keyboard of the input device. The touch information 506 identifies the characteristics of and locations of touches. These characteristics and locations of a touch can include, for example, the size of the touch (e.g., the amount of area touched), changes in the size of the touch over time, the shape of the touch (e.g., a geographic shape or outline of the area touched), changes in the shape of the touch over time, the location of the touch over time, the change in pressure of the touch over time, the movement of the touch (directions and locations that are touched), a velocity of the touch, an acceleration of the touch, a distance the touch travels, and so forth.


Based on the touch information 506, input classification module 504 classifies the intent of the user input, and outputs or otherwise makes available the input intent classification 508. Additional information, such as the touch information 506, can also be output or otherwise made available along with the input intent classification 508. The intent of the user input can be, for example, a key strike, a multi-touch interaction, a mouse click, and so forth as discussed below. The input intent classification 508 can subsequently be used by one or more other modules to take the appropriate action based on the user input. For example, if the intent is a key strike, then the key associated with the location pressed by the user as the user input is recorded as user selection of that key. By way of another example, if the intent is a multi-touch interaction, then the touch information is used to determine which multi-touch interaction was input by the user and the corresponding action taken.


Generally, the intent of the user input is classified based on the touch information 506. The force applied by the user's finger over time, and optionally the one or more locations where that force is applied, is analyzed in view of characteristics of the different intents supported by input classification module 504. A user input that conforms to the characteristics of a particular intent is classified as that particular intent, whereas a user input that does not conform to the characteristics of a particular intent is not classified as that particular intent.


In one or more embodiments, the intent of a user input can be classified as a key strike, a single finger motion, a multi-touch interaction, or other input. A user input is compared to the characteristics of one or more of a key strike, a single finger motion, a multi-touch interaction. If the user input conforms to the key strike characteristics then the intent of the user input is classified as a key strike, if the user input conforms to the single finger motion characteristics then the intent of the user input is classified as a single finger motion, and if the user input conforms to the multi-touch interaction characteristics then the intent of the user input is classified as a multi-touch interaction.


Furthermore, if the user input conforms to the characteristics of no particular intent (e.g., a key strike, a single finger motion, a multi-touch interaction, etc.), then the intent of the user input is classified as an “other” intent. The other intent refers to some other type of intent, typically an intent that is filtered out by system 500. Examples of such user inputs that can be classified as an “other” intent can be the user's hands resting on the keyboard, sensor noise, the device being bumped, and so forth. Thus, not only can input classification module 504 distinguish between different intents of user inputs to the keyboard, but module 504 can also filter out user inputs to that same keyboard that are not intended to be user inputs to the input device that are acted upon by the computing device.


A user input intent may be characterized by a force graph shape indicating the force applied by the user's finger over time for the intent. Different user input intents can have different associated force graph shapes, and input classification module 504 can maintain these different force graph shapes and compare user inputs to these different force graph shapes. A user input (as characterized by touch information 506) that conforms to the force graph shape for an intent is classified as being that particular intent. Whether a user input conforms to a particular force graph shape can be determined in a variety of different manners. For example, a force graph can be generated based on the touch information obtained for the user input, and the generated force graph can be compared to the one or more maintained force graph shapes. If the generated force graph shape matches a particular maintained force graph shape, then the user input conforms to the intent associated with that particular maintained force graph shape. Whether two graph shapes match can be determined using a variety of public and/or proprietary graph or shape matching techniques.


The behavior of a particular touching of the keyboard over a lifetime of the touch may be used to classify a user input as a key strike or other type of input (e.g., a multi-touch interaction or single finger motion). The lifetime of the touch refers to a time duration that begins when the user's finger touching the surface is sensed and ends when the user's finger is no longer sensed as touching the surface. The user's finger can remain approximately stationary for the lifetime of the touch (e.g., which may be typical for a key strike), or can move across the keypad during the lifetime of the touch (e.g., which may be typical for multi-touch interactions and single finger motions).


The intent of a user input may be classified as a movement in response to the touch travelling at least a threshold distance. This threshold distance can be a fixed distance (e.g., 1.5 inches) or a relative distance (e.g., 10% of the width of the keyboard). The travelling of a touch refers to the distance moved by the user's finger while being moved along some path during the lifetime of the touch. If multiple touches are sensed concurrently then the intent of the user input may be classified as a multi-touch interaction, whereas if only a single touch is sensed at a particular time then the intent of the user input at that particular time may be classified as a single finger motion.


The intent of a user input may also be classified as a movement in response to the touch having at least a threshold velocity and a short (e.g., less than a threshold value) lifetime. The threshold value of the lifetime of the touch may be 0.25 seconds, although other threshold values are contemplated. The velocity of a touch refers to the distance moved by the user's finger while being moved along some path during the lifetime of the touch divided by the time duration of the lifetime of the touch. For example, the velocity may be 4 inches/second, although other velocities are contemplated. If multiple touches are sensed concurrently then the intent of the user input may be classified as a multi-touch interaction, whereas if only a single touch is sensed at a particular time then the intent of the user input at that particular time may be classified as a single finger motion.


Generally, a key strike (also referred to as a tap) is characterized by a sharp rise in force followed by a sharp decrease in force as the finger rebounds from the surface, and optionally followed by a rise in force again as the momentum of the remainder of the finger continues towards the surface of the keyboard. The pressure of a touch is analyzed over time, and based on these characteristics of a key strike the user input is classified as either a key strike or some other type of input (e.g., a multi-touch interaction). User inputs that conform to these characteristics are classified as key strikes, and user inputs that do not conform to these characteristics are classified as other types of user inputs. If the user inputs include multiple touches concurrently, then the user input may be classified as multiple key strikes if each touch conforms to the characteristics of a key strike and the touches are for particular locations (e.g., locations corresponding to a modifier key on the keyboard, such as a “shift” key, a “ctrl” key, an “alt” key, and so forth). Alternatively, if the user inputs include multiple touches concurrently, then the user input is classified as some other type of input rather than a key strike.


It should be noted that these characteristics of a key strike allow the intent of a user input to be distinguished between a key strike and an “other” intent (e.g., the user resting his or her hands on the keyboard). The key strike is characterized by a particular force graph shape indicating the force applied by the user's finger over time. If a user intends to strike a key then the input conforms to this particular force graph shape, and if the user intends to rest his or her hands on the keyboard then the input does not conform to this particular force graph shape. Thus a key strike intent or an “other” intent can be readily determined based on whether the user input conforms to this particular force graph shape.


The characteristics of a key strike can also include temporal information. For example, the characteristics of a key strike may include the touch being at the same location (not moving) or having a location that moves less than a threshold amount (e.g., less than the width or height of a key on the keyboard, although other threshold amounts are contemplated).



FIG. 6 depicts an example 600 of a graph of pressure of a touch over time. The vertical axis is pressure or force (e.g., in grams), and the horizontal axis is time (e.g., in milliseconds). A line 602 represents the pressure applied as the touch over time. The pressure applied as the touch over time can be analyzed to determine whether the pressure applied conforms to the characteristics of a key strike.


In analyzing the pressure an initial impact time duration may be determined, which is a duration of time that begins when the pressure applied as the touch rises to (e.g., is equal to or greater than) a start threshold amount. This start threshold amount may be 200 grams, although other start threshold amounts are contemplated. This duration of time ends when the pressure applied as the touch reaches a local minimum before rising again (e.g., as the momentum of the remainder of the finger continues towards the surface of the keyboard) or the pressure applied as the touch drops to (e.g., is equal to or less than) a stop threshold amount. This stop threshold amount may be the same as the start threshold amount (e.g., 200 grams) or a lesser amount (e.g., 100 grams), although other values for the stop threshold amount are contemplated. Thus, the initial impact time duration can be different for different user inputs.


In example 600, an initial impact time duration 604 is illustrated. The initial impact time duration begins at point 606 where the pressure applied as the touch rises to the start threshold amount, and ends at point 608 where the pressure applied as the touch drops to a local minimum before rising again.


Also in analyzing the pressure a maximum force may be identified. This maximum force refers to a local maximum of the pressure applied as the touch during the initial impact time duration 604 and after the pressure applied as the touch rises to the start threshold amount. In example 600, the maximum force is at point 610.


Given the initial impact time duration and/or the maximum force, various rules or criteria are applied to determine whether the pressure applied conforms to the characteristics of a key strike. These rules or criteria may include an initial impact criteria, a maximum force criteria, and/or an initial impact ending force criteria.


The initial impact criteria refers to the duration of the initial impact time duration, indicating that the initial impact time duration is to be within a particular time range. This particular time range can be, for example, between 5 milliseconds (ms) and 25 ms, although other time ranges are contemplated. Thus, for example, the initial impact time duration is to be at least 5 ms and no more than 25 ms in order for the pressure applied as the touch to conform to the characteristics of a key strike.


The maximum force criteria refers to when, within the initial impact time duration, the maximum force occurs. The maximum force criteria indicates that the maximum force is to occur within a threshold amount of time, such as 12 ms although other amounts of time are also contemplated. The maximum force criteria also indicates that the maximum force is not to occur in a last portion of the initial impact time duration. This last portion may be the last third of the initial impact time duration, although other portions are contemplated. Thus, for example, the maximum force is to occur within the first 12 ms of the initial impact time duration and is not to occur within the last ⅓ of the initial impact time duration in order for the pressure applied as the touch to conform to the characteristics of a key strike.


The initial impact ending force criteria refers to the pressure of the touch at the end of the initial impact time duration. The initial impact ending force criteria indicates that the pressure of the touch at the end of the initial impact time duration is to be less than a particular percentage of the maximum force. This particular percentage may be 80%, although other percentages are contemplated. Alternatively, the initial impact ending force criteria can indicate that the pressure of the touch is to be less than another particular percentage of the maximum force in response to the maximum force being reached quickly (within a threshold amount of time). This other particular percentage may be 90% and the threshold amount of time may be 5 ms, although other percentages and threshold amounts of time are contemplated. Thus, for example, at the end of the initial impact time duration the pressure of the touch is to be less than 80% of the maximum force, or less than 90% of the maximum force if the maximum force was reached within 5 ms, in order for the pressure applied as the touch to conform to the characteristics of a key strike.


If the initial impact criteria, the maximum force criteria, and the initial impact ending force criteria all conform to the characteristics of a key strike, then the pressure applied as the touch conforms to the characteristics of a key strike and the user input is classified as a key strike. However, if one or more of the initial impact criteria, the maximum force criteria, and the initial impact ending force criteria does not conform to the characteristics of a key strike, then the pressure applied as the touch does not conform to the characteristics of a key strike and the user input is not classified as a key strike.


Characteristics of a key strike are discussed above. Characteristics of other types of inputs can also be maintained. The pressure of a touch may be analyzed over time and a determination made as to whether the user input conforms to the characteristics of another type of input. User inputs that conform to the characteristics of another type of input are classified as that other type of input, while user inputs that do not conform to the characteristics of that other type input are not classified as that other type of input.


For example, a mouse click may be characterized by a slow rise in force followed by a sharp decrease in force as the finger rebounds from the surface. The pressure of a touch is analyzed over time, and based on these characteristics of a mouse click the user input is classified as either a mouse click or some other type of input (e.g., a key strike). User inputs that conform to these characteristics are classified as mouse clicks, and user inputs that do not conform to these characteristics are classified as other types of user inputs.


The characteristics of a mouse click can also include temporal information. For example, the characteristics of a mouse click may include the touch being at the same location (not moving) or having a location that moves less than a threshold amount (e.g., 0.25 inches, although other threshold amounts are contemplated).



FIG. 7 depicts an example 700 of a graph of pressure of a touch over time. The vertical axis is pressure or force (e.g., in grams), and the horizontal axis is time (e.g., in milliseconds). A line 702 represents the pressure applied as the touch over time. The pressure applied as the touch over time can be analyzed to determine whether the pressure applied conforms to the characteristics of a mouse click.


In analyzing the pressure a touch time duration may be determined, which is a duration of time that begins when the pressure applied as the touch rises to (e.g., is equal to or greater than) a start threshold amount. This start threshold amount may be 200 grams, although other start threshold amounts are contemplated. This duration of time ends when the pressure applied as the touch drops to (e.g., is equal to or less than) a stop threshold amount. This stop threshold amount may be the same as the start threshold amount (e.g., 200 grams) or a lesser amount (e.g., 100 grams), although other values for the stop threshold amount are contemplated.


In example 700, a touch time duration 704 is illustrated. The touch time duration begins at point 706 where the pressure applied as the touch rises to the start threshold amount, and ends at point 708 where the pressure applied as the touch drops to the stop threshold amount.


Also in analyzing the pressure a maximum force may be identified. This maximum force refers to a local maximum of the pressure applied as the touch during the touch time duration 704 and after the pressure applied as the touch rises to the start threshold amount. In example 700, the maximum force is at point 710.


Given the touch time duration and/or the maximum force, various rules or criteria are applied to determine whether the pressure applied conforms to the characteristics of a mouse click. These rules or criteria may include rules or criteria referring to when the maximum force occurs. The rules or criteria may indicate that the maximum force is to occur at least a threshold amount of time (e.g., 25 ms, although other amounts of time are also contemplated) after the pressure applied as the touch rises to the start threshold amount. The rules or criteria may indicate that the maximum force is not to occur in a beginning portion of the touch time duration. This beginning portion may be the beginning half of the touch time duration, although other portions are contemplated. Thus, for example, the rules or criteria may indicate that the maximum force is to occur at least 25 ms after the touch time duration beginning, and is not to occur within the first half of the touch time duration in order for the pressure applied as the touch to correspond to the characteristics of a mouse click.


As indicated above, although discussed herein with reference to force information and a user input device that is a pressure sensitive input device, the classifying the intent of user input techniques can alternatively be implemented using other types of sensors and input sensing technologies. In such situations, rather than using the pressure applied by a user input to classify the intent of the user input, the contact area of the user input (the portions of the touchpad or touchscreen that were touched by the user's finger or other object, the amount of light reflected by the user's finger or other object, etc.) may be used to classify the intent of a user input. The contact area (e.g., measured in millimeters or other units) is treated analogously to the pressure discussed above. For example, referring to FIG. 5, a key strike is characterized by a sharp rise in contact area followed by a sharp decrease in contact area as the finger or other object touching the surface rebounds from the surface, and optionally followed by a rise in contact area again as the momentum of the remainder of the finger or other object touching the surface of the keyboard continues towards the surface.



FIG. 8 is a flowchart illustrating an example process 800 for implementing the techniques described herein in accordance with one or more embodiments. Process 800 is carried out by an input classification module, such as input classification module 504 of FIG. 5, and can be implemented in software, firmware, hardware, or combinations thereof. Process 800 is shown as a set of acts and is not limited to the order shown for performing the operations of the various acts. Process 800 is an example process for implementing the techniques described herein; additional discussions of implementing the techniques described herein are included herein with reference to different figures.


In process 800, touch information regarding a user input to an input device is obtained (act 802). This touch information can be obtained from a pressure sensitive input device, or other types of input devices (e.g., capacitive or resistive) as discussed above.


Based on the obtained touch information, an intent of the user input is classified as being a key strike or other type of input (act 804). The classification is performed based on characteristics of a key strike or other type of input, as discussed above. Various other types of inputs can be received as user inputs, such as multi-touch interactions, a single finger motions, and mouse clicks as discussed above.


An indication of the classification of the user input is output (act 806). This classification can be used by one or more other modules (e.g., of computing device 102 of FIG. 1) to take the appropriate action based on the user input. This classification can be output to, for example, one or more modules of the input device 104 and/or the computing device 102 of FIG. 1.



FIG. 9 is a flowchart illustrating another example process 900 for implementing the techniques described herein in accordance with one or more embodiments. Process 900 is carried out by an input classification module, such as input classification module 504 of FIG. 5, and can be implemented in software, firmware, hardware, or combinations thereof. Process 900 is shown as a set of acts and is not limited to the order shown for performing the operations of the various acts. Process 900 is an example process for implementing the techniques described herein; additional discussions of implementing the techniques described herein are included herein with reference to different figures.


In process 900, when a user input is received (e.g., the user touches the keyboard), a check is made as to whether the touch travels a threshold distance (act 902). This threshold distance can be a fixed or relative distance, such as 1.5 inches or 10% of the width of the keyboard, as discussed above.


If the touch travels a threshold distance (e.g., at least the threshold distance, equal to or greater than the threshold distance, etc.), then the touch is classified as a movement (act 904). A movement can be, for example, a single finger motion or a multi-touch interaction, as discussed above. Whether the movement is a single finger motion or a multi-touch interaction can be determined in different manners. For example, if during the time of the touch multiple touches (e.g., multiple fingers) are sensed concurrently then the movement is a multi-touch interaction, whereas if during the time of the touch just a single touch (e.g., a single finger) is sensed then the movement is a single finger motion. By way of another example, if the input device does not support (or is operating in a mode that does not support) multi-touch interactions, then the movement is a single finger motion. By way of another example, if the input device does not support (or is operating in a mode that does not support) single finger motions, then the movement is a multi-touch interaction.


However, if the touch does not travel the threshold distance, then a check is made as to whether the lifetime of the touch is less than a threshold amount (act 906). If the lifetime of the touch is less than a threshold amount (e.g., less than 0.25 seconds), then a check is made as to whether the touch has a threshold velocity (act 908). This threshold velocity can be, for example, 4 inches/second or other velocities as discussed above. If the touch has a threshold velocity (e.g., at least the threshold velocity, equal to or greater than the threshold velocity, etc.), then the touch is classified as a movement (act 904). A movement can be, for example, a single finger motion or a multi-touch interaction, as discussed above.


However, if the lifetime of the touch is not less than the threshold amount, or if the touch does not have the threshold velocity, then a check is made as to whether initial impact criteria are satisfied (act 910). The initial impact criteria are satisfied if the initial impact time duration is within a particular time range (e.g., at least 5 ms and no more than 25 ms), as discussed above.


If the initial impact criteria are not satisfied, then the touch is classified as neither a movement nor a key strike (act 912). In such situations, the touch may be interpreted as some default input (e.g., the user resting his or her fingers on the keyboard), or alternatively may be further analyzed to determine the intent of the user input (e.g., a mouse click, as discussed above).


However, if the initial impact criteria are satisfied, then a check is made as to whether maximum force criteria are satisfied (act 914). The maximum force criteria are satisfied if the maximum force occurs during a particular portion of the initial impact time duration (e.g., within the first 12 ms of the initial impact time duration, or during the first ⅔ of the initial impact time duration), as discussed above.


If the maximum force criteria are not satisfied, then the touch is classified as neither a movement nor a key strike (act 912).


However, if the maximum force criteria are satisfied, then a check is made as to whether initial impact ending force criteria are satisfied (act 916). The initial impact ending force criteria are satisfied if the pressure on the key at the end of the initial impact time duration is less than a particular percentage of the maximum force (e.g., less than 80% of the maximum force, or less than 90% of the maximum force if the maximum force occurred in the first 5 ms of the initial impact time duration), as discussed above.


If the initial impact ending force criteria are not satisfied, then the touch is classified as neither a movement nor a key strike (act 912). However, if the initial impact ending force criteria are satisfied, then the touch is classified as a key strike (act 918).


Example System and Device



FIG. 10 illustrates an example system generally at 1000 that includes an example computing device 1002 that is representative of one or more computing systems and/or devices that may implement the various techniques described herein. The computing device 1002 may, for example, be configured to assume a mobile configuration through use of a housing formed and size to be grasped and carried by one or more hands of a user, illustrated examples of which include a mobile phone, mobile game and music device, and tablet computer although other examples are also contemplated.


The example computing device 1002 as illustrated includes a processing system 1004, one or more computer-readable media 1006, and one or more I/O interfaces 1008 that are communicatively coupled, one to another. Although not shown, the computing device 1002 may further include a system bus or other data and command transfer system that couples the various components, one to another. A system bus can include any one or combination of different bus structures, such as a memory bus or memory controller, a peripheral bus, a universal serial bus, and/or a processor or local bus that utilizes any of a variety of bus architectures. A variety of other examples are also contemplated, such as control and data lines.


The processing system 1004 is representative of functionality to perform one or more operations using hardware. Accordingly, the processing system 1004 is illustrated as including hardware elements 1010 that may be configured as processors, functional blocks, and so forth. This may include implementation in hardware as an application specific integrated circuit or other logic device formed using one or more semiconductors. The hardware elements 1010 are not limited by the materials from which they are formed or the processing mechanisms employed therein. For example, processors may be comprised of semiconductor(s) and/or transistors (e.g., electronic integrated circuits (ICs)). In such a context, processor-executable instructions may be electronically-executable instructions.


The computer-readable storage media 1006 is illustrated as including memory/storage 1012. The memory/storage 1012 represents memory/storage capacity associated with one or more computer-readable media. The memory/storage component 1012 may include volatile media (such as random access memory (RAM)) and/or nonvolatile media (such as read only memory (ROM), Flash memory, optical disks, magnetic disks, and so forth). The memory/storage component 1012 may include fixed media (e.g., RAM, ROM, a fixed hard drive, and so on) as well as removable media (e.g., Flash memory, a removable hard drive, an optical disc, and so forth). The computer-readable media 1006 may be configured in a variety of other ways as further described below.


Input/output interface(s) 1008 are representative of functionality to allow a user to enter commands and information to computing device 1002, and also allow information to be presented to the user and/or other components or devices using various input/output devices. Examples of input devices include a keyboard, a cursor control device (e.g., a mouse), a microphone, a scanner, touch functionality (e.g., capacitive or other sensors that are configured to detect physical touch), a camera (e.g., which may employ visible or non-visible wavelengths such as infrared frequencies to recognize movement as gestures that do not involve touch), and so forth. Examples of output devices include a display device (e.g., a monitor or projector), speakers, a printer, a network card, tactile-response device, and so forth. Thus, the computing device 1002 may be configured in a variety of ways to support user interaction.


The computing device 1002 is further illustrated as being communicatively and physically coupled to an input device 1014 that is physically and communicatively removable from the computing device 1002. In this way, a variety of different input devices may be coupled to the computing device 1002 having a wide variety of configurations to support a wide variety of functionality. In this example, the input device 1014 includes one or more keys 1016, which may be configured as pressure sensitive keys, keys on a touchpad or touchscreen, mechanically switched keys, and so forth.


The input device 1014 is further illustrated as including one or more modules 1018 that may be configured to support a variety of functionality. The one or more modules 1018, for instance, may be configured to process analog and/or digital signals received from the keys 1016 to determine whether a keystroke was intended, determine whether an input is indicative of resting pressure, support authentication of the input device 1014 for operation with the computing device 1002, classify the intent of user input, and so on. Modules 1018 may include, for example, the input classification module 504 of FIG. 5.


Although illustrated as separate from the computing device 1002, the input device 1014 can alternatively be included as part of the computing device 1002 as discussed above. In such situations, the keys 1016 and the modules 1018 are included as part of the computing device 1002. Additionally, in such situations the keys 1016 may be keys of a virtual keyboard and/or keys of a non-virtual keyboard (e.g., a pressure sensitive input device).


Various techniques may be described herein in the general context of software, hardware elements, or program modules. Generally, such modules include routines, programs, objects, elements, components, data structures, and so forth that perform particular tasks or implement particular abstract data types. The terms “module,” “functionality,” and “component” as used herein generally represent software, firmware, hardware, or a combination thereof. The features of the techniques described herein are platform-independent, meaning that the techniques may be implemented on a variety of commercial computing platforms having a variety of processors.


An implementation of the described modules and techniques may be stored on or transmitted across some form of computer-readable media. The computer-readable media may include a variety of media that may be accessed by the computing device 1002. By way of example, and not limitation, computer-readable media may include “computer-readable storage media” and “computer-readable signal media.”


“Computer-readable storage media” may refer to media and/or devices that enable persistent and/or non-transitory storage of information in contrast to mere signal transmission, carrier waves, or signals per se. Thus, computer-readable storage media refers to non-signal bearing media. The computer-readable storage media includes hardware such as volatile and nonvolatile, removable and non-removable media and/or storage devices implemented in a method or technology suitable for storage of information such as computer readable instructions, data structures, program modules, logic elements/circuits, or other data. Examples of computer-readable storage media may include, but are not limited to, RAM, ROM, EEPROM, flash memory or other memory technology, CD-ROM, digital versatile disks (DVD) or other optical storage, hard disks, magnetic cassettes, magnetic tape, magnetic disk storage or other magnetic storage devices, or other storage device, tangible media, or article of manufacture suitable to store the desired information and which may be accessed by a computer.


“Computer-readable signal media” may refer to a signal-bearing medium that is configured to transmit instructions to the hardware of the computing device 1002, such as via a network. Signal media typically may embody computer readable instructions, data structures, program modules, or other data in a modulated data signal, such as carrier waves, data signals, or other transport mechanism. Signal media also include any information delivery media. The term “modulated data signal” means a signal that has one or more of its characteristics set or changed in such a manner as to encode information in the signal. By way of example, and not limitation, communication media include wired media such as a wired network or direct-wired connection, and wireless media such as acoustic, RF, infrared, and other wireless media.


As previously described, hardware elements 1010 and computer-readable media 1006 are representative of modules, programmable device logic and/or fixed device logic implemented in a hardware form that may be employed in some embodiments to implement at least some aspects of the techniques described herein, such as to perform one or more instructions. Hardware may include components of an integrated circuit or on-chip system, an application-specific integrated circuit (ASIC), a field-programmable gate array (FPGA), a complex programmable logic device (CPLD), and other implementations in silicon or other hardware. In this context, hardware may operate as a processing device that performs program tasks defined by instructions and/or logic embodied by the hardware as well as a hardware utilized to store instructions for execution, e.g., the computer-readable storage media described previously.


Combinations of the foregoing may also be employed to implement various techniques described herein. Accordingly, software, hardware, or executable modules may be implemented as one or more instructions and/or logic embodied on some form of computer-readable storage media and/or by one or more hardware elements 1010. The computing device 1002 may be configured to implement particular instructions and/or functions corresponding to the software and/or hardware modules. Accordingly, implementation of a module that is executable by the computing device 1002 as software may be achieved at least partially in hardware, e.g., through use of computer-readable storage media and/or hardware elements 1010 of the processing system 1004. The instructions and/or functions may be executable/operable by one or more articles of manufacture (for example, one or more computing devices 1002 and/or processing systems 1004) to implement techniques, modules, and examples described herein.


CONCLUSION

Although the example implementations have been described in language specific to structural features and/or methodological acts, it is to be understood that the implementations defined in the appended claims is not necessarily limited to the specific features or acts described. Rather, the specific features and acts are disclosed as example forms of implementing the claimed features.

Claims
  • 1. A method comprising: obtaining touch information regarding a single finger touching a keyboard of an input device, the input device supporting multiple different user intents for the single finger touching the keyboard, the obtaining comprising obtaining force information; andautomatically determining, based on the touch information, which of the multiple different user intents is intended for the single finger touching the keyboard, the multiple different user intents including a key strike, a mouse click, and a single finger motion, the determining including: determining that the user intent intended for the single finger touching the keyboard is the key strike in response to the touch information conforming to characteristics of the key strike,determining that the user intent intended for the single finger touching the keyboard is the mouse click in response to the touch information conforming to characteristics of the mouse click, the characteristics of the key strike indicating that the force applied by the single finger touching the keyboard rises faster than the characteristics of the mouse click indicate,determining that the user intent intended for the single finger touching the keyboard is the single finger motion in response to the touch information conforming to characteristics of the single finger motion.
  • 2. The method as recited in claim 1, the input device comprising multiple pressure sensors that sense the single finger touching the keyboard, and the obtaining comprising obtaining force information.
  • 3. The method as recited in claim 1, the input device comprising a capacitive system that senses the single finger touching the keyboard, and the obtaining comprising obtaining contact information.
  • 4. The method as recited in claim 1, the determining that the user intent is the single finger motion comprising determining that the user intent is the single finger motion in response to the single finger travelling a threshold distance across the keyboard.
  • 5. The method as recited in claim 1, the determining that the user intent is the single finger motion comprising determining that the user intent is the single finger motion in response to the single finger having a threshold velocity across the keyboard.
  • 6. The method as recited in claim 1, the input device comprising an input device that is physically separate from and communicatively removable from a computing device.
  • 7. The method as recited in claim 1, the determining comprising determining which of the multiple different user intents is intended by comparing the touch information to one or more graphs representing user input intents, and determining whether the key strike or the mouse click is intended based on an intent associated with the one of the one or more graphs with which the touch information conforms.
  • 8. The method as recited in claim 1, the determining comprising determining which of the multiple different user intents is intended for the single finger touching the keyboard in the absence of another option or button selection by a user to indicate which of the multiple different user intents is intended.
  • 9. The method as recited in claim 1, further comprising: using the touch information to identify a location of the keyboard touched by the single finger touching the keyboard; andrecording a user selection of a key associated with the location in response to determining that the user intent intended for the single finger touching the keyboard is the key strike, but not recording the user selection of the key associated with the location in response to determining that the user intent intended for the single finger touching the keyboard is the mouse click or the single finger motion.
  • 10. An apparatus comprising an input device configured to provide an output that indicates touch information regarding a single finger touching a keyboard of the input device, the touch information including force information, the input device supporting multiple different user intents for the single finger touching the keyboard of the input device, the multiple different user intents including a key strike, a mouse click, and a single finger motion, the output being usable by one or more modules to automatically classify the user intent for the single finger touching the keyboard as the key strike in response to the touch information conforming to characteristics of the key strike, classify the user intent for the single finger touching the keyboard as the mouse click in response to the touch information conforming to characteristics of the mouse click, and classify the user intent for the single finger touching the keyboard as the single finger motion in response to the touch information conforming to characteristics of the single finger motion, the characteristics of the key strike indicating that the force applied by the single finger touching the keyboard rises faster than the characteristics of the mouse click indicate.
  • 11. The apparatus as recited in claim 10, the input device comprising a pressure sensitive input device.
  • 12. The apparatus as recited in claim 11, the touch information comprising information indicating changes in locations where pressure is applied to the keyboard of the pressure sensitive input device over time as well as changes in pressure applied to the keyboard over time.
  • 13. The apparatus as recited in claim 10, the apparatus including the one or more modules.
  • 14. The apparatus as recited in claim 10, the output being usable by the one or more modules to automatically classify the user intent for the single finger touching keyboard in the absence of another option or button selection by a user to indicate which of the multiple different user intents is intended.
  • 15. The apparatus as recited in claim 10, the input device comprising an input device that is physically separate from and communicatively removable from a computing device.
  • 16. The apparatus as recited in claim 10, further comprising: the output being usable by the one or more modules to record a user selection of a key associated with a location of the keyboard touched by the single finger in response to classifying the user intent for the single finger touching the keyboard as the key strike, but not record the user selection of the key associated with the location in response to classifying the user intent for the single finger touching the keyboard as the mouse click or the single finger motion.
  • 17. An apparatus comprising: a hardware processor; andcomputer-readable storage devices having stored thereon multiple instructions that, responsive to execution by the processor, cause the processor to perform operations comprising: obtaining touch information regarding a single finger touching a keyboard of an input device, the touch information comprising force information, the input device supporting multiple different user intents for the single finger touching the keyboard; andautomatically determining, based on the touch information, which of the multiple different user intents is intended for the single finger touching the keyboard, the multiple different user intents including a key strike, a mouse click, and a single finger motion, and the determining including: determining that the user intent intended for the single finger touching the keyboard is the key strike in response to the touch information conforming to characteristics of a key strike,determining that the user intent intended for the single finger touching the keyboard is the mouse click in response to the touch information conforming to characteristics of the mouse click, the characteristics of the key strike indicating that the force applied by the single finger touching the keyboard rises faster than the characteristics of the mouse click indicate, anddetermining that the user intent intended for the single finger touching the keyboard is the single finger motion in response to the touch information conforming to characteristics of the single finger motion.
  • 18. The apparatus as recited in claim 17, the apparatus including the input device.
  • 19. The apparatus as recited in claim 17, the single finger touching the keyboard comprising a single touch user input, and the touch information comprising force information indicating pressure applied to the keyboard by the single touch user input.
  • 20. The apparatus as recited in claim 17, the single finger touching the keyboard comprising a single touch user input, and the touch information comprising contact information indicating an area of the keyboard that is touched by the single touch user input.
RELATED APPLICATIONS

This application claims priority under 35 U.S.C. § 119(e) to the following U.S. Provisional Patent Applications, the entire disclosures of each of these applications being incorporated by reference in their entirety: U.S. Provisional Patent Application No. 61/606,321, filed Mar. 2, 2012, and titled “Screen Edge;” U.S. Provisional Patent Application No. 61/606,301, filed Mar. 2, 2012, and titled “Input Device Functionality;” U.S. Provisional Patent Application No. 61/606,313, filed Mar. 2, 2012, and titled “Functional Hinge;” U.S. Provisional Patent Application No. 61/606,333, filed Mar. 2, 2012, and titled “Usage and Authentication;” U.S. Provisional Patent Application No. 61/613,745, filed Mar. 21, 2012, and titled “Usage and Authentication;” U.S. Provisional Patent Application No. 61/606,336, filed Mar. 2, 2012, and titled “Kickstand and Camera;” and U.S. Provisional Patent Application No. 61/607,451, filed Mar. 6, 2012, and titled “Spanaway Provisional.”

US Referenced Citations (910)
Number Name Date Kind
578325 Fleming Mar 1897 A
3600528 Leposavic Aug 1971 A
3777082 Hatley Dec 1973 A
3879586 DuRocher et al. Apr 1975 A
3968336 Johnson Jul 1976 A
4046975 Seeger, Jr. Sep 1977 A
4065649 Carter et al. Dec 1977 A
4086451 Boulanger Apr 1978 A
4243861 Strandwitz Jan 1981 A
4261042 Ishiwatari et al. Apr 1981 A
4302648 Sado et al. Nov 1981 A
4317011 Mazurk Feb 1982 A
4317013 Larson Feb 1982 A
4323740 Balash Apr 1982 A
4365130 Christensen Dec 1982 A
4375018 Petersen Feb 1983 A
4492829 Rodrique Jan 1985 A
4503294 Matsumaru Mar 1985 A
4527021 Morikawa et al. Jul 1985 A
4559426 Van Zeeland et al. Dec 1985 A
4577822 Wilkerson Mar 1986 A
4588187 Dell May 1986 A
4607147 Ono et al. Aug 1986 A
4651133 Ganesan et al. Mar 1987 A
4652704 Franklin Mar 1987 A
4724605 Fiorella Feb 1988 A
4735394 Facco Apr 1988 A
4795977 Frost et al. Jan 1989 A
4801771 Mizuguchi et al. Jan 1989 A
4824268 Diernisse Apr 1989 A
4864084 Cardinale Sep 1989 A
4983787 Kunikane Jan 1991 A
4990900 Kikuchi Feb 1991 A
5008497 Asher Apr 1991 A
5021638 Nopper et al. Jun 1991 A
5053585 Yaniger Oct 1991 A
5107401 Youn Apr 1992 A
5128829 Loew Jul 1992 A
5138119 Demeo Aug 1992 A
5218177 Coleman et al. Jun 1993 A
5220318 Staley Jun 1993 A
5220521 Kikinis Jun 1993 A
5235495 Blair et al. Aug 1993 A
5243162 Kobayashi Sep 1993 A
5253362 Nolan et al. Oct 1993 A
5283559 Kalendra et al. Feb 1994 A
5331443 Stanisci Jul 1994 A
5363075 Fanucchi Nov 1994 A
5375076 Goodrich et al. Dec 1994 A
5404133 Moriike et al. Apr 1995 A
5480118 Cross Jan 1996 A
5491313 Bartley et al. Feb 1996 A
5546271 Gut et al. Aug 1996 A
5548477 Kumar et al. Aug 1996 A
5558577 Kato Sep 1996 A
5581682 Anderson et al. Dec 1996 A
5596700 Darnell et al. Jan 1997 A
5617343 Danielson et al. Apr 1997 A
5618232 Martin Apr 1997 A
5661279 Kenmochi Aug 1997 A
5666112 Crowley et al. Sep 1997 A
5681220 Bertram et al. Oct 1997 A
5737183 Kobayashi et al. Apr 1998 A
5745376 Barker et al. Apr 1998 A
5748114 Koehn May 1998 A
5781406 Hunte Jul 1998 A
5803748 Maddrell et al. Sep 1998 A
5807175 Davis et al. Sep 1998 A
5818361 Acevedo Oct 1998 A
5828770 Leis et al. Oct 1998 A
5842027 Oprescu et al. Nov 1998 A
5874697 Selker et al. Feb 1999 A
5905485 Podoloff May 1999 A
5920317 McDonald Jul 1999 A
5924555 Sadamori et al. Jul 1999 A
5926170 Oba Jul 1999 A
5971635 Wise Oct 1999 A
5995026 Sellers Nov 1999 A
5995081 Kato Nov 1999 A
5995084 Chan et al. Nov 1999 A
6002389 Kasser Dec 1999 A
6002581 Lindsey Dec 1999 A
6005209 Burleson et al. Dec 1999 A
6012714 Worley et al. Jan 2000 A
6014800 Lee Jan 2000 A
6040823 Seffernick et al. Mar 2000 A
6042075 Burch, Jr. Mar 2000 A
6044717 Biegelsen et al. Apr 2000 A
6055705 Komatsu et al. May 2000 A
6061644 Leis May 2000 A
6108200 Fullerton Aug 2000 A
6112797 Colson et al. Sep 2000 A
6128007 Seybold Oct 2000 A
6141388 Servais et al. Oct 2000 A
6178085 Leung Jan 2001 B1
6178443 Lin Jan 2001 B1
6188391 Seely et al. Feb 2001 B1
6254105 Rinde et al. Jul 2001 B1
6279060 Luke et al. Aug 2001 B1
6305073 Badders Oct 2001 B1
6329617 Burgess Dec 2001 B1
6344791 Armstrong Feb 2002 B1
6366440 Kung Apr 2002 B1
6380497 Hashimoto et al. Apr 2002 B1
6437682 Vance Aug 2002 B1
6450046 Maeda Sep 2002 B1
6506983 Babb et al. Jan 2003 B1
6511378 Bhatt et al. Jan 2003 B1
6532147 Christ, Jr. Mar 2003 B1
6543949 Ritchey et al. Apr 2003 B1
6565439 Shinohara et al. May 2003 B2
6585435 Fang Jul 2003 B2
6597347 Yasutake Jul 2003 B1
6600121 Olodort et al. Jul 2003 B1
6603408 Gaba Aug 2003 B1
6603461 Smith, Jr. et al. Aug 2003 B2
6608664 Hasegawa Aug 2003 B1
6617536 Kawaguchi Sep 2003 B2
6651943 Cho et al. Nov 2003 B2
6684166 Bellwood et al. Jan 2004 B2
6685369 Lien Feb 2004 B2
6687614 Ihara et al. Feb 2004 B2
6695273 Iguchi Feb 2004 B2
6704005 Kato et al. Mar 2004 B2
6704864 Philyaw Mar 2004 B1
6721019 Kono et al. Apr 2004 B2
6725318 Sherman et al. Apr 2004 B1
6774888 Genduso Aug 2004 B1
6776546 Kraus et al. Aug 2004 B2
6780019 Ghosh et al. Aug 2004 B1
6781819 Yang et al. Aug 2004 B2
6784869 Clark et al. Aug 2004 B1
6798887 Andre Sep 2004 B1
6813143 Makela Nov 2004 B2
6819316 Schulz et al. Nov 2004 B2
6856506 Doherty et al. Feb 2005 B2
6856789 Pattabiraman et al. Feb 2005 B2
6861961 Sandbach et al. Mar 2005 B2
6898315 Guha May 2005 B2
6909354 Baker et al. Jun 2005 B2
6914197 Doherty et al. Jul 2005 B2
6950950 Sawyers et al. Sep 2005 B2
6962454 Costello Nov 2005 B1
6970957 Oshins et al. Nov 2005 B1
6976799 Kim et al. Dec 2005 B2
7007238 Glaser Feb 2006 B2
7051149 Wang et al. May 2006 B2
7068496 Wong et al. Jun 2006 B2
7083295 Hanna Aug 2006 B1
7091436 Serban Aug 2006 B2
7095404 Vincent et al. Aug 2006 B2
7099149 Krieger et al. Aug 2006 B2
7106222 Ward et al. Sep 2006 B2
7116309 Kimura et al. Oct 2006 B1
7123292 Seeger et al. Oct 2006 B1
7136282 Rebeske Nov 2006 B1
D535292 Shi et al. Jan 2007 S
7159132 Takahashi et al. Jan 2007 B2
7194662 Do et al. Mar 2007 B2
7213323 Baker et al. May 2007 B2
7213991 Chapman et al. May 2007 B2
7224830 Nefian et al. May 2007 B2
7252512 Tai et al. Aug 2007 B2
7260221 Atsmon Aug 2007 B1
7277087 Hill et al. Oct 2007 B2
7280348 Ghosh Oct 2007 B2
7301759 Hsiung Nov 2007 B2
7365967 Zheng Apr 2008 B2
7374312 Feng et al. May 2008 B2
7415676 Fujita Aug 2008 B2
7447922 Asbury et al. Nov 2008 B1
7447934 Dasari et al. Nov 2008 B2
7454712 Schultz Nov 2008 B2
7457108 Ghosh Nov 2008 B2
7469386 Bear et al. Dec 2008 B2
7486165 Ligtenberg et al. Feb 2009 B2
7499037 Lube Mar 2009 B2
7502803 Culter et al. Mar 2009 B2
7539882 Jessup et al. May 2009 B2
7541907 Wang et al. Jun 2009 B2
7542052 Solomon et al. Jun 2009 B2
7558594 Wilson Jul 2009 B2
7559834 York Jul 2009 B1
RE40891 Yasutake Sep 2009 E
7594638 Chan et al. Sep 2009 B2
7620244 Collier Nov 2009 B1
7629966 Anson Dec 2009 B2
7636921 Louie Dec 2009 B2
7639329 Takeda et al. Dec 2009 B2
7639876 Clary et al. Dec 2009 B2
7656392 Bolender Feb 2010 B2
7693654 Dietsch et al. Apr 2010 B1
7722358 Chatterjee et al. May 2010 B2
7728923 Kim et al. Jun 2010 B2
7729493 Krieger et al. Jun 2010 B2
7731147 Rha Jun 2010 B2
7733326 Adiseshan Jun 2010 B1
7761119 Patel Jul 2010 B2
7773076 Pittel et al. Aug 2010 B2
7773121 Huntsberger et al. Aug 2010 B1
7774155 Sato et al. Aug 2010 B2
7777972 Chen et al. Aug 2010 B1
7782342 Koh Aug 2010 B2
7813715 McKillop et al. Oct 2010 B2
7815358 Inditsky Oct 2010 B2
7822338 Wernersson Oct 2010 B2
7865639 McCoy et al. Jan 2011 B2
7884807 Hovden et al. Feb 2011 B2
7893921 Sato Feb 2011 B2
D636397 Green Apr 2011 S
7928964 Kolmykov-Zotov et al. Apr 2011 B2
7932890 Onikiri et al. Apr 2011 B2
7936501 Smith et al. May 2011 B2
7944520 Ichioka et al. May 2011 B2
7945717 Rivalsi May 2011 B2
7970246 Travis et al. Jun 2011 B2
7973771 Geaghan Jul 2011 B2
7978281 Vergith et al. Jul 2011 B2
8016255 Lin Sep 2011 B2
8018386 Qi et al. Sep 2011 B2
8018579 Krah Sep 2011 B1
8026904 Westerman Sep 2011 B2
8053688 Conzola et al. Nov 2011 B2
8059384 Park et al. Nov 2011 B2
8065624 Morin et al. Nov 2011 B2
8069356 Rathi et al. Nov 2011 B2
8077160 Land et al. Dec 2011 B2
8098233 Hotelling et al. Jan 2012 B2
8115499 Osoinach et al. Feb 2012 B2
8117362 Rodriguez et al. Feb 2012 B2
8118274 McClure et al. Feb 2012 B2
8120166 Koizumi et al. Feb 2012 B2
8130203 Westerman Mar 2012 B2
8149219 Lii et al. Apr 2012 B2
8154524 Wilson et al. Apr 2012 B2
8159372 Sherman Apr 2012 B2
8162282 Hu et al. Apr 2012 B2
D659139 Gengler May 2012 S
8169421 Wright et al. May 2012 B2
8189973 Travis et al. May 2012 B2
8229509 Paek et al. Jul 2012 B2
8229522 Kim et al. Jul 2012 B2
8231099 Chen Jul 2012 B2
8243432 Duan et al. Aug 2012 B2
8248791 Wang et al. Aug 2012 B2
8255708 Zhang Aug 2012 B1
8264310 Lauder et al. Sep 2012 B2
8267368 Torii et al. Sep 2012 B2
8269731 Molne Sep 2012 B2
8274784 Franz et al. Sep 2012 B2
8279589 Kim Oct 2012 B2
8289115 Cretella, Jr. et al. Oct 2012 B2
8322290 Mignano Dec 2012 B1
8346206 Andrus et al. Jan 2013 B1
8373664 Wright Feb 2013 B2
8384566 Bocirnea Feb 2013 B2
8387078 Memmott Feb 2013 B2
8387938 Lin Mar 2013 B2
8403576 Merz Mar 2013 B2
8416559 Agata et al. Apr 2013 B2
8424160 Chen Apr 2013 B2
8446359 Doczy et al. May 2013 B2
8464079 Chueh et al. Jun 2013 B2
8477100 Wang et al. Jul 2013 B2
8498100 Whitt, III et al. Jul 2013 B1
8514568 Qiao et al. Aug 2013 B2
8520371 Peng et al. Aug 2013 B2
8543227 Perek et al. Sep 2013 B1
8548608 Perek et al. Oct 2013 B2
8564944 Whitt, III et al. Oct 2013 B2
8569640 Yamada et al. Oct 2013 B2
8570725 Whitt, III et al. Oct 2013 B2
8576031 Lauder et al. Nov 2013 B2
8582280 Ryu Nov 2013 B2
8587701 Tatsuzawa Nov 2013 B2
8596881 Umeno Dec 2013 B2
8599542 Healey et al. Dec 2013 B1
8610015 Whitt, III Dec 2013 B2
8614666 Whitman et al. Dec 2013 B2
8633898 Westerman et al. Jan 2014 B2
8646999 Shaw et al. Feb 2014 B2
8674941 Casparian et al. Mar 2014 B2
8699215 Whitt, III et al. Apr 2014 B2
8719603 Belesiu May 2014 B2
8724302 Whitt et al. May 2014 B2
8744070 Zhang et al. Jun 2014 B2
8744391 Tenbrook et al. Jun 2014 B2
8754854 Hamburgen Jun 2014 B1
8762746 Lachwani et al. Jun 2014 B1
8767388 Ahn et al. Jul 2014 B2
8780540 Whitt, III et al. Jul 2014 B2
8780541 Whitt et al. Jul 2014 B2
8791382 Whitt, III et al. Jul 2014 B2
8797765 Lin et al. Aug 2014 B2
8823652 Linegar et al. Sep 2014 B2
8825187 Hamrick et al. Sep 2014 B1
8830668 Whit, III et al. Sep 2014 B2
8850241 Oler et al. Sep 2014 B2
8854799 Whitt, III et al. Oct 2014 B2
8873227 Whitt, III et al. Oct 2014 B2
8891232 Wang Nov 2014 B2
8896993 Belesiu et al. Nov 2014 B2
8903517 Perek et al. Dec 2014 B2
8908858 Chiu et al. Dec 2014 B2
8918546 Cheah et al. Dec 2014 B2
8934221 Guo Jan 2015 B2
8935774 Belesiu et al. Jan 2015 B2
8939422 Liu et al. Jan 2015 B2
8947864 Whitt, III et al. Feb 2015 B2
8949477 Drasnin Feb 2015 B2
8964376 Chen Feb 2015 B2
8997983 Sajid Apr 2015 B2
9047207 Belesiu et al. Jun 2015 B2
9064654 Whitt, III et al. Jun 2015 B2
9075566 Whitt, III et al. Jul 2015 B2
9098117 Lutz, III et al. Aug 2015 B2
9111703 Whitt, III et al. Aug 2015 B2
9116550 Siddiqui et al. Aug 2015 B2
9134807 Shaw et al. Sep 2015 B2
9134808 Siddiqui et al. Sep 2015 B2
9146620 Whitt et al. Sep 2015 B2
9158383 Shaw et al. Oct 2015 B2
9158384 Whitt, III et al. Oct 2015 B2
9176900 Whitt, III et al. Nov 2015 B2
9176901 Whitt, III et al. Nov 2015 B2
9189428 Pollmann et al. Nov 2015 B2
9268373 Whitt et al. Feb 2016 B2
9275809 Panay et al. Mar 2016 B2
9298236 Oler et al. Mar 2016 B2
9304549 Siddiqui Apr 2016 B2
9304948 Whitman et al. Apr 2016 B2
9304949 Whitman et al. Apr 2016 B2
9348605 Drasnin May 2016 B2
9360893 Bathiche et al. Jun 2016 B2
9411751 Whitt et al. Aug 2016 B2
9426905 Bathiche et al. Aug 2016 B2
9447620 Park et al. Sep 2016 B2
9460029 Shaw et al. Oct 2016 B2
9465412 Belesiu et al. Oct 2016 B2
9618977 Whitt et al. Apr 2017 B2
9619071 Perek et al. Apr 2017 B2
9678542 Whitt et al. Jun 2017 B2
9706089 Beck et al. Jul 2017 B2
9710093 Shaw et al. Jul 2017 B2
9766663 Siddiqui et al. Sep 2017 B2
9793073 Marwah et al. Oct 2017 B2
9852855 Shaw et al. Dec 2017 B2
20010023818 Masaru et al. Sep 2001 A1
20020005108 Ludwig Jan 2002 A1
20020044216 Cha Apr 2002 A1
20020070883 Dosch Jun 2002 A1
20020093436 Lien Jul 2002 A1
20020126446 Miyako et al. Sep 2002 A1
20020134828 Sandbach et al. Sep 2002 A1
20020135457 Sandbach et al. Sep 2002 A1
20020195177 Hinkley et al. Dec 2002 A1
20030000821 Takahashi et al. Jan 2003 A1
20030007648 Currell Jan 2003 A1
20030011576 Sandbach et al. Jan 2003 A1
20030016282 Koizumi Jan 2003 A1
20030044216 Fang Mar 2003 A1
20030051983 Lahr Mar 2003 A1
20030067450 Thursfield et al. Apr 2003 A1
20030108720 Kashino Jun 2003 A1
20030160712 Levy Aug 2003 A1
20030163611 Nagao Aug 2003 A1
20030173195 Federspiel Sep 2003 A1
20030197687 Shetter Oct 2003 A1
20030198008 Leapman et al. Oct 2003 A1
20030231243 Shibutani Dec 2003 A1
20040005184 Kim et al. Jan 2004 A1
20040039862 Hunt et al. Feb 2004 A1
20040046796 Fujita Mar 2004 A1
20040056843 Lin et al. Mar 2004 A1
20040085716 Uke May 2004 A1
20040113956 Bellwood et al. Jun 2004 A1
20040150631 Fleck et al. Aug 2004 A1
20040156168 LeVasseur et al. Aug 2004 A1
20040160734 Yim Aug 2004 A1
20040169641 Bean et al. Sep 2004 A1
20040212598 Kraus et al. Oct 2004 A1
20040212601 Cake et al. Oct 2004 A1
20040258924 Berger et al. Dec 2004 A1
20040268000 Barker et al. Dec 2004 A1
20050030728 Kawashima et al. Feb 2005 A1
20050047773 Satake et al. Mar 2005 A1
20050052831 Chen Mar 2005 A1
20050055498 Beckert et al. Mar 2005 A1
20050057515 Bathiche Mar 2005 A1
20050059489 Kim Mar 2005 A1
20050062715 Tsuji et al. Mar 2005 A1
20050099400 Lee May 2005 A1
20050134717 Misawa Jun 2005 A1
20050146512 Hill et al. Jul 2005 A1
20050153753 Cheng Jul 2005 A1
20050206737 Gim et al. Sep 2005 A1
20050236848 Kim et al. Oct 2005 A1
20050264653 Starkweather et al. Dec 2005 A1
20050264988 Nicolosi Dec 2005 A1
20050283731 Saint-Hilaire et al. Dec 2005 A1
20050285703 Wheeler et al. Dec 2005 A1
20060049920 Sadler et al. Mar 2006 A1
20060049993 Lin et al. Mar 2006 A1
20060061555 Mullen Mar 2006 A1
20060085658 Allen et al. Apr 2006 A1
20060092139 Sharma May 2006 A1
20060096392 Inkster et al. May 2006 A1
20060102020 Takada et al. May 2006 A1
20060102914 Smits et al. May 2006 A1
20060125799 Hillis et al. Jun 2006 A1
20060132423 Travis Jun 2006 A1
20060132456 Anson Jun 2006 A1
20060154725 Glaser et al. Jul 2006 A1
20060155391 Pistemaa et al. Jul 2006 A1
20060156415 Rubinstein et al. Jul 2006 A1
20060174143 Sawyers et al. Aug 2006 A1
20060176377 Miyasaka Aug 2006 A1
20060181514 Newman Aug 2006 A1
20060187216 Trent, Jr. et al. Aug 2006 A1
20060192763 Ziemkowski Aug 2006 A1
20060195522 Miyazaki Aug 2006 A1
20060220465 Kingsmore et al. Oct 2006 A1
20060265617 Priborsky Nov 2006 A1
20060267931 Vainio et al. Nov 2006 A1
20060272429 Ganapathi et al. Dec 2006 A1
20070003267 Shibutani Jan 2007 A1
20070047221 Park Mar 2007 A1
20070056385 Lorenz Mar 2007 A1
20070062089 Homer et al. Mar 2007 A1
20070069153 Pai-Paranjape et al. Mar 2007 A1
20070072474 Beasley et al. Mar 2007 A1
20070117600 Robertson et al. May 2007 A1
20070121956 Bai et al. May 2007 A1
20070127205 Kuo Jun 2007 A1
20070145945 McGinley et al. Jun 2007 A1
20070172229 Wernersson Jul 2007 A1
20070176902 Newman et al. Aug 2007 A1
20070178891 Louch et al. Aug 2007 A1
20070182663 Biech Aug 2007 A1
20070182722 Hotelling et al. Aug 2007 A1
20070185590 Reindel et al. Aug 2007 A1
20070200830 Yamamoto Aug 2007 A1
20070220708 Lewis Sep 2007 A1
20070222766 Bolender Sep 2007 A1
20070230227 Palmer Oct 2007 A1
20070234420 Novotney et al. Oct 2007 A1
20070236408 Yamaguchi et al. Oct 2007 A1
20070236475 Wherry Oct 2007 A1
20070236873 Yukawa et al. Oct 2007 A1
20070242037 Son Oct 2007 A1
20070247432 Oakley Oct 2007 A1
20070252674 Nelson et al. Nov 2007 A1
20070257821 Son et al. Nov 2007 A1
20070260892 Paul et al. Nov 2007 A1
20070268273 Westerman et al. Nov 2007 A1
20070274094 Schultz et al. Nov 2007 A1
20070274095 Destain Nov 2007 A1
20070283179 Burnett et al. Dec 2007 A1
20070296709 Guanghai Dec 2007 A1
20070297125 Maatta Dec 2007 A1
20070297625 Hjort et al. Dec 2007 A1
20080001924 de los Reyes et al. Jan 2008 A1
20080005423 Jacobs et al. Jan 2008 A1
20080013809 Zhu et al. Jan 2008 A1
20080018613 Kim et al. Jan 2008 A1
20080042978 Perez-Noguera Feb 2008 A1
20080053222 Ehrensvard et al. Mar 2008 A1
20080059888 Dunko Mar 2008 A1
20080074398 Wright Mar 2008 A1
20080104437 Lee May 2008 A1
20080129520 Lee Jun 2008 A1
20080151478 Chern Jun 2008 A1
20080158185 Westerman Jul 2008 A1
20080167832 Soss Jul 2008 A1
20080174570 Jobs et al. Jul 2008 A1
20080186660 Yang Aug 2008 A1
20080186683 Ligtenberg et al. Aug 2008 A1
20080219025 Spitzer et al. Sep 2008 A1
20080228969 Cheah et al. Sep 2008 A1
20080238884 Harish Oct 2008 A1
20080253822 Matias Oct 2008 A1
20080273297 Kumar Nov 2008 A1
20080297878 Brown et al. Dec 2008 A1
20080307242 Qu Dec 2008 A1
20080309636 Feng et al. Dec 2008 A1
20080316002 Brunet et al. Dec 2008 A1
20080316183 Westerman et al. Dec 2008 A1
20080320190 Lydon et al. Dec 2008 A1
20090009476 Daley, III Jan 2009 A1
20090073060 Shimasaki et al. Mar 2009 A1
20090073957 Newland et al. Mar 2009 A1
20090079639 Hotta et al. Mar 2009 A1
20090083562 Park et al. Mar 2009 A1
20090089600 Nousiainen Apr 2009 A1
20090096756 Lube Apr 2009 A1
20090102805 Meijer et al. Apr 2009 A1
20090127005 Zachut et al. May 2009 A1
20090131134 Baerlocher et al. May 2009 A1
20090140985 Liu Jun 2009 A1
20090158221 Nielsen et al. Jun 2009 A1
20090163147 Steigerwald et al. Jun 2009 A1
20090174687 Ciesla et al. Jul 2009 A1
20090174759 Yeh et al. Jul 2009 A1
20090177906 Paniagua, Jr. et al. Jul 2009 A1
20090187860 Fleck et al. Jul 2009 A1
20090189873 Peterson Jul 2009 A1
20090195497 Fitzgerald et al. Aug 2009 A1
20090195518 Mattice et al. Aug 2009 A1
20090201254 Rais Aug 2009 A1
20090207144 Bridger Aug 2009 A1
20090231275 Odgers Sep 2009 A1
20090239586 Boeve et al. Sep 2009 A1
20090244009 Staats et al. Oct 2009 A1
20090244832 Behar et al. Oct 2009 A1
20090244872 Yan Oct 2009 A1
20090251008 Sugaya Oct 2009 A1
20090259865 Sheynblat et al. Oct 2009 A1
20090262492 Whitchurch et al. Oct 2009 A1
20090265670 Kim et al. Oct 2009 A1
20090269943 Palli et al. Oct 2009 A1
20090285491 Ravenscroft et al. Nov 2009 A1
20090296331 Choy Dec 2009 A1
20090302175 Torii et al. Dec 2009 A1
20090303137 Kusaka et al. Dec 2009 A1
20090303204 Nasiri et al. Dec 2009 A1
20090315830 Westerman Dec 2009 A1
20090320244 Lin Dec 2009 A1
20090321490 Groene et al. Dec 2009 A1
20100001963 Doray et al. Jan 2010 A1
20100006412 Wang et al. Jan 2010 A1
20100013319 Kamiyama et al. Jan 2010 A1
20100023869 Saint-Hilaire et al. Jan 2010 A1
20100026656 Hotelling et al. Feb 2010 A1
20100038821 Jenkins et al. Feb 2010 A1
20100039081 Sip Feb 2010 A1
20100039764 Locker et al. Feb 2010 A1
20100045609 Do et al. Feb 2010 A1
20100045633 Gettemy Feb 2010 A1
20100051356 Stern et al. Mar 2010 A1
20100051432 Lin et al. Mar 2010 A1
20100052880 Laitinen et al. Mar 2010 A1
20100053534 Hsieh et al. Mar 2010 A1
20100054435 Louch et al. Mar 2010 A1
20100056130 Louch et al. Mar 2010 A1
20100072334 Le Gette et al. Mar 2010 A1
20100073329 Raman et al. Mar 2010 A1
20100077237 Sawyers Mar 2010 A1
20100079379 Demuynck et al. Apr 2010 A1
20100081377 Chatterjee et al. Apr 2010 A1
20100083108 Rider et al. Apr 2010 A1
20100085321 Pundsack Apr 2010 A1
20100100752 Chueh et al. Apr 2010 A1
20100102182 Lin Apr 2010 A1
20100103112 Yoo et al. Apr 2010 A1
20100103611 Yang et al. Apr 2010 A1
20100105443 Vaisanen Apr 2010 A1
20100106983 Kasprzak et al. Apr 2010 A1
20100115309 Carvalho et al. May 2010 A1
20100117993 Kent May 2010 A1
20100123686 Klinghult et al. May 2010 A1
20100128427 Iso May 2010 A1
20100133398 Chiu et al. Jun 2010 A1
20100142130 Wang et al. Jun 2010 A1
20100146317 Challener et al. Jun 2010 A1
20100148995 Elias Jun 2010 A1
20100148999 Casparian et al. Jun 2010 A1
20100149104 Sim et al. Jun 2010 A1
20100149111 Olien Jun 2010 A1
20100149134 Westerman et al. Jun 2010 A1
20100149377 Shintani et al. Jun 2010 A1
20100156798 Archer Jun 2010 A1
20100156913 Ortega et al. Jun 2010 A1
20100161522 Tirpak et al. Jun 2010 A1
20100164857 Liu et al. Jul 2010 A1
20100164897 Morin et al. Jul 2010 A1
20100171891 Kaji et al. Jul 2010 A1
20100174421 Tsai et al. Jul 2010 A1
20100180063 Ananny et al. Jul 2010 A1
20100185877 Chueh et al. Jul 2010 A1
20100185989 Shiplacoff et al. Jul 2010 A1
20100188299 Rinehart et al. Jul 2010 A1
20100201308 Lindholm Aug 2010 A1
20100205472 Tupman et al. Aug 2010 A1
20100206614 Park et al. Aug 2010 A1
20100206644 Yeh Aug 2010 A1
20100207774 Song Aug 2010 A1
20100214257 Wussler et al. Aug 2010 A1
20100220205 Lee et al. Sep 2010 A1
20100222110 Kim et al. Sep 2010 A1
20100231498 Large et al. Sep 2010 A1
20100231510 Sampsell et al. Sep 2010 A1
20100231522 Li Sep 2010 A1
20100231556 Mines et al. Sep 2010 A1
20100235546 Terlizzi et al. Sep 2010 A1
20100238075 Pourseyed Sep 2010 A1
20100238138 Goertz et al. Sep 2010 A1
20100238620 Fish Sep 2010 A1
20100245106 Miller Sep 2010 A1
20100245221 Khan Sep 2010 A1
20100250975 Gill et al. Sep 2010 A1
20100250988 Okuda et al. Sep 2010 A1
20100259482 Ball Oct 2010 A1
20100259876 Kim Oct 2010 A1
20100265182 Ball et al. Oct 2010 A1
20100271771 Wu et al. Oct 2010 A1
20100274932 Kose Oct 2010 A1
20100279768 Huang et al. Nov 2010 A1
20100289457 Onnerud et al. Nov 2010 A1
20100295812 Burns et al. Nov 2010 A1
20100302378 Marks et al. Dec 2010 A1
20100304793 Kim et al. Dec 2010 A1
20100306538 Thomas et al. Dec 2010 A1
20100308778 Yamazaki et al. Dec 2010 A1
20100308844 Day et al. Dec 2010 A1
20100309617 Wang et al. Dec 2010 A1
20100313680 Joung et al. Dec 2010 A1
20100315345 Laitinen Dec 2010 A1
20100315348 Jellicoe et al. Dec 2010 A1
20100315373 Steinhauser et al. Dec 2010 A1
20100321339 Kimmel Dec 2010 A1
20100321877 Moser Dec 2010 A1
20100324457 Bean et al. Dec 2010 A1
20100325155 Skinner et al. Dec 2010 A1
20100331059 Apgar et al. Dec 2010 A1
20110011650 Klinghult Jan 2011 A1
20110012873 Prest et al. Jan 2011 A1
20110015894 Chu Jan 2011 A1
20110019123 Prest et al. Jan 2011 A1
20110031287 Le Gette et al. Feb 2011 A1
20110032127 Roush Feb 2011 A1
20110036965 Zhang et al. Feb 2011 A1
20110037721 Cranfill et al. Feb 2011 A1
20110043990 Mickey et al. Feb 2011 A1
20110050576 Forutanpour et al. Mar 2011 A1
20110050626 Porter et al. Mar 2011 A1
20110055407 Lydon et al. Mar 2011 A1
20110057724 Pabon Mar 2011 A1
20110060926 Brooks et al. Mar 2011 A1
20110069148 Jones et al. Mar 2011 A1
20110074688 Hull et al. Mar 2011 A1
20110102326 Casparian et al. May 2011 A1
20110102356 Kemppinen et al. May 2011 A1
20110102752 Chen et al. May 2011 A1
20110107958 Pance et al. May 2011 A1
20110108401 Yamada et al. May 2011 A1
20110109572 Deslippe et al. May 2011 A1
20110109573 Deslippe et al. May 2011 A1
20110113368 Carvajal et al. May 2011 A1
20110115738 Suzuki et al. May 2011 A1
20110115747 Powell et al. May 2011 A1
20110117970 Choi May 2011 A1
20110134032 Chiu et al. Jun 2011 A1
20110134043 Chen Jun 2011 A1
20110134112 Koh et al. Jun 2011 A1
20110149094 Chen et al. Jun 2011 A1
20110157037 Shamir et al. Jun 2011 A1
20110157046 Lee et al. Jun 2011 A1
20110157087 Kanehira et al. Jun 2011 A1
20110163955 Nasiri et al. Jul 2011 A1
20110164357 Yeom et al. Jul 2011 A1
20110164370 McClure et al. Jul 2011 A1
20110167181 Minoo et al. Jul 2011 A1
20110167287 Walsh et al. Jul 2011 A1
20110167391 Momeyer et al. Jul 2011 A1
20110167992 Eventoff et al. Jul 2011 A1
20110169762 Weiss Jul 2011 A1
20110176035 Poulsen Jul 2011 A1
20110179864 Raasch et al. Jul 2011 A1
20110184646 Wong et al. Jul 2011 A1
20110184824 George et al. Jul 2011 A1
20110188199 Pan Aug 2011 A1
20110191480 Kobayashi Aug 2011 A1
20110193787 Morishige et al. Aug 2011 A1
20110193938 Oderwald et al. Aug 2011 A1
20110199389 Lu et al. Aug 2011 A1
20110202878 Park et al. Aug 2011 A1
20110205372 Miramontes Aug 2011 A1
20110216266 Travis Sep 2011 A1
20110221678 Davydov Sep 2011 A1
20110227913 Hyndman Sep 2011 A1
20110231682 Kakish et al. Sep 2011 A1
20110234494 Peterson et al. Sep 2011 A1
20110241999 Thier Oct 2011 A1
20110242138 Tribble Oct 2011 A1
20110248152 Svajda et al. Oct 2011 A1
20110248920 Larsen Oct 2011 A1
20110248941 Abdo et al. Oct 2011 A1
20110261001 Liu Oct 2011 A1
20110261083 Wilson Oct 2011 A1
20110265287 Li et al. Nov 2011 A1
20110266672 Sylvester Nov 2011 A1
20110267272 Meyer et al. Nov 2011 A1
20110273475 Herz et al. Nov 2011 A1
20110285555 Bocirnea Nov 2011 A1
20110290686 Huang Dec 2011 A1
20110291958 Wu et al. Dec 2011 A1
20110295697 Boston et al. Dec 2011 A1
20110297566 Gallagher et al. Dec 2011 A1
20110298919 Maglaque Dec 2011 A1
20110302518 Zhang Dec 2011 A1
20110304577 Brown Dec 2011 A1
20110305875 Sanford et al. Dec 2011 A1
20110310548 Ahn et al. Dec 2011 A1
20110314425 Chiang Dec 2011 A1
20110316807 Corrion Dec 2011 A1
20110320204 Locker et al. Dec 2011 A1
20120002820 Leichter Jan 2012 A1
20120007821 Zaliva Jan 2012 A1
20120011462 Westerman et al. Jan 2012 A1
20120013490 Pance Jan 2012 A1
20120013519 Hakansson et al. Jan 2012 A1
20120019445 Liu Jan 2012 A1
20120020490 Leichter Jan 2012 A1
20120023401 Arscott et al. Jan 2012 A1
20120023459 Westerman Jan 2012 A1
20120024682 Huang et al. Feb 2012 A1
20120026048 Vazquez et al. Feb 2012 A1
20120026096 Ku Feb 2012 A1
20120026110 Yamano Feb 2012 A1
20120032887 Chiu et al. Feb 2012 A1
20120032891 Parivar Feb 2012 A1
20120032901 Kwon Feb 2012 A1
20120032917 Yamaguchi Feb 2012 A1
20120038495 Ishikawa Feb 2012 A1
20120044179 Hudson Feb 2012 A1
20120047368 Chinn et al. Feb 2012 A1
20120050975 Garelli et al. Mar 2012 A1
20120062564 Miyashita Mar 2012 A1
20120068919 Lauder et al. Mar 2012 A1
20120069540 Lauder et al. Mar 2012 A1
20120075249 Hoch Mar 2012 A1
20120077384 Bar-Niv et al. Mar 2012 A1
20120081316 Sirpal et al. Apr 2012 A1
20120092279 Martin Apr 2012 A1
20120094257 Pillischer et al. Apr 2012 A1
20120099749 Rubin et al. Apr 2012 A1
20120103778 Obata et al. May 2012 A1
20120113137 Nomoto May 2012 A1
20120113579 Agata et al. May 2012 A1
20120115553 Mahe et al. May 2012 A1
20120117409 Lee et al. May 2012 A1
20120127118 Nolting et al. May 2012 A1
20120127126 Mattice et al. May 2012 A1
20120139727 Houvener et al. Jun 2012 A1
20120140396 Zeliff et al. Jun 2012 A1
20120145525 Ishikawa Jun 2012 A1
20120156875 Srinivas et al. Jun 2012 A1
20120162693 Ito Jun 2012 A1
20120162889 Han Jun 2012 A1
20120175487 Goto Jul 2012 A1
20120182242 Lindahl et al. Jul 2012 A1
20120182249 Endo et al. Jul 2012 A1
20120194393 Uttermann et al. Aug 2012 A1
20120194448 Rothkopf Aug 2012 A1
20120200802 Large Aug 2012 A1
20120206937 Travis et al. Aug 2012 A1
20120212438 Vaisanen Aug 2012 A1
20120218194 Silverman Aug 2012 A1
20120221877 Prabu Aug 2012 A1
20120223866 Ayala et al. Sep 2012 A1
20120224073 Miyahara Sep 2012 A1
20120227259 Badaye et al. Sep 2012 A1
20120229634 Laett et al. Sep 2012 A1
20120235635 Sato Sep 2012 A1
20120242584 Tuli Sep 2012 A1
20120243165 Chang et al. Sep 2012 A1
20120246377 Bhesania Sep 2012 A1
20120249443 Anderson et al. Oct 2012 A1
20120250873 Bakalos et al. Oct 2012 A1
20120256829 Dodge Oct 2012 A1
20120256959 Ye et al. Oct 2012 A1
20120260177 Sehrer Oct 2012 A1
20120274811 Bakin Nov 2012 A1
20120298491 Ozias et al. Nov 2012 A1
20120299872 Nishikawa et al. Nov 2012 A1
20120300275 Vilardell et al. Nov 2012 A1
20120306747 Davidson et al. Dec 2012 A1
20120312955 Randolph Dec 2012 A1
20130002562 Leskela Jan 2013 A1
20130009413 Chiu et al. Jan 2013 A1
20130015311 Kim Jan 2013 A1
20130016468 Oh Jan 2013 A1
20130021289 Chen et al. Jan 2013 A1
20130027867 Lauder et al. Jan 2013 A1
20130031353 Noro Jan 2013 A1
20130038541 Bakker Feb 2013 A1
20130044059 Fu Feb 2013 A1
20130044074 Park et al. Feb 2013 A1
20130046397 Fadell et al. Feb 2013 A1
20130063873 Wodrich et al. Mar 2013 A1
20130067126 Casparian et al. Mar 2013 A1
20130067259 Freiwald et al. Mar 2013 A1
20130073877 Radke Mar 2013 A1
20130076614 Ive et al. Mar 2013 A1
20130076617 Csaszar et al. Mar 2013 A1
20130076635 Lin Mar 2013 A1
20130082824 Colley Apr 2013 A1
20130088431 Ballagas et al. Apr 2013 A1
20130093679 Dickinson et al. Apr 2013 A1
20130100030 Los et al. Apr 2013 A1
20130100082 Bakin et al. Apr 2013 A1
20130106766 Yilmaz et al. May 2013 A1
20130107144 Marhefka et al. May 2013 A1
20130118878 Purcocks May 2013 A1
20130135214 Li et al. May 2013 A1
20130151944 Lin Jun 2013 A1
20130154959 Lindsay et al. Jun 2013 A1
20130159749 Moeglein et al. Jun 2013 A1
20130162554 Lauder et al. Jun 2013 A1
20130172906 Olson et al. Jul 2013 A1
20130191741 Dickinson et al. Jul 2013 A1
20130212483 Brakensiek et al. Aug 2013 A1
20130215035 Guard Aug 2013 A1
20130217451 Komiyama et al. Aug 2013 A1
20130222272 Martin, Jr. Aug 2013 A1
20130222274 Mori et al. Aug 2013 A1
20130222275 Byrd et al. Aug 2013 A1
20130222323 McKenzie Aug 2013 A1
20130226794 Englebardt Aug 2013 A1
20130227836 Whitt, III Sep 2013 A1
20130228023 Drasnin Sep 2013 A1
20130228433 Shaw Sep 2013 A1
20130228434 Whitt, III Sep 2013 A1
20130228435 Whitt, III Sep 2013 A1
20130228439 Whitt, III Sep 2013 A1
20130229100 Siddiqui Sep 2013 A1
20130229335 Whitman Sep 2013 A1
20130229347 Lutz, III Sep 2013 A1
20130229350 Shaw et al. Sep 2013 A1
20130229351 Whitt, III Sep 2013 A1
20130229354 Whitt, III et al. Sep 2013 A1
20130229356 Marwah et al. Sep 2013 A1
20130229363 Whitman Sep 2013 A1
20130229366 Dighde Sep 2013 A1
20130229386 Bathiche Sep 2013 A1
20130229534 Panay Sep 2013 A1
20130229568 Belesiu Sep 2013 A1
20130229570 Beck et al. Sep 2013 A1
20130229756 Whitt, III Sep 2013 A1
20130229757 Whitt, III et al. Sep 2013 A1
20130229758 Belesiu Sep 2013 A1
20130229759 Whitt, III Sep 2013 A1
20130229760 Whitt, III Sep 2013 A1
20130229761 Shaw Sep 2013 A1
20130229762 Whitt, III Sep 2013 A1
20130229773 Siddiqui Sep 2013 A1
20130230346 Shaw Sep 2013 A1
20130231755 Perek Sep 2013 A1
20130232280 Perek Sep 2013 A1
20130232348 Oler Sep 2013 A1
20130232349 Oler Sep 2013 A1
20130232350 Belesiu et al. Sep 2013 A1
20130232353 Belesiu Sep 2013 A1
20130232571 Belesiu Sep 2013 A1
20130232742 Burnett et al. Sep 2013 A1
20130241860 Ciesla et al. Sep 2013 A1
20130242495 Bathiche et al. Sep 2013 A1
20130262886 Nishimura Oct 2013 A1
20130268897 Li et al. Oct 2013 A1
20130285922 Alberth, Jr. et al. Oct 2013 A1
20130300590 Dietz Nov 2013 A1
20130300647 Drasnin Nov 2013 A1
20130301199 Whitt Nov 2013 A1
20130301206 Whitt Nov 2013 A1
20130304941 Drasnin Nov 2013 A1
20130321992 Liu et al. Dec 2013 A1
20130322000 Whitt Dec 2013 A1
20130322001 Whitt Dec 2013 A1
20130329360 Aldana Dec 2013 A1
20130332628 Panay Dec 2013 A1
20130339757 Reddy Dec 2013 A1
20130346882 Shiplacoff et al. Dec 2013 A1
20140012401 Perek Jan 2014 A1
20140043275 Whitman Feb 2014 A1
20140048399 Whitt, III Feb 2014 A1
20140069791 Chu et al. Mar 2014 A1
20140078063 Bathiche et al. Mar 2014 A1
20140085814 Kielland Mar 2014 A1
20140119802 Shaw May 2014 A1
20140139989 Mori et al. May 2014 A1
20140167585 Kuan et al. Jun 2014 A1
20140185215 Whitt Jul 2014 A1
20140185220 Whitt Jul 2014 A1
20140204514 Whitt Jul 2014 A1
20140204515 Whitt Jul 2014 A1
20140247546 Whitt et al. Sep 2014 A1
20140291134 Whitt et al. Oct 2014 A1
20140293534 Siddiqui Oct 2014 A1
20140313665 Delpier et al. Oct 2014 A1
20140342976 Goad et al. Nov 2014 A1
20140362506 Whitt, III et al. Dec 2014 A1
20140372914 Byrd et al. Dec 2014 A1
20140374230 Shaw et al. Dec 2014 A1
20140379942 Perek et al. Dec 2014 A1
20150005953 Fadell et al. Jan 2015 A1
20150022995 Peterson et al. Jan 2015 A1
20150036274 Belesiu et al. Feb 2015 A1
20150227212 Whitt, III et al. Aug 2015 A1
20150234478 Belesiu et al. Aug 2015 A1
20150261262 Whitt, III et al. Sep 2015 A1
20150311014 Shaw et al. Oct 2015 A1
20150378392 Siddiqui et al. Dec 2015 A1
20160124467 Whitt et al. May 2016 A1
20160170935 Drasnin Jun 2016 A1
20160209884 Whitt et al. Jul 2016 A1
20160299537 Whitt et al. Oct 2016 A1
20160306396 Bathiche et al. Oct 2016 A1
20170131819 Shaw et al. May 2017 A1
20170147084 Whitt et al. May 2017 A1
20170177038 Whitt et al. Jun 2017 A1
Foreign Referenced Citations (144)
Number Date Country
990023 Jun 1976 CA
2363007 Feb 2000 CN
1352767 Jun 2002 CN
1489031 Apr 2004 CN
1537223 Oct 2004 CN
1653411 Aug 2005 CN
1787605 Jun 2006 CN
1808362 Jul 2006 CN
200947406 Sep 2007 CN
101198925 Jun 2008 CN
101335147 Dec 2008 CN
101366001 Feb 2009 CN
101410781 Apr 2009 CN
101452334 Jun 2009 CN
101464750 Jun 2009 CN
101490642 Jul 2009 CN
101500388 Aug 2009 CN
101644979 Feb 2010 CN
101675406 Mar 2010 CN
101681189 Mar 2010 CN
101765825 Jun 2010 CN
101904661 Dec 2010 CN
101908428 Dec 2010 CN
102004559 Apr 2011 CN
102004577 Apr 2011 CN
1102012763 Apr 2011 CN
102096494 Jun 2011 CN
102112947 Jun 2011 CN
201853163 Jun 2011 CN
102117121 Jul 2011 CN
102124532 Jul 2011 CN
102138113 Jul 2011 CN
102147643 Aug 2011 CN
102156510 Aug 2011 CN
201936231 Aug 2011 CN
102214040 Oct 2011 CN
102292687 Dec 2011 CN
102339683 Feb 2012 CN
102356624 Feb 2012 CN
103455149 Dec 2013 CN
203480365 Mar 2014 CN
19924694 Mar 2001 DE
10116556 Oct 2002 DE
645726 Mar 1995 EP
1003188 May 2000 EP
1223722 Jul 2002 EP
1480029 Nov 2004 EP
1591891 Nov 2005 EP
1983411 Oct 2008 EP
2006869 Dec 2008 EP
2009660 Dec 2008 EP
2026178 Feb 2009 EP
2207076 Jul 2010 EP
2353978 Aug 2011 EP
2400365 Dec 2011 EP
2410408 Jan 2012 EP
2423787 Feb 2012 EP
2068643 Aug 1981 GB
2123213 Jan 1984 GB
2305780 Apr 1997 GB
2381584 May 2003 GB
2402460 Dec 2004 GB
2482931 Feb 2012 GB
2482932 Feb 2012 GB
52107722 Sep 1977 JP
56108127 Aug 1981 JP
S57126617 Aug 1982 JP
S5810335 Jan 1983 JP
S93824 Jan 1984 JP
6014315 Jan 1985 JP
S6037923 Aug 1985 JP
S60216479 Oct 1985 JP
H0195596 Apr 1989 JP
H0245820 Feb 1990 JP
04363823 Dec 1992 JP
4363823 Dec 1992 JP
06250761 Sep 1994 JP
H06250761 Sep 1994 JP
08273471 Oct 1996 JP
H1078357 Mar 1998 JP
10326124 Dec 1998 JP
1173239 Mar 1999 JP
11338575 Dec 1999 JP
2000010654 Jan 2000 JP
2000035849 Feb 2000 JP
2001142564 May 2001 JP
2002170458 Jun 2002 JP
2002358852 Dec 2002 JP
2002366277 Dec 2002 JP
2003256106 Sep 2003 JP
2003257282 Sep 2003 JP
2003529837 Oct 2003 JP
2004038950 Feb 2004 JP
2004234662 Aug 2004 JP
2004259173 Sep 2004 JP
2005031555 Feb 2005 JP
2005078104 Mar 2005 JP
2005117161 Apr 2005 JP
2005071728 Aug 2005 JP
2005302447 Oct 2005 JP
2006004063 Jan 2006 JP
2006093461 Apr 2006 JP
2006127486 May 2006 JP
2006163459 Jun 2006 JP
2006266812 Oct 2006 JP
2006292375 Oct 2006 JP
2006294361 Oct 2006 JP
2007042004 Feb 2007 JP
2008140182 Jun 2008 JP
2008152805 Jul 2008 JP
2008181523 Aug 2008 JP
2008298435 Dec 2008 JP
2009009854 Jan 2009 JP
2010244514 Oct 2010 JP
2010272384 Dec 2010 JP
2011150737 Aug 2011 JP
2011170834 Sep 2011 JP
2011221640 Nov 2011 JP
2012027875 Feb 2012 JP
2003077368 Mar 2014 JP
20010107055 Dec 2001 KR
20050014299 Feb 2005 KR
20060003093 Jan 2006 KR
20080006404 Jan 2008 KR
20090029411 Mar 2009 KR
20100022059 Feb 2010 KR
20100067366 Jun 2010 KR
20100115675 Oct 2010 KR
102011008717 Aug 2011 KR
20110109791 Oct 2011 KR
20110120002 Nov 2011 KR
20110122333 Nov 2011 KR
101113530 Feb 2012 KR
WO-9919995 Apr 1999 WO
WO-2006044818 Apr 2006 WO
WO-2007103631 Sep 2007 WO
WO-2007112172 Oct 2007 WO
WO-2008117529 Oct 2008 WO
WO-2009034484 Mar 2009 WO
WO-2010074116 Jul 2010 WO
WO-2011049609 Apr 2011 WO
WO-2011071096 Jun 2011 WO
WO-2011138200 Nov 2011 WO
WO-2014209818 Dec 2014 WO
Non-Patent Literature Citations (560)
Entry
FingerWorks (NPL dated 2002, attached as pdf file named fingerworks_touchstream.pdf).
“ACPI Docking for Windows Operating Systems”, Retrieved from: <http://www.scritube.com/limba/engleza/software/ACPI-Docking-for-Windows-Opera331824193.php> on Jul. 6, 2012, 10 pages.
“Accessing Device Sensors”, retrieved from <https://developer.palm.com/content/api/dev-guide/pdk/accessing-device-sensors.html> on May 25, 2012, 4 pages.
“First One Handed Fabric Keyboard with Bluetooth Wireless Technology”, Retrieved from: <http://press.xtvworld.com/article3817.html> on May 8, 2012,(Jan. 6, 2005),2 pages.
“Force and Position Sensing Resistors: An Emerging Technology”, Interlink Electronics, Available at <http://staff.science.uva.nl/˜vlaander/docu/FSR/An_Exploring_Technology.pdf>,(Feb. 1990), pp. 1-6.
“Frogpad Introduces Weareable Fabric Keyboard with Bluetooth Technology”, Retrieved from: <http://www.geekzone.co.nz/content.asp?contentid=3898> on May 7, 2012,(Jan. 7, 2005),3 pages.
“Incipio LG G-Slate Premium Kickstand Case—Black Nylon”, Retrieved from: <http://www.amazon.com/Incipio-G-Slate-Premium-Kickstand-Case/dp/B004ZKP916> on May 8, 2012,4 pages.
“Membrane Keyboards & Membrane Keypads”, Retrieved from: <http://www.pannam.com/> on May 9, 2012,(Mar. 4, 2009),2 pages.
“Motion Sensors”, Android Developers, retrieved from <http://developer.android.com/guide/topics/sensors/sensors_motion.html> on May 25, 2012,7 pages.
“Position Sensors”, Android Developers, retrieved from <http://developer.android.com/guide/topics/sensors/sensors_position.html> on May 25, 2012,5 pages.
“SolRxTM E-Series Multidirectional Phototherapy ExpandableTM 2-Bulb Full Body Panel System”, Retrieved from: <http://www.solarcsystems.com/us_multidirectional_uv_light_therapy_1_intro.html > on Jul. 25, 2012,(2011), 4 pages.
“Virtualization Getting Started Guide”, Red Hat Enterprise Linux 6, Edition 0.2, retrieved from <http://docs.redhat.com/docs/en-US/Red_Hat_Enterprise_Linux/6/html-single/Virtualization_Getting_Started_Guide/index.html> on Jun. 13, 2012, 24 pages.
Block, Steve et al., “DeviceOrientation Event Specification”, W3C, Editors Draft, retrieved from <https://developer.palm.com/content/api/dev-guide/pdk/accessing-device-sensors.html> on May 25, 2012,(Jul. 12, 2011), 14 pages.
Brown, Rich “Microsoft Shows Off Pressure-Sensitive Keyboard”, retrieved from <http://news.cnet.com/8301-17938_105-10304792-1.html> on May 7, 2012, (Aug. 6, 2009), 2 pages.
Butler, Alex et al., “SideSight: Multi-“touch” Interaction around Small Devices”, In the proceedings of the 21st annual ACM symposium on User interface software and technology., retrieved from <http://research.microsoft.com/pubs/132534/sidesight_crv3.pdf> on May 29, 2012,(Oct. 19, 2008), 4 pages.
Crider, Michael “Sony Slate Concept Tablet “Grows” a Kickstand”, Retrieved from: <http://androidcommunity.com/sony-slate-concept-tablet-grows-a-kickstand-20120116/> on May 4, 2012,(Jan. 16, 2012), 9 pages.
Dietz, Paul H., et al., “A Practical Pressure Sensitive Computer Keyboard”, In Proceedings of UIST 2009,(Oct. 2009), 4 pages.
Glatt, Jeff “Channel and Key Pressure (Aftertouch).”, Retrieved from: <http://home.roadrunner.com/˜jgglatt/tutr/touch.htm> on Jun. 11, 2012, 2 pages.
Hanlon, Mike “ElekTex Smart Fabric Keyboard Goes Wireless”, Retrieved from: <http://www.gizmag.com/go/5048/ > on May 7, 2012,(Jan. 15, 2006), 5 pages.
Kaur, Sukhmani “Vincent Liew's redesigned laptop satisfies ergonomic needs”, Retrieved from: <http://www.designbuzz.com/entry/vincent-liew-s-redesigned-laptop-satisfies-ergonomic-needs/> on Jul. 27, 2012,(Jun. 21, 2010), 4 pages.
Khuntontong, Puttachat et al., “Fabrication of Molded Interconnection Devices by Ultrasonic Hot Embossing on Thin Polymer Films”, IEEE Transactions on Electronics Packaging Manufacturing, vol. 32, No. 3,(Jul. 2009), pp. 152-156.
Linderholm, Owen “Logitech Shows Cloth Keyboard for PDAs”, Retrieved from: <http://www.pcworld.com/article/89084/logitech_shows_cloth_keyboard_for_pdas.html> on May 7, 2012,(Mar. 15, 2002), 5 pages.
McLellan, Charles “Eleksen Wireless Fabric Keyboard: a first look”, Retrieved from: <http://www.zdnetasia.com/eleksen-wireless-fabric-keyboard-a-first-look-40278954.htm> on May 7, 2012,(Jul. 17, 2006), 9 pages.
Post, E.R. et al., “E-Broidery: Design and Fabrication of Textile-Based Computing”, IBM Systems Journal, vol. 39, Issue 3 & 4,(Jul. 2000), pp. 840-860.
Purcher, Jack “Apple is Paving the Way for a New 3D GUI for IOS Devices”, Retrieved from: <http://www.patentlyapple.com/patently-apple/2012/01/apple-is-paving-the-way-for-a-new-3d-gui-for-ios-devices.html> on Jun. 4, 2012,(Jan. 12, 2012), 15 pages.
Takamatsu, Seiichi et al., “Flexible Fabric Keyboard with Conductive Polymer-Coated Fibers”, In Proceedings of Sensors 2011,(Oct. 28, 2011), 4 pages.
Zhang, et al., “Model-Based Development of Dynamically Adaptive Software”, In Proceedings of ICSE 2006, Available at <http://www.irisa.fr/lande/lande/icse-proceedings/icse/p371.pdf>,(May 20, 2006), pp. 371-380.
“Cholesteric Liquid Crystal”, Retrieved from: <http://en.wikipedia.org/wiki/Cholesteric_liquid_crystal> on Aug. 6, 2012,(Jun. 10, 2012), 2 pages.
“Cirago Slim Case®—Protective case with built-in kickstand for your iPhone 5®”, Retrieved from <http://cirago.com/wordpress/wp-content/uploads/2012/10/ipc1500brochure1.pdf> on Jan. 29, 2013, (Jan. 2013),1 page.
“Corrected Notice of Allowance”, U.S. Appl. No. 13/470,633, (Apr. 9, 2013), 2 pages.
“DR2PA”, retrieved from <http://www.architainment.co.uk/wp-content/uploads/2012/08/DR2PA-AU-US-size-Data-Sheet-Rev-H_LOGO.pdf> on Sep. 17, 2012, 4 pages.
“Final Office Action”, U.S. Appl. No. 13/651,195, (dated Apr. 18, 2013),13 pages.
“Final Office Action”, U.S. Appl. No. 13/651,232, (dated May 21, 2013), 21 pages.
“Final Office Action”, U.S. Appl. No. 13/651,287, (dated May 3, 2013), 16 pages.
“How to Use the iPad's Onscreen Keyboard”, Retrieved from <http://www.dummies.com/how-to/content/how-to-use-the-ipads-onscreen-keyboard.html> on Aug. 28, 2012, 3 pages.
“i-Interactor electronic pen”, Retrieved from: <http://www.alibaba.com/product-gs/331004878/i_Interactor_electronic_pen.html> on Jun. 19, 2012, 5 pages.
“MPC Fly Music Production Controller”, AKAI Professional, Retrieved from: <http://www.akaiprompc.com/mpc-fly> Jul. 9, 2012, 4 pages.
“NI Releases New Maschine & Maschine Mikro”, Retrieved from <http://www.djbooth.net/index/dj-equipment/entry/ni-releases-new-maschine-mikro/> on Sep. 17, 2012, 19 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/471,001, (dated Feb. 19, 2013), 15 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/471,139, (dated Mar. 21, 2013), 12 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/471,202, (dated Feb. 11, 2013), 10 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/563,435, (dated Jun. 14, 2013), 6 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/651,195, (dated Jan. 2, 2013), 14 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/651,232, (dated Jan. 17, 2013), 15 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/651,272, (dated Feb. 12, 2013), 10 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/651,287, (dated Jan. 29, 2013), 13 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/651,304, (dated Mar. 22, 2013), 9 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/651,327, (dated Mar. 22, 2013), 6 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/651,726, (dated Apr. 15, 2013), 6 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/651,871, (dated Mar. 18, 2013), 14 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/651,976, (dated Feb. 22, 2013), 16 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/653,321, (dated Feb. 1, 2013), 13 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/653,682, (dated Feb. 7, 2013), 11 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/653,682, (dated Jun. 3, 2013), 14 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/656,055, (dated Apr. 23, 2013), 11 pages.
“Notice of Allowance”, U.S. Appl. No. 13/470,633, (dated Mar. 22, 2013), 7 pages.
“Notice of Allowance”, U.S. Appl. No. 13/471,202, (dated May 28, 2013), 7 pages.
“Notice of Allowance”, U.S. Appl. No. 13/651,272, (dated May 2, 2013), 7 pages.
“Notice of Allowance”, U.S. Appl. No. 13/651,327, (dated Jun. 11, 2013), 7 pages.
“Notice of Allowance”, U.S. Appl. No. 13/651,726, (dated May 31, 2013), 5 pages.
“On-Screen Keyboard for Windows 7, Vista, XP with Touchscreen”, Retrieved from <www.comfort-software.com/on-screen-keyboard.html> on Aug. 28, 2012, (Feb. 2, 2011), 3 pages.
“Reflex LCD Writing Tablets”, retrieved from <http://www.kentdisplays.com/products/lcdwritingtablets.html> on Jun. 27, 2012, 3 pages.
“Restriction Requirement”, U.S. Appl. No. 13/471,139, (dated Jan. 17, 2013), 7 pages.
“Restriction Requirement”, U.S. Appl. No. 13/651,304, (dated Jan. 18, 2013), 7 pages.
“Restriction Requirement”, U.S. Appl. No. 13/651,726, (dated Feb. 22, 2013), 6 pages.
“Restriction Requirement”, U.S. Appl. No. 13/651,871, (dated Feb. 7, 2013), 6 pages.
“SMART Board™ Interactive Display Frame Pencil Pack”, Available at <http://downloads01.smarttech.com/media/sitecore/en/support/product/sbfpd/400series(interactivedisplayframes)/guides/smartboardinteractivedisplayframepencilpackv12mar09.pdf>,(2009), 2 pages.
“The Microsoft Surface Tablets Comes With Impressive Design and Specs”, Retrieved from <http://microsofttabletreview.com/the-microsoft-surface-tablets-comes-with-impressive-design-and-specs> on Jan. 30, 2013, (Jun. 2012), 2 pages.
“Tilt Shift Lenses: Perspective Control”, retrieved from http://www.cambridgeincolour.com/tutorials/tilt-shift-lenses1.htm, (Mar. 28, 2008),11 Pages.
“What is Active Alignment?”, http://www.kasalis.com/active_alignment.html, retrieved on Nov. 22, 2012, 2 Pages.
Das, Apurba et al., “Study of Heat Transfer through Multilayer Clothing Assemblies: A Theoretical Prediction”, Retrieved from <http://www.autexrj.com/cms/zalaczone_pliki/5_013_11.pdf>, (Jun. 2011), 7 pages.
Iwase, Eiji “Multistep Sequential Batch Assembly of Three-Dimensional Ferromagnetic Microstructures with Elastic Hinges”, Retrieved at <<http://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=1549861>> Proceedings: Journal of Microelectromechanical Systems, (Dec. 2005), 7 pages.
Piltch, Avram “ASUS Eee Pad Slider SL101 Review”, Retrieved from <http://www.laptopmag.com/review/tablets/asus-eee-pad-slider-sl101.aspx>, (Sep. 22, 2011), 5 pages.
Qin, Yongqiang et al., “pPen: Enabling Authenticated Pen and Touch Interaction on Tabletop Surfaces”, In Proceedings of ITS 2010, Available at <http://www.dfki.de/its2010/papers/pdf/po172.pdf>,(Nov. 2010), pp. 283-284.
Sumimoto, Mark “Touch & Write: Surface Computing With Touch and Pen Input”, Retrieved from: <http://www.gottabemobile.com/2009/08/07/touch-write-surface-computing-with-touch-and-pen-input/> on Jun. 19, 2012,(Aug. 7, 2009), 4 pages.
Valliath, G T., “Design of Hologram for Brightness Enhancement in Color LCDs”, Retrieved from <http://www.loreti.it/Download/PDF/LCD/44_05.pdf> on Sep. 17, 2012, 5 pages.
Williams, Jim “A Fourth Generation of LCD Backlight Technology”, Retrieved from <http://cds.linear.com/docs/Application%20Note/an65f.pdf>, (Nov. 1995),124 pages
“Corrected Notice of Allowance”, U.S. Appl. No. 13/470,633, (dated Jul. 2, 2013), 2 pages.
“Corrected Notice of Allowance”, U.S. Appl. No. 13/651,327, (dated Sep. 12, 2013), 2 pages.
“Final Office Action”, U.S. Appl. No. 13/471,001, (dated Jul. 25, 2013), 20 pages.
“Final Office Action”, U.S. Appl. No. 13/471,139, (dated Sep. 16, 2013), 13 pages.
“Final Office Action”, U.S. Appl. No. 13/651,976, (dated Jul. 25, 2013), 21 pages.
“Final Office Action”, U.S. Appl. No. 13/653,321, (dated Aug. 2, 2013), 17 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/527,263, (dated Jul. 19, 2013), 5 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/564,520, (dated Jun. 19, 2013), 8 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/565,124, (dated Jun. 17, 2013), 5 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/651,871, (dated Jul. 1, 2013), 5 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/938,930, (dated Aug. 29, 2013), 9 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/939,002, (dated Aug. 28, 2013), 6 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/939,032, (dated Aug. 29, 2013), 7 pages.
“Notice of Allowance”, U.S. Appl. No. 13/651,195, (dated Jul. 8, 2013), 9 pages.
“Notice of Allowance”, U.S. Appl. No. 13/651,304, (dated Jul. 1, 2013), 5 pages.
“PCT Search Report and Written Opinion”, Application No. PCT/US2013/029461, (dated Jun. 21, 2013), 11 pages.
“PCT Search Report and Written Opinion”, Application No. PCT/US2013/028948, (dated Jun. 21, 2013), 11 pages.
“Advisory Action”, U.S. Appl. No. 13/939,032, dated Feb. 24, 2014, 2 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/599,635, dated Feb. 25, 2014, 13 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/471,186, dated Feb. 27, 2014, 8 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/471,405, dated Feb. 20, 2014, 37 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/564,520, dated Feb. 14, 2014, 5 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/653,682, dated Feb. 26, 2014, 10 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/656,055, dated Mar. 12, 2014, 17 pages.
“Notice of Allowance”, U.S. Appl. No. 13/471,139, dated Mar. 17, 2014, 4 pages.
“Notice of Allowance”, U.S. Appl. No. 13/938,930, dated Feb. 20, 2014, 4 pages.
“Notice of Allowance”, U.S. Appl. No. 13/939,002, dated Mar. 3, 2014, 4 pages.
“Non-Final Office Action”, U.S. Appl. No. 14/063,912, dated Jan. 2, 2014, 10 pages.
“Corrected Notice of Allowance”, U.S. Appl. No. 13/563,435, dated Jan. 14, 2014, 2 pages.
“Corrected Notice of Allowance”, U.S. Appl. No. 13/563,435, dated Jan. 22, 2014, 2 pages.
“Foreign Office Action”, CN Application No. 201320097066.8, dated Oct. 24, 2013, 5 Pages.
“Final Office Action”, U.S. Appl. No. 13/564,520, dated Jan. 15, 2014, 7 pages.
“Advanced Configuration and Power Management Specification”, Intel Corporation, Microsoft Corporation, Toshiba Corp. Revision 1, (Dec. 22, 1996), 364 pages.
“Corrected Notice of Allowance”, U.S. Appl. No. 13/651,327, (dated Sep. 23, 2013), 2 pages.
“Corrected Notice of Allowance”, U.S. Appl. No. 13/651,726, (dated Sep. 17, 2013), 2 pages.
“Developing Next-Generation Human Interfaces using Capacitive and Infrared Proximity Sensing”, Silicon Laboratories, Inc., Available at <http://www.silabs.com/pages/DownloadDoc.aspx?FILEURL=support%20documents/technicaldocs/capacitive%20and%20proximity%20sensing_wp.pdf&src=SearchResults>,(Aug. 30, 2010), pp. 1-10.
“Directional Backlighting for Display Panels”, U.S. Appl. No. 13/021,448, (Feb. 4, 2011),38 pages.
“Final Office Action”, U.S. Appl. No. 13/653,682, (dated Oct. 18, 2013), 16 pages.
“Final Office Action”, U.S. Appl. No. 13/656,055, (dated Oct. 23, 2013), 14 pages.
“Final Office Action”, U.S. Appl. No. 13/938,930, (dated Nov. 8, 2013), 10 pages.
“Final Office Action”, U.S. Appl. No. 13/939,002, (dated Nov. 8, 2013), 7 pages.
“Final Office Action”, U.S. Appl. No. 13/939,032, (dated Dec. 20, 2013), 5 pages.
“FingerWorks Installation and Operation Guide for the TouchStream ST and TouchStream LP”, FingerWorks, Inc. Retrieved from <http://ec1.images-amazon.com/media/i3d/01/A/man-migrate/MANUAL000049862.pdf>, (2002), 14 pages.
“International Search Report and Written Opinion”, Application No. PCT/US2013/040968, (dated Sep. 5, 2013), 12 pages.
“International Search Report and Written Opinion”, Application No. PCT/US2013/045049, (dated Sep. 16, 2013), 9 pages.
“International Search Report and Written Opinion”, Application No. PCT/US2013/042550, (dated Sep. 24, 2013), 14 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/021,448, (dated Dec. 13, 2012), 9 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/468,918, (dated Dec. 26, 2013), 18 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/651,232, (dated Dec. 5, 2013), 15 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/656,520, (dated Feb. 1, 2013), 15 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/656,520, (dated Jun. 5, 2013), 8 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/780,228, (dated Oct. 30, 2013), 12 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/939,002, (dated Dec. 20, 2013), 5 pages.
“Notice of Allowance”, U.S. Appl. No. 13/563,435, (dated Nov. 12, 2013), 5 pages.
“Notice of Allowance”, U.S. Appl. No. 13/565,124, (dated Dec. 24, 2013), 6 pages.
“Notice of Allowance”, U.S. Appl. No. 13/651,871, (dated Oct. 2, 2013), 7 pages.
“Notice of Allowance”, U.S. Appl. No. 13/653,321, (dated Dec. 18, 2013), 41 pages.
“Notice to Grant”, CN Application No. 201320097089.9, (dated Sep. 29, 2013), 2 Pages.
“Notice to Grant”, CN Application No. 201320097124.7, (dated Oct. 8, 2013),2 pages.
“Optical Sensors in Smart Mobile Devices”, ON Semiconductor, TND415/D, Available at <http://www.onsemi.jp/pub_link/Collateral/TND415-D.PDF>,(Nov. 2010), pp. 1-13.
“Optics for Displays: Waveguide-based Wedge Creates Collimated Display Backlight”, OptoIQ, retrieved from <http://www.optoiq.com/index/photonics-technologies-applications/lfw-display/lfw-article-display.articles.laser-focus-world.volume-46.issue-1.world-news.optics-for_displays.html> on Nov. 2, 2010,(Jan. 1, 2010), 3 pages.
“Restriction Requirement”, U.S. Appl. No. 13/468,918, (dated Nov. 29, 2013), 6 pages.
“Welcome to Windows 7”, Retrieved from: <http://www.microsoft.com/en-us/download/confirmation.aspx?id=4984> on Aug. 1, 2013, (Sep. 16, 2009), 3 pages.
Gaver, William W., et al., “A Virtual Window on Media Space”, retrieved from <http://www.gold.ac.uk/media/15gaver-smets-overbeeke.MediaSpaceWindow.chi95.pdf> on Jun. 1 , 2012, retrieved from <http://www.gold.ac.uk/media/15gaver-smets-overbeeke.MediaSpaceWindow.chi95.pdf> on Jun. 1, 2012,(May 7, 1995), 9 pages.
Harada, Susumu et al., “VoiceDraw: A Hands-Free Voice-Driven Drawing Application for People With Motor Impairment”, In Proceedings of Ninth International ACM SIGACCESS Conference on Computers and Accessibility, retrieved from <http://citeerx.ist.psu.edu/viewdoc/download?doi=10.1.1.113.7211&rep=rep1&type=pdf> on Jun. 1, 2012,(Oct. 15, 2007), 8 pages.
Hinckley, Ken et al., “Codex: A Dual Screen Tablet Computer”, Conference on Human Factors in Computing Systems, (Apr. 9, 2009), 10 pages.
Kaufmann, Benoit et al., “Hand Posture Recognition Using Real-time Artificial Evolution”, EvoApplications'09, retrieved from <http://evelyne.lutton.free.fr/Papers/KaufmannEvolASP2010.pdf> on Jan. 5, 2012,(Apr. 3, 2010), 10 pages.
Li, et al., “Characteristic Mode Based Tradeoff Analysis of Antenna-Chassis Interactions for Multiple Antenna Terminals”, In IEEE Transactions on Antennas and Propagation, Retrieved from <http://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=6060882>,(Feb. 2012),13 pages.
Manresa-Yee, Cristina et al., “Experiences Using a Hands-Free Interface”, In Proceedings of the 10th International ACM SIGACCESS Conference on Computers and Accessibility, retrieved from <http://dmi.uib.es/˜cmanresay/Research/%5BMan08%5DAssets08.pdf> on Jun. 1, 2012,(Oct. 13, 2008),pp. 261-262.
Nakanishi, Hideyuki et al., “Movable Cameras Enhance Social Telepresence in Media Spaces”, In Proceedings of the 27th International Conference on Human Factors In Computing Systems, retrieved from <http://smg.ams.eng.osaka-u.ac.jp/˜nakanishi/hnp_2009_chi.pdf> on Jun. 1, 2012,(Apr. 6, 2009),10 pages.
Prospero, Michael “Samsung Outs Series 5 Hybrid PC Tablet”, Retrieved from: <http://blog.laptopmag.com/samsung-outs-series-5-hybrid-pc-tablet-running-windows-8> on Oct. 31, 2013, (Jun. 4, 2012),7 pages.
Reilink, Rob et al., “Endoscopic Camera Control by Head Movements for Thoracic Surgery”, In Proceedings of 3rd IEEE RAS & EMBS International Conference of Biomedical Robotics and Biomechatronics, retrieved from <http://doc.utwente.nl/74929/1/biorob_online.pdf> on Jun. 1, 2012,(Sep. 26, 2010),pp. 510-515.
Sundstedt, Veronica “Gazing at Games: Using Eye Tracking to Control Virtual Characters”, In ACM SIGGRAPH 2010 Courses, retrieved from <http://www.tobii.com/Global/Analysis/Training/EyeTrackAwards/veronica_sundstedt.pdf> on Jun. 1, 2012,(Jul. 28, 2010),85 pages.
Travis, Adrian et al., “Collimated Light from a Waveguide for a Display Backlight”, Optics Express, 19714, vol. 17, No. 22, retrieved from <http://download.microsoft.com/download/D/2/E/D2E425F8-CF3C-4C71-A4A2-70F9D4081007/OpticsExpressbacklightpaper.pdf> on Oct. 15, 2009,(Oct. 15, 2009), 6 pages.
Travis, Adrian et al., “The Design of Backlights for View-Sequential 3D”, retrieved from <http://download.microsoft.com/download/D/2/E/D2E425F8-CF3C-4C71-A4A2-70F9D4081007/Backlightforviewsequentialautostereo.docx> on Nov. 1, 2010, 4 pages.
Valli, Alessandro “Notes on Natural Interaction”, retrieved from <http://www.idemployee.id.tue.nl/g.w.m.rauterberg/lecturenotes/valli-2004.pdf> on Jan. 5, 2012,(Sep. 2005), 80 pages.
Vaucelle, Cati “Scopemate, a Robotic Microscope!”, Architectradure, retrieved from <http://architectradure.blogspot.com/2011/10/at-uist-this-monday-scopemate-robotic.html> on Jun. 6, 2012,(Oct. 17, 2011), 2 pages.
Xu, Zhang et al., “Hand Gesture Recognition and Virtual Game Control Based on 3D Accelerometer and EMG Sensors”, IUI'09, Feb. 8-11, 2009, retrieved from <http://sclab.yonsei.ac.kr/courses/10TPR/10TPR.files/Hand%20Gesture%20Recognition%20and%20Virtual%20Game%20Control%20based%20on%203d%20accelerometer%20and%20EMG%20sensors.pdf> on Jan. 5, 2012,(Feb. 8, 2009), 5 pages.
Xu, Zhi-Gang et al., “Vision-based Detection of Dynamic Gesture”, ICTM'09, Dec. 5-6, 2009, retrieved from <http://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=5412956> on Jan. 5, 2012,(Dec. 5, 2009), pp. 223-226.
Zhu, Dingyun et al., “Keyboard before Head Tracking Depresses User Success in Remote Camera Control”, In Proceedings of 12th IFIP TC 13 International Conference on Human-Computer Interaction, Part II, retrieved from <http://csiro.academia.edu/Departments/CSIRO_ICT_Centre/Papers?page=5> on Jun. 1, 2012,(Aug. 24, 2009), 14 pages.
“Corrected Notice of Allowance”, U.S. Appl. No. 13/651,232, dated Jul. 31, 2014, 2 pages.
“Corrected Notice of Allowance”, U.S. Appl. No. 13/651,287, dated Aug. 21, 2014, 2 pages.
“Corrected Notice of Allowance”, U.S. Appl. No. 14/199,924, dated Aug. 29, 2014, 2 pages.
“Corrected Notice of Allowance”, U.S. Appl. No. 14/199,924, dated Sep. 5, 2014, 2 pages.
“Final Office Action”, U.S. Appl. No. 13/471,376, dated Aug. 18, 2014, 24 pages.
“Final Office Action”, U.S. Appl. No. 13/595,700, dated Aug. 15, 2014, 6 pages.
“Final Office Action”, U.S. Appl. No. 13/599,635, dated Aug. 8, 2014, 16 pages.
“Foreign Notice of Allowance”, CN Application No. 201320097065.3, dated Nov. 21, 2013, 2 pages.
“Foreign Office Action”, CN Application No. 201320097065.3, dated Jun. 18, 2013, 2 pages.
“International Search Report and Written Opinion”, Application No. PCT/US2013/028485, dated Jun. 25, 2014, 10 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/471,282, dated Sep. 3, 2014, 13 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/712,111, dated Aug. 1, 2014, 11 pages.
“Non-Final Office Action”, U.S. Appl. No. 14/063,912, dated Sep. 2, 2014, 11 pages.
“Notice of Allowance”, U.S. Appl. No. 13/471,030, dated Sep. 5, 2014, 6 pages.
“Restriction Requirement”, U.S. Appl. No. 13/653,184, dated Sep. 5, 2014, 6 pages.
“Supplemental Notice of Allowance”, U.S. Appl. No. 13/471,405, dated Aug. 29, 2014, 5 pages.
“Advisory Action”, U.S. Appl. No. 14/199,924, dated May 28, 2014, 2 pages.
“Corrected Notice of Allowance”, U.S. Appl. No. 13/563,435, dated Mar. 20, 2014, 2 pages.
“Corrected Notice of Allowance”, U.S. Appl. No. 13/565,124, dated Apr. 3, 2014, 4 pages.
“Corrected Notice of Allowance”, U.S. Appl. No. 13/565,124, dated Mar. 10, 2014, 2 pages.
“Corrected Notice of Allowance”, U.S. Appl. No. 13/565,124, dated Apr. 14, 2014, 2 pages.
“Corrected Notice of Allowance”, U.S. Appl. No. 13/938,930, dated May 6, 2014, 2 pages.
“Corrected Notice of Allowance”, U.S. Appl. No. 13/938,930, dated Jun. 6, 2014, 2 pages.
“Corrected Notice of Allowance”, U.S. Appl. No. 13/939,002, dated May 22, 2014, 2 pages.
“Corrected Notice of Allowance”, U.S. Appl. No. 13/939,002, dated Jun. 19, 2014, 2 pages.
“Corrected Notice of Allowance”, U.S. Appl. No. 13/939,002, dated May 5, 2014, 2 pages.
“Corrected Notice of Allowance”, U.S. Appl. No. 13/939,032, dated Jun. 26, 2014, 2 pages.
“Corrected Notice of Allowance”, U.S. Appl. No. 13/939,032, dated Jul. 15, 2014, 2 pages.
“Final Office Action”, U.S. Appl. No. 13/653,682, dated Jun. 11, 2014, 11 pages.
“Final Office Action”, U.S. Appl. No. 14/063,912, dated Apr. 29, 2014, 10 pages.
“Final Office Action”, U.S. Appl. No. 14/199,924, dated May 6, 2014, 5 pages.
“Foreign Notice of Allowance”, CN Application No. 201320096755.7, dated Jan. 27, 2014, 2 pages.
“Foreign Office Action”, CN Application No. 201320097079.5, dated Sep. 26, 2013, 4 pages.
“Foreign Office Action”, CN Application No. 201320328022.1, dated Feb. 17, 2014, 4 Pages.
“Foreign Office Action”, CN Application No. 201320328022.1, dated Oct. 18, 2013, 3 Pages.
“Interlink Electronics FSR (TM) Force Sensing Resistors (TM)”, Retrieved at <<http://akizukidenshi.com/download/ds/ interlinkelec/94-00004+Rev+B%20FSR%201ntegration%20Guide.pdf on Mar. 21, 2013, 36 pages.
“International Search Report and Written Opinion”, Application No. PCT/US2014/031531, dated Jun. 20, 2014, 10 Pages.
“International Search Report and Written Opinion”, Application No. PCT/US2013/028483, dated Jun. 24, 2014, 10 pages.
“International Search Report and Written Opinion”, Application No. PCT/US2013/028484, dated Jun. 24, 2014, 10 pages.
“International Search Report and Written Opinion”, Application No. PCT/US2013/028769, dated Jun. 26, 2014, 10 pages.
“International Search Report and Written Opinion”, Application No. PCT/US2013/028771, dated Jun. 19, 2014, 10 pages.
“International Search Report and Written Opinion”, Application No. PCT/US2013/028486, dated Jun. 20, 2014, 10 pages.
“International Search Report and Written Opinion”, Application No. PCT/US2013/041017, dated Jul. 17, 2014, 10 pages.
“International Search Report and Written Opinion”, Application No. PCT/US2013/028489, dated Jun. 20, 2014, 11 pages.
“International Search Report and Written Opinion”, Application No. PCT/US2013/028488, dated Jun. 24, 2014, 11 pages.
“International Search Report and Written Opinion”, Application No. PCT/US2013/028767, dated Jun. 24, 2014, 11 pages.
“International Search Report and Written Opinion”, Application No. PCT/US2013/028481, dated Jun. 19, 2014, 11 pages.
“International Search Report and Written Opinion”, Application No. PCT/US2013/028490, dated Jun. 24, 2014, 11 pages.
“International Search Report and Written Opinion”, Application No. PCT/US2013/028766, dated Jun. 26, 2014, 11 pages.
“International Search Report and Written Opinion”, Application No. PCT/US2013/028772, dated Jun. 30, 2014, 11 pages.
“International Search Report and Written Opinion”, Application No. PCT/US2013/028768, dated Jun. 24, 2014, 12 pages.
“International Search Report and Written Opinion”, Application No. PCT/US2013/028482, dated Jun. 20, 2014, 13 pages.
“International Search Report and Written Opinion”, Application No. PCT/US2013/028487, dated May 27, 2014, 9 pages.
“International Search Report and Written Opinion”, Application No. PCT/US2013/028770, dated Jun. 26, 2014, 9 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/468,882, dated Jul. 9, 2014, 9 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/468,949, dated Jun. 20, 2014, 10 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/470,951, dated Jul. 2, 2014, 19 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/471,001, dated Jun. 17, 2014, 23 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/471,030, dated May 15, 2014, 10 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/471,054, dated Jun. 3, 2014, 15 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/471,237, dated Mar. 24, 2014, 7 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/471,376, dated Apr. 2, 2014, 17 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/471,412, dated Jul. 11, 2014, 22 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/492,232, dated Apr. 30, 2014, 9 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/527,263, dated Apr. 3, 2014, 6 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/564,520, dated Jun. 16, 2014, 5 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/595,700, dated Jun. 18, 2014, 8 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/647,479, dated Jul. 3, 2014, 10 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/651,976, dated Jun. 16, 2014, 23 pages.
“Non-Final Office Action”, U.S. Appl. No. 14/199,924, dated Apr. 10, 2014, 6 pages.
“Non-Final Office Action”, U.S. Appl. No. 14/200,595, dated Apr. 11, 2014, 4 pages.
“Non-Final Office Action”, U.S. Appl. No. 14/225,250, dated Jun. 17, 2014, 5 pages.
“Non-Final Office Action”, U.S. Appl. No. 14/225,276, dated Jun. 13, 2014, 6 pages.
“Non-Final Office Action”, U.S. Appl. No. 14/277,240, dated Jun. 13, 2014, 6 pages.
“Notice of Allowance”, U.S. Appl. No. 13/468,918, dated Jun. 17, 2014, 5 pages.
“Notice of Allowance”, U.S. Appl. No. 13/471,186, dated Jul. 3, 2014, 7 pages.
“Notice of Allowance”, U.S. Appl. No. 13/471,237, dated May 12, 2014, 8 pages.
“Notice of Allowance”, U.S. Appl. No. 13/471,405, dated Jun. 24, 2014, 9 pages.
“Notice of Allowance”, U.S. Appl. No. 13/651,232, dated Apr. 25, 2014, 9 pages.
“Notice of Allowance”, U.S. Appl. No. 13/651,287, dated May 2, 2014, 6 pages.
“Notice of Allowance”, U.S. Appl. No. 13/939,032, dated Apr. 3, 2014, 4 pages.
“Notice of Allowance”, U.S. Appl. No. 14/018,286, dated May 23, 2014, 8 pages.
“Notice of Allowance”, U.S. Appl. No. 14/199,924, dated Jun. 10, 2014, 4 pages.
“Restriction Requirement”, U.S. Appl. No. 13/595,700, dated May 28, 2014, 6 pages.
“Supplemental Notice of Allowance”, U.S. Appl. No. 13/653,321, dated Mar. 28, 2014, 4 pages.
“Supplemental Notice of Allowance”, U.S. Appl. No. 14/018,286, dated Jun. 11, 2014, 5 pages.
“Corrected Notice of Allowance”, U.S. Appl. No. 13/651,232, Apr. 24, 2015, 2 pages.
“Corrected Notice of Allowance”, U.S. Appl. No. 13/656,055, dated Apr. 13, 2015, 2 pages.
“Final Office Action”, U.S. Appl. No. 13/525,614, dated Apr. 29, 2015, 20 pages.
“Final Office Action”, U.S. Appl. No. 13/780,228, dated Apr. 10, 2015, 19 pages.
“Final Office Action”, U.S. Appl. No. 14/225,250, dated Mar. 13, 2015, 7 pages.
“Foreign Notice on Reexamination”, CN Application No. 201320097066.8, dated Apr. 3, 2015, 7 Pages.
“Non-Final Office Action”, U.S. Appl. No. 13/471,054, dated Mar. 13, 2015, 18 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/471,376, dated Mar. 27, 2015, 28 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/471,393, dated Mar. 26, 2015, 13 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/471,412, dated Jun. 1, 2015, 31 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/647,479, dated Apr. 28, 2015, 11 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/653,218, dated Mar. 4, 2015, 16 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/852,848, dated Mar. 26, 2015, 7 pages.
“Non-Final Office Action”, U.S. Appl. No. 14/059,280, dated Mar. 3, 2015, 18 pages.
“Non-Final Office Action”, U.S. Appl. No. 14/063,912, dated May 7, 2015, 12 pages.
“Non-Final Office Action”, U.S. Appl. No. 14/225,276, dated Apr. 23, 2015, 7 pages.
“Notice of Allowance”, U.S. Appl. No. 13/468,949, dated Apr. 24, 2015, 9 pages.
“Notice of Allowance”, U.S. Appl. No. 13/468,918, dated Apr. 8, 2015, 7 pages.
“Notice of Allowance”, U.S. Appl. No. 13/468,949, dated Apr. 24, 2015, 8 pages.
“Notice of Allowance”, U.S. Appl. No. 13/471,030, dated Apr. 6, 2015, 6 pages.
“Notice of Allowance”, U.S. Appl. No. 13/471,282, dated Apr. 30, 2015, 8 pages.
“Notice of Allowance”, U.S. Appl. No. 13/564,520, dated May 8, 2015, 4 pages.
“Notice of Allowance”, U.S. Appl. No. 13/651,232, dated Mar. 30, 2015, 7 pages.
“Notice of Allowance”, U.S. Appl. No. 13/653,184, dated Mar. 10, 2015, 6 pages.
“Notice of Allowance”, U.S. Appl. No. 13/656,055, dated Mar. 4, 2015, 7 pages.
“Supplemental Notice of Allowance”, U.S. Appl. No. 13/595,700, dated Apr. 10, 2015, 2 pages.
“Supplemental Notice of Allowance”, U.S. Appl. No. 13/595,700, dated May 4, 2015, 2 pages.
“Supplemental Notice of Allowance”, U.S. Appl. No. 13/595,700, dated May 22, 2015, 2 pages.
“Supplemental Notice of Allowance”, U.S. Appl. No. 13/656,055, dated May 15, 2015, 2 pages.
Schafer,“Using Interactive Maps for Navigation and Collaboration”, CHI '01 Extended Abstracts on Human Factors in Computing Systems, Mar. 31, 2001, 2 pages.
“Corrected Notice of Allowance”, U.S. Appl. No. 13/471,030, dated Sep. 30, 2014, 2 pages.
“Corrected Notice of Allowance”, U.S. Appl. No. 14/199,924, dated Sep. 19, 2014, 2 pages.
“Final Office Action”, U.S. Appl. No. 13/468,949, dated Oct. 6, 2014, 11 pages.
“Final Office Action”, U.S. Appl. No. 13/471,054, dated Oct. 23, 2014, 17 pages.
“Final Office Action”, U.S. Appl. No. 13/492,232, dated Nov. 17, 2014, 13 pages.
“Final Office Action”, U.S. Appl. No. 13/595,700, dated Oct. 9, 2014, 8 pages.
“Final Office Action”, U.S. Appl. No. 13/656,055, dated Sep. 17, 2014, 10 pages.
“Final Office Action”, U.S. Appl. No. 14/200,595, dated Nov. 19, 2014, 5 pages.
“International Search Report and Written Opinion”, Application No. PCT/US2014/043546, dated Oct. 9, 2014, 10 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/471,393, dated Oct. 20, 2014, 12 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/525,614, dated Nov. 24, 2014, 19 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/653,184, dated Dec. 1, 2014, 7 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/780,228, dated Sep. 15, 2014, 18 pages.
“Notice of Allowance”, U.S. Appl. No. 13/653,682, dated Sep. 24, 2014, 4 pages.
“Notice of Allowance”, U.S. Appl. No. 14/277,240, dated Sep. 16, 2014, 4 pages.
“Restriction Requirement”, U.S. Appl. No. 13/653,218, dated Nov. 7, 2014, 6 pages.
“Restriction Requirement”, U.S. Appl. No. 14/147,252, dated Dec. 1, 2014, 6 pages.
Harrison, “UIST 2009 Student Innovation Contest-Demo Video”, Retrieved From: <https://www.youtube.com/watch?v=PDI8eYIASf0> Sep. 16, 2014, Jul. 23, 2009, 1 pages.
“Final Office Action”, U.S. Appl. No. 13/468,882, dated Feb. 12, 2015, 9 pages.
“First Examination Report”, NZ Application No. 628690, dated Nov. 27, 2014, 2 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/492,232, dated Feb. 24, 2015, 12 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/599,635, dated Feb. 12, 2015, 16 pages.
“Non-Final Office Action”, U.S. Appl. No. 14/147,252, dated Feb. 23, 2015, 11 pages.
“Notice of Allowance”, U.S. Appl. No. 14/200,595, dated Feb. 17, 2015, 2 pages.
“Notice of Allowance”, U.S. Appl. No. 14/200,595, dated Feb. 25, 2015, 4 pages.
“Final Office Action”, U.S. Appl. No. 13/471,412, dated Dec. 15, 2014, 11 pages.
“Final Office Action”, U.S. Appl. No. 13/647,479, dated Dec. 12, 2014, 12 pages.
“Final Office Action”, U.S. Appl. No. 13/712,111, dated Dec. 18, 2014, 13 pages.
“Final Office Action”, U.S. Appl. No. 14/225,276, dated Dec. 17, 2014, 6 pages.
“Supplemental Notice of Allowance”, U.S. Appl. No. 13/471,405, dated Dec. 17, 2014, 5 pages.
“Corrected Notice of Allowance”, U.S. Appl. No. 14/277,240, dated Jan. 8, 2015, 2 pages.
“Final Office Action”, U.S. Appl. No. 13/470,951, dated Jan. 12, 2015, 20 pages.
“Final Office Action”, U.S. Appl. No. 13/527,263, dated Jan. 27, 2015, 7 pages.
“Final Office Action”, U.S. Appl. No. 14/063,912, dated Jan. 12, 2015, 12 pages.
“Foreign Office Action”, CN Application No. 201320097079.5, dated Jul. 28, 2014, 4 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/471,030, dated Jan. 15, 2015, 7 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/564,520, dated Jan. 26, 2015, 6 pages.
“Notice of Allowance”, U.S. Appl. No. 13/595,700, dated Jan. 21, 2015, 4 pages.
“Notice of Allowance”, U.S. Appl. No. 13/651,976, dated Jan. 21, 2015, 10 pages.
“Advisory Action”, U.S. Appl. No. 13/471,376, dated Sep. 23, 2015, 7 pages.
“Corrected Notice of Allowance”, U.S. Appl. No. 13/471,030, dated Aug. 10, 2015, 2 pages.
“Corrected Notice of Allowance”, U.S. Appl. No. 13/564,520, dated Aug. 14, 2015, 2 pages.
“Corrected Notice of Allowance”, U.S. Appl. No. 13/564,520, dated Sep. 17, 2015, 2 pages.
“Corrected Notice of Allowance”, U.S. Appl. No. 13/651,232, dated Jun. 10, 2015, 2 pages.
“Corrected Notice of Allowance”, U.S. Appl. No. 13/651,232, dated Jul. 6, 2015, 2 pages.
“Corrected Notice of Allowance”, U.S. Appl. No. 13/656,055, dated Jul. 1, 2015, 2 pages.
“Corrected Notice of Allowance”, U.S. Appl. No. 14/200,595, dated Jun. 4, 2015, 3 pages.
“Corrected Notice of Allowance”, U.S. Appl. No. 14/225,276, dated Aug. 27, 2015, 2 pages.
“Corrected Notice of Allowance”, U.S. Appl. No. 14/457,881, dated Aug. 20, 2015, 2 pages.
“Extended European Search Report”, EP Application No. 13859280.3, dated Sep. 7, 2015, 6 pages.
“Extended European Search Report”, EP Application No. 13859406.4, dated Sep. 8, 2015, 6 pages.
“Final Office Action”, U.S. Appl. No. 13/471,376, dated Jul. 28, 2015, 35 pages.
“Final Office Action”, U.S. Appl. No. 13/492,232, dated Jul. 10, 2015, 11 pages.
“Final Office Action”, U.S. Appl. No. 13/599,635, dated Jul. 30, 2015, 23 pages.
“Final Office Action”, U.S. Appl. No. 13/647,479, dated Sep. 17, 2015, 11 pages.
“Final Office Action”, U.S. Appl. No. 13/852,848, dated Jul. 20, 2015, 9 pages.
“Final Office Action”, U.S. Appl. No. 14/059,280, dated Jul. 22, 2015, 25 pages.
“Final Office Action”, U.S. Appl. No. 14/063,912, dated Sep. 3, 2015, 13 pages.
“Final Office Action”, U.S. Appl. No. 14/147,252, dated Jun. 25, 2015, 11 pages.
“Foreign Office Action”, CN Application No. 201310067335.0, dated Jun. 12, 2015, 15 Pages.
“Foreign Office Action”, CN Application No. 201310067385.9, dated Aug. 6, 2015, 16 pages.
“Foreign Office Action”, CN Application No. 201310067808.7, dated May 28, 2015, 14 Pages.
“Foreign Office Action”, CN Application No. 201310225788.1, dated Jun. 23, 2015, 14 Pages.
“International Preliminary Report on Patentability”, Application No. PCT/US2014/031531, dated Jun. 9, 2015, 7 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/525,614, dated Jul. 31, 2015, 20 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/712,111, dated Jul. 17, 2015, 14 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/780,228, dated Sep. 18, 2015, 86 pages.
“Non-Final Office Action”, U.S. Appl. No. 14/225,250, dated Aug. 19, 2015, 9 pages.
“Non-Final Office Action”, U.S. Appl. No. 14/225,276, dated Aug. 19, 2015, 9 pages.
“Non-Final Office Action”, U.S. Appl. No. 14/727,001, dated Jul. 10, 2015, 7 pages.
“Notice of Allowance”, U.S. Appl. No. 14/225,276, dated Jun. 22, 2015, 4 pages.
“Notice of Allowance”, U.S. Appl. No. 14/457,881, dated Jul. 22, 2015, 7 pages.
“Restriction Requirement”, U.S. Appl. No. 13/891,109, dated Sep. 22, 2015, 6 pages.
“Supplemental Notice of Allowance”, U.S. Appl. No. 13/468,918, dated Jun. 4, 2015, 2 pages.
“Supplemental Notice of Allowance”, U.S. Appl. No. 13/468,918, dated Aug. 7, 2015, 4 pages.
“Supplemental Notice of Allowance”, U.S. Appl. No. 13/468,949, dated Jun. 5, 2015, 2 pages.
“Supplemental Notice of Allowance”, U.S. Appl. No. 13/468,949, dated Sep. 14, 2015, 2 pages.
“Supplemental Notice of Allowance”, U.S. Appl. No. 13/653,184, dated Jun. 24, 2015, 2 pages.
“Supplemental Notice of Allowance”, U.S. Appl. No. 13/656,055, dated Jun. 10, 2015, 2 pages.
Cunningham,“Software Infrastructure for Natural Language Processing”, In Proceedings of the fifth conference on Applied natural language processing, Mar. 31, 1997, pp. 237-244.
“Corrected Notice of Allowance”, U.S. Appl. No. 13/470,951, dated Jul. 8, 2016, 2 pages.
“Corrected Notice of Allowance”, U.S. Appl. No. 13/891,109, dated Jun. 21, 2016, 2 pages.
“Foreign Notice of Allowance”, CN Application No. 201310067531.8, dated Jun. 7, 2016, 4 pages.
“Foreign Notice of Allowance”, CN Application No. 201310067592.4, dated May 17, 2016, 4 pages.
“Foreign Office Action”, CL Application No. 14-211.785, dated Apr. 26, 2016, 12 pages.
“Foreign Office Action”, CN Application No. 201310067622.1, dated Jun. 21, 2016, 13 Pages.
“Foreign Office Action”, CN Application No. 201310096345.7, dated May 25, 2016, 16 Pages.
“Foreign Office Action”, CN Application No. 201380025290.9, dated May 10, 2016, 15 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/780,228, dated Jun. 17, 2016, 18 pages.
“Non-Final Office Action”, U.S. Appl. No. 14/225,250, dated Jul. 1, 2016, 9 pages.
“Notice of Allowance”, U.S. Appl. No. 14/517,048, dated Jul. 7, 2016, 7 pages.
“Corrected Notice of Allowance”, U.S. Appl. No. 13/527,263, dated Apr. 12, 2016, 2 pages.
“Corrected Notice of Allowance”, U.S. Appl. No. 13/527,263, dated Apr. 25, 2016, 2 pages.
“Corrected Notice of Allowance”, U.S. Appl. No. 14/517,048, dated Apr. 13, 2016, 2 pages.
“Extended European Search Report”, EP Application No. 13858403.2, dated Mar. 16, 2016, 12 pages.
“Extended European Search Report”, EP Application No. 13860271.9, dated May 4, 2016, 8 pages.
“Extended European Search Report”, EP Application No. 13861059.7, dated Apr. 29, 2016, 8 pages.
“Final Office Action”, U.S. Appl. No. 13/492,232, dated May 25, 2016, 12 pages.
“Final Office Action”, U.S. Appl. No. 13/780,228, dated Mar. 23, 2016, 16 pages.
“Foreign Notice of Allowance”, CN Application No. 201310065273.X, dated Mar. 31, 2016, 4 Pages.
“Foreign Notice of Allowance”, CN Application No. 201320097079.5, dated Apr. 1, 2016, 4 Pages.
“Foreign Notice of Allowance”, CN Application No. 201310067808.7, dated May 4, 2016, 4 pages.
“Foreign Office Action”, CN Application No. 201310067385.9, dated Apr. 14, 2016, 14 Pages.
“Foreign Office Action”, CN Application No. 201310067627.4, dated May 3, 2016, 7 pages.
“Foreign Office Action”, CN Application No. 201310225788.1, dated Feb. 29, 2016, 11 pages.
“Foreign Office Action”, CN Application No. 201310316114.2, dated Apr. 18, 2016, 11 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/653,218, dated Apr. 20, 2016, 17 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/689,541, dated Apr. 14, 2016, 23 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/712,111, dated Mar. 24, 2016, 16 pages.
“Non-Final Office Action”, U.S. Appl. No. 14/307,262, dated Apr. 20, 2016, 10 pages.
“Non-Final Office Action”, U.S. Appl. No. 14/794,182, dated Apr. 13, 2016, 15 pages.
“Non-Final Office Action”, U.S. Appl. No. 14/994,737, dated Apr. 5, 2016, 6 pages.
“Notice of Allowance”, U.S. Appl. No. 13/468,882, dated May 24, 2016, 5 pages.
“Notice of Allowance”, U.S. Appl. No. 13/470,951, dated Mar. 24, 2016, 8 pages.
“Notice of Allowance”, U.S. Appl. No. 13/891,109, dated May 6, 2016, 7 pages.
“Advisory Action”, U.S. Appl. No. 14/059,280, dated Sep. 25, 2015, 7 pages.
“Corrected Notice of Allowance”, U.S. Appl. No. 14/225,276, dated Sep. 29, 2015, 2 pages.
“Corrected Notice of Allowance”, U.S. Appl. No. 14/457,881, dated Oct. 02, 2015, 2 pages.
“Decision on Reexamination”, CN Application No. 201320097079.5, dated Sep. 7, 2015, 8 Pages.
“Extended European Search Report”, EP Application No. 13858620.1, dated Sep. 18, 2015, 6 pages.
“Extended European Search Report”, EP Application No. 13858834.8, dated Oct. 29, 2015, 8 pages.
“Final Office Action”, U.S. Appl. No. 13/653,218, dated Oct. 5, 2015, 16 pages.
“Final Office Action”, U.S. Appl. No. 13/689,541, dated Nov. 2, 2015, 21 pages.
“Final Office Action”, Application No. 13/712,111, dated Nov. 4, 2015, 14 pages.
“Foreign Office Action”, CN Application No. 201310065273.X, dated Oct. 28, 2015, 14 pages.
“Foreign Office Action”, CN Application No. 201310067592.4, dated Oct. 23, 2015, 12 Pages.
“Foreign Office Action”, CN Application No. 201310067622.1, dated Oct. 27, 2015, 14 pages.
“Foreign Office Action”, CN Application No. 201310067627.4, dated Sep. 28, 2015, 14 pages.
“Foreign Office Action”, CN Application No. 201310096345.7, dated Oct. 19, 2015, 16 Pages.
“Foreign Office Action”, CN Application No. 201310316114.2, dated Sep. 29, 2015, 13 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/468,882, dated Nov. 13, 2015, 9 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/470,951, dated Oct. 1, 2015, 29 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/471,393, dated Sep. 30, 2015, 15 pages.
“Non-Final Office Action”, U.S. Appl. No. 14/162,529, dated Sep. 18, 2015, 13 pages.
“Notice of Allowance”, U.S. Appl. No. 13/471,054, dated Sep. 25, 2015, 7 pages.
“Notice of Allowance”, U.S. Appl. No. 13/471,376, dated Nov. 23, 2015, 9 pages.
“Notice of Allowance”, U.S. Appl. No. 13/471,412, dated Nov. 20, 2015, 10 pages.
“Notice of Allowance”, U.S. Appl. No. 13/852,848, dated Nov. 19, 2015, 4 pages.
“Notice of Allowance”, U.S. Appl. No. 14/059,280, dated Nov. 23, 2015, 9 pages.
“Notice of Allowance”, U.S. Appl. No. 14/727,001, dated Oct. 2, 2015, 4 pages.
“Supplemental Notice of Allowance”, U.S. Appl. No. 13/471,054, dated Nov. 19, 2015, 2 pages.
“Supplementary European Search Report”, EP Application No. 13728568.0, dated Oct. 30, 2015, 7 pages.
“Corrected Notice of Allowance”, U.S. Appl. No. 13/471,412, dated Mar. 3, 2016, 2 pages.
“Corrected Notice of Allowance”, U.S. Appl. No. 13/527,263, dated Mar. 7, 2016, 2 pages.
“Corrected Notice of Allowance”, U.S. Appl. No. 13/852,848, dated Mar. 2, 2016, 2 pages.
“Extended European Search Report”, EP Application No. 13728568.0, dated Mar. 14, 2016, 16 pages.
“Final Office Action”, U.S. Appl. No. 13/471,393, dated Mar. 9, 2016, 17 pages.
“Foreign Office Action”, CN Application No. 201310067356.2, dated Feb. 4, 2016, 15 Pages.
“Foreign Office Action”, CN Application No. 201310067603.9, dated Feb. 15, 2016, 12 Pages.
“Notice of Allowance”, U.S. Appl. No. 14/517,048, dated Feb. 24, 2016, 8 pages.
“Restriction Requirement”, U.S. Appl. No. 14/307,262, dated Mar. 21, 2016, 6 pages.
“Corrected Notice of Allowance”, U.S. Appl. No. 13/471,054, dated Jan. 11, 2016, 2 pages.
“Corrected Notice of Allowance”, U.S. Appl. No. 13/471,412, dated Feb. 16, 2016, 2 pages.
“Corrected Notice of Allowance”, U.S. Appl. No. 13/527,263, dated Jan. 4, 2016, 2 pages.
“Corrected Notice of Allowance”, U.S. Appl. No. 13/527,263, dated Jan. 11, 2016, 2 pages.
“Corrected Notice of Allowance”, U.S. Appl. No. 13/852,848, dated Jan. 29, 2016, 2 pages.
“Corrected Notice of Allowance”, U.S. Appl. No. 14/727,001, dated Jan. 25, 2016, 2 pages.
“Corrected Notice of Allowance”, U.S. Appl. No. 14/727,001, dated Dec. 15, 2015, 2 pages.
“Extended European Search Report”, EP Application No. 13857958.6, dated Dec. 18, 2015, 8 pages.
“Extended European Search Report”, EP Application No. 13858283.8, dated Nov. 23, 2015, 10 pages.
“Extended European Search Report”, EP Application No. 13858397.6, dated Nov. 30, 2015, 7 pages.
“Extended European Search Report”, EP Application No. 13858674.8, dated Nov. 27, 2015, 6 pages.
“Extended European Search Report”, EP Application No. 13860272.7, dated Dec. 14, 2015, 9 pages.
“Extended European Search Report”, EP Application No. 13860836.9, dated Nov. 27, 2015, 9 pages.
“Extended European Search Report”, EP Application No. 13861292.4, dated Nov. 23, 2015, 7 pages.
“Final Office Action”, U.S. Appl. No. 14/225,250, dated Jan. 29, 2016, 10 pages.
“Foreign Office Action”, CN Application No. 201310067373.6, dated Dec. 23, 2015, 15 Pages.
“Foreign Office Action”, CN Application No. 201310067429.8, dated Nov. 25, 2015, 12 Pages.
“Foreign Office Action”, CN Application No. 201310067631.0, dated Dec. 10, 2015, 11 Pages.
“Foreign Office Action”, CN Application No. 201310067641.4, dated Dec. 30, 2015, 12 Pages.
“Foreign Office Action”, CN Application No. 201310067808.7, dated Jan. 7, 2016, 7 Pages.
“Non-Final Office Action”, U.S. Appl. No. 13/492,232, dated Dec. 17, 2015, 11 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/891,109, dated Jan. 29, 2016, 8 pages.
“Notice of Allowance”, U.S. Appl. No. 13/527,263, dated Dec. 9, 2015, 6 pages.
“Notice of Allowance”, U.S. Appl. No. 13/647,479, dated Jan. 14, 2016, 8 pages.
“Notice of Allowance”, U.S. Appl. No. 14/727,001, dated Dec. 15, 2015, 2 pages.
“Restriction Requirement”, U.S. Appl. No. 14/794,182, dated Dec. 22, 2015, 6 pages.
“Corrected Notice of Allowance”, U.S. Appl. No. 14/994,737, dated Mar. 20, 2017, 2 pages.
“Corrected Notice of Allowance”, U.S. Appl. No. 14/994,737, dated Mar. 30, 2017, 2 pages.
“Foreign Office Action”, CN Application No. 201310067641.4, dated Jan. 23, 2017, 7 pages.
“Foreign Office Action”, CN Application No. 201380025290.9, dated Mar. 27, 2017, 13 pages.
“Foreign Office Action”, EP Application No. 13861292.4, dated Jan. 3, 2017, 4 pages.
“Foreign Office Action”, EP Application No. 14720018.2, dated Mar. 7, 2017, 7 pages.
“Non-Final Office Action”, U.S. Appl. No. 14/225,250, dated Mar. 20, 2017, 7 pages.
“Non-Final Office Action”, U.S. Appl. No. 15/083,008, dated Mar. 24, 2017, 13 pages.
“Non-Final Office Action”, U.S. Appl. No. 15/184,566, dated Mar. 30, 2017, 9 pages.
“Notice of Allowance”, U.S. Appl. No. 14/147,252, dated Apr. 14, 2017, 7 pages.
“Supplemental Notice of Allowance”, U.S. Appl. No. 13/653,218, dated Apr. 7, 2017, 3 pages.
“Corrected Notice of Allowance”, U.S. Appl. No. 14/482,983, dated Mar. 13, 2017, 5 pages.
“Corrected Notice of Allowance”, U.S. Appl. No. 14/994,737, dated Mar. 2, 2017, 2 pages.
“Foreign Notice of Allowance”, CN Application No. 201310067385.9, dated Feb. 20, 2017, 4 pages.
“Foreign Office Action”, JP Application No. 2014-560121, dated Dec. 20, 2016, 8 pages.
“Foreign Office Action”, JP Application No. 2015-512767, dated Jan. 24, 2017, 8 pages.
“Notice of Allowance”, U.S. Appl. No. 13/653,218, dated Mar. 10, 2017, 13 pages.
“Corrected Notice of Allowance”, U.S. Appl. No. 14/482,983, dated Feb. 22, 2017, 6 pages.
“Corrected Notice of Allowance”, U.S. Appl. No. 14/944,737, dated Jan. 18, 2017, 2 pages.
“Examiner's Answer to Appeal Brief”, U.S. Appl. No. 13/492,232, dated Feb. 9, 2017, 5 pages.
“Final Office Action”, U.S. Appl. No. 13/712,111, dated Aug. 11, 2016, 14 pages.
“Final Office Action”, U.S. Appl. No. 14/225,250, dated Dec. 27, 2016, 10 pages.
“Final Office Action”, U.S. Appl. No. 14/307,262, dated Sep. 29, 2016, 7 pages.
“Final Office Action”, U.S. Appl. No. 14/794,182, dated Sep. 15, 2016, 25 pages.
“Final Office Action”, U.S. Appl. No. 14/994,737, dated Aug. 26, 2016, 6 pages.
“Foreign Notice of Allowance”, CN Application No. 201310067429.8, dated Dec. 13, 2016, 4 pages.
“Foreign Notice of Allowance”, CN Application No. 201310067627.4, dated Oct. 24, 2016, 4 pages.
“Foreign Notice of Allowance”, CN Application No. 201310067631.0, dated Aug. 5, 2016, 4 pages.
“Foreign Notice of Allowance”, CN Application No. 201310096345.7, dated Feb. 4, 2017, 4 pages.
“Foreign Notice of Allowance”, CN Application No. 201310316114.2, dated Aug. 2, 2016, 4 pages.
“Foreign Office Action”, CN Application No. 201310067356.2, dated Oct. 9, 2016, 13 pages.
“Foreign Office Action”, CN Application No. 201310067373.6, dated Aug. 18, 2016, 16 pages.
“Foreign Office Action”, CN Application No. 201310067373.6, dated Dec. 27, 2016, 12 pages.
“Foreign Office Action”, CN Application No. 201310067385.9, dated Oct. 8, 2016, 6 pages.
“Foreign Office Action”, CN Application No. 201310067429.8, dated Jul. 20, 2016, 11 pages.
“Foreign Office Action”, CN Application No. 201310067523.3, dated Nov. 7, 2016, 12 pages.
“Foreign Office Action”, CN Application No. 201310067603.9, dated Oct. 17, 2016, 6 pages.
“Foreign Office Action”, CN Application No. 201310067622.1, dated Nov. 23, 2016, 16 pages.
“Foreign Office Action”, CN Application No. 201310067641.4, dated Jul. 27, 2016, 8 pages.
“Foreign Office Action”, CN Application No. 201310096345.7, dated Oct. 9, 2016, 15 pages.
“Foreign Office Action”, CN Application No. 201310225788.1, dated Sep. 1, 2016, 8 pages.
“Foreign Office Action”, CN Application No. 201380025290.9, dated Oct. 19, 2016, 8 pages.
“Foreign Office Action”, EP Application No. 13858283.8, dated Nov. 2, 2016, 5 pages.
“Foreign Office Action”, EP Application No. 13858674.8, dated Oct. 26, 2016, 5 pages.
“Foreign Office Action”, JP Application No. 2014-560073, dated Jan. 4, 2017, 7 pages.
“Foreign Office Action”, JP Application No. 2014-560119, dated Jan. 4, 2017, 6 pages.
“Foreign Office Action”, JP Application No. 2014-560120, dated Jan. 10, 2017, 5 pages.
“Foreign Office Action”, JP Application No. 2014-560122, dated Dec. 6, 2016, 10 pages.
“Non-Final Office Action”, U.S. Appl. No. 13/471,393, dated Oct. 5, 2016, 17 pages.
“Non-Final Office Action”, U.S. Appl. No. 14/147,252, dated Oct. 12, 2016, 17 pages.
“Non-Final Office Action”, U.S. Appl. No. 14/692,497, dated Feb. 22, 2017, 8 pages.
“Non-Final Office Action”, U.S. Appl. No. 14/704,423, dated Sep. 26, 2016, 11 pages.
“Non-Final Office Action”, U.S. Appl. No. 14/847,838, dated Dec. 16, 2016, 24 pages.
“Non-Final Office Action”, U.S. Appl. No. 15/050,072, dated Sep. 23, 2016, 5 pages.
“Notice of Allowance”, U.S. Appl. No. 14/307,262, dated Nov. 30, 2016, 5 pages.
“Notice of Allowance”, U.S. Appl. No. 14/482,983, dated Jan. 4, 2017, 10 pages.
“Notice of Allowance”, U.S. Appl. No. 14/994,737, dated Nov. 30, 2016, 5 pages.
“Restriction Requirement”, U.S. Appl. No. 13/653,218, dated Oct. 21, 2016, 6 pages.
“Supplemental Notice of Allowance”, U.S. Appl. No. 14/517,048, dated Aug. 10, 2016, 2 pages.
“Supplemental Notice of Allowance”, U.S. Appl. No. 14/517,048, dated Aug. 19, 2016, 2 pages.
“Corrected Notice of Allowance”, U.S. Appl. No. 14/847,838, dated Jun. 26, 2017, 3 pages.
“Final Office Action”, U.S. Appl. No. 13/471,393, dated Jul. 12, 2017, 20 pages.
“Final Office Action”, U.S. Appl. No. 14/162,529, dated Jun. 26, 2017, 28 pages.
“Foreign Office Action”, CN Application No. 201310067523.3, dated May 17, 2017, 8 pages.
“Foreign Office Action”, CN Application No. 201480036444.9, dated May 4, 2017, 8 pages.
“Foreign Office Action”, JP Application No. 2014-560072, dated Jun. 27, 2017, 4 pages.
“Foreign Office Action”, JP Application No. 2014-560120, dated Jun. 13, 2017, 5 pages.
“Foreign Office Action”, TW Application No. 103103802, dated May 11, 2017, 9 pages.
“Notice of Allowance”, U.S. Appl. No. 14/794,182, dated Jul. 13, 2017, 7 pages.
“Supplemental Notice of Allowance”, U.S. Appl. No. 13/653,218, dated Jun. 19, 2017, 2 pages.
“Supplemental Notice of Allowance”, U.S. Appl. No. 14/147,252, dated Jun. 21, 2017, 2 pages.
“Corrected Notice of Allowance”, U.S. Appl. No. 13/780,228, dated Apr. 17, 2017, 3 pages.
“Corrected Notice of Allowance”, U.S. Appl. No. 14/994,737, dated May 10, 2017, 2 pages.
“Foreign Notice of Allowance”, JP Application No. 2014-560119, dated Apr. 18, 2017, 4 pages.
“Foreign Office Action”, CN Application No. 201310067356.2, dated Mar. 20, 2017, 9 pages.
“Foreign Office Action”, CN Application No. 201310067603.9, dated Apr. 19, 2017, 12 pages.
“Foreign Office Action”, CN Application No. 201480006798.9, dated Apr. 13, 2017, 12 pages.
“Foreign Office Action”, EP Application No. 13730686.6, dated Mar. 27, 2017, 6 pages.
“Non-Final Office Action”, U.S. Appl. No. 15/050,072, dated May 12, 2017, 9 pages.
“Non-Final Office Action”, U.S. Appl. No. 15/425,799, dated Jun. 2, 2017, 12 pages.
“Notice of Allowance”, U.S. Appl. No. 13/780,228, dated Mar. 13, 2017, 7 pages.
“Notice of Allowance”, U.S. Appl. No. 14/847,838, dated May 4, 2017, 10 pages.
“Supplemental Notice of Allowance”, U.S. Appl. No. 13/653,218, dated May 16, 2017, 2 pages.
“Supplemental Notice of Allowance”, U.S. Appl. No. 14/147,252, dated May 3, 2017, 2 pages.
“Supplemental Notice of Allowance”, U.S. Appl. No. 14/147,252, dated May 15, 2017, 2 pages.
“Corrected Notice of Allowance”, U.S. Appl. No. 15/426,116, dated Oct. 10, 2017, 2 pages.
“Final Office Action”, U.S. Appl. No. 14/225,250, dated Aug. 17, 2017, 8 pages.
“Foreign Office Action”, CN Application No. 201310225788.1, dated Sep. 20, 2017, 10 pages.
“Foreign Office Action”, CN Application No. 201380025290.9, dated Jul. 18, 2017, 9 pages.
“Foreign Office Action”, EP Application No. 13858283.8, dated Oct. 18, 2017, 4 pages.
“Foreign Office Action”, EP Application No. 13858403.2, dated Aug. 1, 2017, 5 pages.
“Foreign Office Action”, JP Application No. 2014-560073, dated Jul. 25, 2017, 12 pages.
“Foreign Office Action”, JP Application No. 2015-512767, dated Jul. 25, 2017, 6 pages.
“Non-Final Office Action”, U.S. Appl. No. 15/083,008, dated Oct. 17, 2017, 16 pages.
“Non-Final Office Action”, U.S. Appl. No. 15/190,961, dated Nov. 6, 2017, 11 pages.
“Notice of Allowance”, U.S. Appl. No. 14/225,250, dated Nov. 6, 2017, 5 pages.
“Notice of Allowance”, U.S. Appl. No. 14/692,497, dated Sep. 8, 2017, 7 pages.
“Notice of Allowance”, U.S. Appl. No. 15/426,116, dated Sep. 14, 2017, 7 pages.
“Supplemental Notice of Allowance”, U.S. Appl. No. 14/692,497, dated Sep. 18, 2017, 2 pages.
“Supplemental Notice of Allowance”, U.S. Appl. No. 14/692,497, dated Oct. 6, 2017, 2 pages.
“Supplemental Notice of Allowance”, U.S. Appl. No. 14/794,182, dated Aug. 14, 2017, 2 pages.
“Supplemental Notice of Allowance”, U.S. Appl. No. 14/847,838, 0/17/2017, 4 pages.
“Notice of Allowance Issued in Japanese Patent Application No. 2014-560073”, dated Oct. 31, 2017, 4 Pages.
“Final Office Action Issued in U.S. Appl. No. 15/425,779”, dated Nov. 20, 2017, 7 Pages.
“Corrected Notice of Allowance Issued in U.S. Appl. No. 141225,250”, dated Nov. 16, 2017, 2 Pages.
“Foreign Office Action Issued in European Application No. 13860271.9”, dated Nov. 7, 2017, 5 Pages.
“Foreign Office Action Issued in European Application No. 14703759.2”, dated Oct. 31, 2017, 8 Pages.
“Foreign Office Action”, EP Application No. 13858834.8, dated Jan. 18, 2018, 5 pages.
“Non-Final Office Action”, U.S. Appl. No. 15/249,952, dated Jan. 24, 2018, 8 pages.
“Foreign Office Action”, CN Application No. 201310067373.6, dated Dec. 11, 2017, 9 pages.
Related Publications (1)
Number Date Country
20130229380 A1 Sep 2013 US
Provisional Applications (7)
Number Date Country
61606321 Mar 2012 US
61606301 Mar 2012 US
61606313 Mar 2012 US
61606333 Mar 2012 US
61613745 Mar 2012 US
61606336 Mar 2012 US
61607451 Mar 2012 US