The present disclosure relates generally to a sensor module, and more particularly to a hybrid sensor module including an optical navigation sensor and a touch sensor and a method of operating the same.
Data processing systems, such as personal computers, tablet computers, entertainment systems, game consoles, and cellular telephones, commonly include a human interface device (HID) for data input and/or controlling cursor movement. One widely utilized HID is a touch pad or touchscreen utilizing a touch sensor, such as a capacitive or resistive sensing system, to sense and measure proximity, position, displacement or acceleration of an object, such as a finger or stylus. Another HID utilizes an optical finger navigation system or ONS, which uses changes in an image or pattern of light reflected from an object position on or over a surface of the HID to detect and measure proximity, position, displacement or acceleration of the object. Existing capacitive or resistive touch sensors suffer from a number of disadvantages including limited tracking resolution, and limited tracking speed. In addition, touchscreens utilizing capacitive sensing systems have difficulty tracking objects covered by most types of electrically insulating material, such as a gloved finger. However, ONS utilized in existing optical track-pad and optical track-ball applications are prone to manufacturing process variation, requiring a high level of optical calibration, have large performance variation across tracking speeds, and have a relatively small active-area due to limitations of light sources used. In addition, because ONS work well only within a narrow range around a nominal design height separating the object from the ONS and due to the lift-cutoff requirement in most applications that using ONS lift detection is necessary to detect when the ONS is separated from the object by more than a maximum lift-height to prevent spurious data from the HID. Previous ONS that rely on an optical lift-cut-off mechanism based on image defocus are not satisfactory for non-image based ONS. In addition, non-image-defocus based optical lift-cut-off mechanism employed in the non-image based ONS, is sensitive to manufacturing process variation, requiring rather tight manufacturing tolerances, and/or reducing tracking.
These and various other features of a hybrid sensor module and methods of operation will be apparent upon reading of the following detailed description in conjunction with the accompanying drawings and the appended claims provided below, where:
Hybrid sensor modules and methods are provided for use in an input device to sense relative movement between the hybrid sensor module and a tracking surface, such as a surface of a stylus or finger. In one embodiment, the hybrid sensor module includes: (i) a touch sensor configured to sense motion of an object in proximity to a surface of the touch sensor; (ii) an optical navigation sensor (ONS) configured to illuminate the object through the surface of the touch sensor and to sense motion of the object based on light returned from the object; and (iii) a controller electrically coupled to the touch sensor and ONS to process the sensed motion of the object and to generate an output signal in response to the sensed motion, wherein the controller is configured to dynamically adjust a tracking resolution of the hybrid sensor module based on a characteristic of the sensed motion.
The drawings described are only schematic and are non-limiting. In the drawings, the size of some of the elements may be exaggerated and not drawn to scale for illustrative purposes. The dimensions and the relative dimensions may not correspond to actual reductions to practice of the invention. For purposes of clarity, many details of optical navigation sensors and touch sensors in general, and principles of operation of capacitive and resistive touch sensors in particular, which are widely known and not relevant to the present apparatus and method have been omitted from the following description.
An electronic system or device 100, such as a tablet computer, smartphone or human interface device (HID), utilizing a hybrid position sensor for detecting a presence of an object 102, such as a stylus or finger, is shown in
A block diagram of a hybrid sensor module 200 is illustrated in
The touch sensor 204 is also configured to sense the presence, location and/or motion of an object or objects in proximity to the surface of the touch sensor, and to provide data or signals to controller 206. In one embodiment, the hybrid sensor module is a capacitive sensor, and can include a matrix or capacitive sensing array 208 of capacitive sensor elements 210, each formed from an intersection of one of a number of receive electrodes 212 and transmit electrodes 214. The sensor array 208 is coupled to capacitive sensor control circuitry, which can be embodied in the controller 206, as shown, or can include a separate, distinct component or integrated circuit (IC). Generally, the capacitive sensor control circuitry can include a relaxation oscillator or other means to convert a capacitance, a counter or timer to measure the oscillator output, and processing logic implemented in firmware, hardware or software to convert the count value (e.g., capacitance value) into a sensor element detection decision (also referred to as switch detection decision) representing a measured value or relative magnitude. It should be noted that there are various known methods for measuring capacitance, such as current versus voltage phase shift measurement, resistor-capacitor charge timing, capacitive bridge divider, charge transfer, successive approximation, sigma-delta modulators, charge-accumulation circuits, field effect, mutual capacitance, frequency shift, or other capacitance measurement algorithms, any one or more of which can be used without departing from the scope of the present disclosure.
Alternatively, in another embodiment the hybrid sensor module includes a resistive sensing array, and the sensor elements 210 of the resistive sensor array 208 can include resistive elements formed from an intersection of a number of top flexible electrodes 212 one or more of a number of lower electrodes 214 separated by an air gap or microdots. As with the capacitive sensor embodiment described above, the resistive sensor array 208 is coupled to resistive sensor control circuitry embodied in the controller 206, as shown, or can in a separate, distinct component or IC. The sensor elements 210 of the resistive sensor array 208 respond to pressure on the surface of the touch sensor caused by a finger or other object in contact with the surface.
Signals provided to the controller 206 from a capacitive or resistive touch sensor 204 can include data representing XY motion or position of the object relative to the surface of the touch sensor 204, data representing gestures, such as tapping or double tapping the surface, and data representing lift-off detection of an object from the surface. By lift-off detection it is meant the touch sensor 204 detects when the object is no longer in contact with a portion of the surface overlying the ONS 202 or has been lifted to a lift height separating the object from the OFN exceeding a manufacturer or user specified maximum lift height. Lift-off detection data or signals can be used by the controller 206 to cut-off or to interrupt data from the ONS 202 representing XY motion or position of the object relative to the ONS, and/or data representing gestures. It is noted that capacitive or resistive sensors can generally provide touch detection, lift-detection, and accurate tracking resolution of XY motion at relatively low tracking speeds.
The controller 206 includes control circuitry to operate the ONS 202, and, as noted above, to operate the capacitive or resistive touch sensor 204, and digital signal processor (DSP) or DSP circuitry to process signals and data received from the ONS 202 and touch sensor 204 to generate one or more output signals including, for example, output signals relating to an XY motion or position of one or more objects proximal to the surface of the hybrid sensor module 200, gesture data, lift-detection and an optical signal. In addition, the controller 206 further includes circuitry and/or firmware to automatically and dynamically adjust a tracking resolution of the hybrid sensor module 200 module based on a characteristic of the signals and/or the sensed motion received from the ONS 202 and/or touch sensor 204. In certain embodiments, the controller 206 is configured to generate output signals in response to a weighted combination of the sensed motion from the ONS 202 and from the touch sensor 204. For example, the controller 206 can be configured to give greater weight to the sensed motion from the ONS 202 when the speed of the sensed motion is more than a specified speed, and to give greater weight to the sensed motion from the touch sensor 204 when the speed of the sensed motion is less than the specified speed. Exemplary values for the specified speed in tracking an object, such as a finger, can include speeds in the range of from about 1 cm/sec to about 50 cm/sec.
In other embodiments, controller 206 further includes circuitry and/or firmware to detect a lift height between the object and the surface, and to give less weight to the sensed motion from the ONS 202 when the lift height exceeds the specified maximum lift height, or to give less weight to the sensed motion from the touch sensor 204 when strength of a signal from the touch sensor 204 is less than a specified amount. One example of the later embodiment can be particularly useful when the touch sensor 204 is a capacitive sensor and the object it is attempting to sense is an insulating object, such as a gloved finger.
In one embodiment, shown in
Referring to
The substrate 308 further includes a controller 318 electrically coupled to the sensor IC 312 and through a connector 320, such as a ribbon connector, to the touch sensor 306. The controller 318 can be integrally formed on a common IC shared with the sensor IC 312 or can include a separate, distinct component or IC mounted on or electrically coupled to the substrate 308.
The hybrid position sensor 300 further includes a light baffle or blocking element 322 having a number of openings formed therein to enable light to be transmitted from the light source 310 to the tracking-surface 302 and from the tracking-surface to the PDA 314 while shielding the PDA from light shining directly from the light source and/or ambient or environmental light passing through the surface 304 but not returned from the tracking surface. A first opening or aperture 324 in a light path 326 between the light source 310 and the tracking-surface 302 permits illumination of the tracking surface. A second opening or aperture 328 in a light path 330 between the tracking-surface 302 and the PDA 314 controls the light scattered onto the PDA and blocking environmental light.
In another embodiment, shown in
The surface of the touch sensor 402 includes a first portion 428 extending over the display 404 and a second portion 430 not extending over the display. The second portion 430 of the surface of the touch sensor 402 includes an optical hot-spot 432 overlying the ONS and adjacent to the display 404. In addition, the second portion 430 of the surface of the touch sensor 402 can further include a number of areas or regions 434 for implementing a number of touch sensitive buttons or keys using the touch sensor.
In one embodiment, shown in
In certain embodiments, the ONS described above is a dark field or speckle-based ONS detecting and utilizing changes in a light pattern in light scattered from the object or tracking-surface to sense motion, position or gestures, rather than correlating successive images of the object or tracking surface. Operating principles of a speckle-based ONS will now be described with reference to
Referring to
A speckle sensitive photo-detector array can include one or more linear or one-dimensional (1D) or a two-dimensional (2D) comb-array having multiple detectors or photosensitive elements arranged in a two-dimensional configuration.
A linear or 1D comb-array is an array having multiple photosensitive elements that are connected in a periodic manner, so the array acts as a fixed template that integrates one spatial frequency component of the signal. An embodiment of one such 1D comb-array is shown in
Referring to
In one embodiment the photo-detector array includes photodiodes or photosensitive elements are arranged in two dimensions (2D), as shown in
Embodiments of methods for operating a hybrid sensor module will now be described with reference to the flowchart of
In a first block, motion of a first object or finger in proximity to a surface of a touch sensor is sensed by the touch sensor (902). As noted above, in embodiments in which the touch sensor is a resistive touch sensor or includes a resistive array, the object or finger must be in contact with the surface. Alternatively, in embodiments in which the touch sensor is a capacitive touch sensor or includes a capacitive array, the object may be proximal to surface without being in contact. The first object is also illuminated through the surface of the touch sensor with a light source in the ONS, and motion of the first object sensed by the ONS based on light returned from the object (904). Next, a controller electrically coupled to the touch sensor and the ONS processes the motion of the first object sensed by the touch sensor and the ONS to generate an output signal in response to the sensed motion (906), and automatically or dynamically adjusts a tracking resolution of the hybrid position module based on a characteristic of the sensed motion (908).
As noted above, adjusting the tracking resolution can include weighting and combining the sensed motion from the ONS and from the touch sensor. In one embodiment, the controller gives greater weight to the sensed motion from the ONS when the speed or acceleration of the sensed motion is greater than a specified amount, and gives greater weight to the sensed motion from the touch sensor when the speed or acceleration is less than the specified speed. Alternatively or additionally, the method can further include detecting with the touch sensor when a lift height between the first object and the surface exceeds a maximum lift height adjusting the tracking resolution can include giving less weight to the sensed motion from the ONS and correspondingly greater weight to the sensed motion from the touch sensor. Similarly, adjusting the tracking resolution can alternatively or additionally include giving greater weight to the sensed motion from the ONS when strength of a signal representing the sensed motion from the touch sensor is less than a specified amount.
Optionally, in certain embodiments, such as that described above with respect to
Thus, embodiments of a hybrid sensor module and methods for operating the same have been described. Although the present disclosure has been described with reference to specific exemplary embodiments, it will be evident that various modifications and changes may be made to these embodiments without departing from the broader spirit and scope of the disclosure. Accordingly, the specification and drawings are to be regarded in an illustrative rather than a restrictive sense.
The Abstract of the Disclosure is provided to comply with 37 C.F.R. §1.72(b), requiring an abstract that will allow the reader to quickly ascertain the nature of one or more embodiments of the technical disclosure. It is submitted with the understanding that it will not be used to interpret or limit the scope or meaning of the claims. In addition, in the foregoing Detailed Description, it can be seen that various features are grouped together in a single embodiment for the purpose of streamlining the disclosure. This method of disclosure is not to be interpreted as reflecting an intention that the claimed embodiments require more features than are expressly recited in each claim. Rather, as the following claims reflect, inventive subject matter lies in less than all features of a single disclosed embodiment. Thus, the following claims are hereby incorporated into the Detailed Description, with each claim standing on its own as a separate embodiment.
In the forgoing description, for purposes of explanation, numerous specific details have been set forth in order to provide a thorough understanding of the hot de-latch system and method of the present disclosure. It will be evident however to one skilled in the art that the present interface device and method may be practiced without these specific details. In other instances, well-known structures, and techniques are not shown in detail or are shown in block diagram form in order to avoid unnecessarily obscuring an understanding of this description.
Reference in the description to “one embodiment” or “an embodiment” means that a particular feature, structure, or characteristic described in connection with the embodiment is included in at least one embodiment of the system or method. The appearances of the phrase “one embodiment” in various places in the specification do not necessarily all refer to the same embodiment. The term “to couple” as used herein may include both to directly electrically connect two or more components or elements and to indirectly connect through one or more intervening components.
Number | Name | Date | Kind |
---|---|---|---|
3922093 | Dandliker et al. | Nov 1975 | A |
4546347 | Kirsch | Oct 1985 | A |
4799055 | Nestler et al. | Jan 1989 | A |
5015096 | Kowalski et al. | May 1991 | A |
5288993 | Bidiville et al. | Feb 1994 | A |
5473344 | Bacon et al. | Dec 1995 | A |
5578813 | Allen et al. | Nov 1996 | A |
5644139 | Allen et al. | Jul 1997 | A |
D382550 | Kaneko et al. | Aug 1997 | S |
D385542 | Kaneko et al. | Oct 1997 | S |
5703356 | Bidiville et al. | Dec 1997 | A |
5729008 | Blalock et al. | Mar 1998 | A |
5729009 | Daendliker et al. | Mar 1998 | A |
5786804 | Gordon | Jul 1998 | A |
5825044 | Allen et al. | Oct 1998 | A |
5854482 | Bidiville et al. | Dec 1998 | A |
5907152 | Daendliker et al. | May 1999 | A |
5933102 | Miller et al. | Aug 1999 | A |
5963197 | Bacon et al. | Oct 1999 | A |
5994710 | Knee et al. | Nov 1999 | A |
6031218 | Piot et al. | Feb 2000 | A |
6037643 | Knee | Mar 2000 | A |
6057540 | Gordon et al. | May 2000 | A |
6097371 | Siddiqui et al. | Aug 2000 | A |
6151015 | Badyal et al. | Nov 2000 | A |
6172354 | Adan et al. | Jan 2001 | B1 |
6215417 | Krass et al. | Apr 2001 | B1 |
6225617 | Daendliker et al. | May 2001 | B1 |
6233368 | Badyal et al. | May 2001 | B1 |
6281881 | Siddiqui et al. | Aug 2001 | B1 |
6281882 | Gordon et al. | Aug 2001 | B1 |
6300936 | Braun et al. | Oct 2001 | B1 |
6326950 | Liu | Dec 2001 | B1 |
6330057 | Lederer et al. | Dec 2001 | B1 |
6351257 | Liu | Feb 2002 | B1 |
6396479 | Gordon | May 2002 | B2 |
6421045 | Venkat et al. | Jul 2002 | B1 |
6424407 | Kinrot et al. | Jul 2002 | B1 |
6433780 | Gordon et al. | Aug 2002 | B1 |
6452683 | Kinrot et al. | Sep 2002 | B1 |
6455840 | Oliver et al. | Sep 2002 | B1 |
D464352 | Kerestegian | Oct 2002 | S |
6462330 | Venkat et al. | Oct 2002 | B1 |
6476970 | Smith | Nov 2002 | B1 |
6504530 | Wilson et al. | Jan 2003 | B1 |
6529184 | Julienne | Mar 2003 | B1 |
6585158 | Norskog | Jul 2003 | B2 |
6603111 | Dietz et al. | Aug 2003 | B2 |
6621483 | Wallace et al. | Sep 2003 | B2 |
6642506 | Nahum et al. | Nov 2003 | B1 |
6657184 | Anderson et al. | Dec 2003 | B2 |
6664948 | Crane et al. | Dec 2003 | B2 |
6674475 | Anderson | Jan 2004 | B1 |
6677929 | Gordon et al. | Jan 2004 | B2 |
6703599 | Casebolt et al. | Mar 2004 | B1 |
6737636 | Dietz et al. | May 2004 | B2 |
6774351 | Black | Aug 2004 | B2 |
6774915 | Rensberger | Aug 2004 | B2 |
6795056 | Norskog et al. | Sep 2004 | B2 |
6809723 | Davis | Oct 2004 | B2 |
6819314 | Black | Nov 2004 | B2 |
6823077 | Dietz et al. | Nov 2004 | B2 |
6859196 | Kehlstadt | Feb 2005 | B2 |
7030860 | Hsu et al. | Apr 2006 | B1 |
7042575 | Carlisle et al. | May 2006 | B2 |
7116117 | Nakano et al. | Oct 2006 | B2 |
7161582 | Bathiche et al. | Jan 2007 | B2 |
7199350 | Chien | Apr 2007 | B2 |
7221356 | Oliver et al. | May 2007 | B2 |
7274808 | Baharav et al. | Sep 2007 | B2 |
7295186 | Brosnan | Nov 2007 | B2 |
7313255 | Machida et al. | Dec 2007 | B2 |
7324086 | Kong | Jan 2008 | B2 |
7420542 | Butterworth et al. | Sep 2008 | B2 |
7525082 | Itagaki | Apr 2009 | B2 |
7557338 | Gruhlke et al. | Jul 2009 | B2 |
7868281 | Leong et al. | Jan 2011 | B2 |
8031176 | Sanders et al. | Oct 2011 | B1 |
8081162 | Teoh et al. | Dec 2011 | B2 |
8111344 | Moon et al. | Feb 2012 | B2 |
8400409 | Francis | Mar 2013 | B1 |
8416191 | Gao et al. | Apr 2013 | B2 |
20020130835 | Brosnan | Sep 2002 | A1 |
20020158300 | Gee | Oct 2002 | A1 |
20020190953 | Gordon et al. | Dec 2002 | A1 |
20030034959 | Davis et al. | Feb 2003 | A1 |
20040084610 | Leong et al. | May 2004 | A1 |
20040189593 | Koay | Sep 2004 | A1 |
20040208348 | Baharav et al. | Oct 2004 | A1 |
20050024336 | Xie et al. | Feb 2005 | A1 |
20050024623 | Xie et al. | Feb 2005 | A1 |
20050083303 | Schroeder et al. | Apr 2005 | A1 |
20050248534 | Kehlstadt | Nov 2005 | A1 |
20060048071 | Jarrett et al. | Mar 2006 | A1 |
20060091298 | Xie et al. | May 2006 | A1 |
20060279548 | Geaghan | Dec 2006 | A1 |
20070008286 | Theytaz et al. | Jan 2007 | A1 |
20070052684 | Gruhlke et al. | Mar 2007 | A1 |
20070057157 | Trisnadi et al. | Mar 2007 | A1 |
20070075710 | Hargreaves et al. | Apr 2007 | A1 |
20070139381 | Spurlock et al. | Jun 2007 | A1 |
20070146318 | Juh et al. | Jun 2007 | A1 |
20070215793 | Gruhlke et al. | Sep 2007 | A1 |
20070291001 | Trisnadi et al. | Dec 2007 | A1 |
20080006762 | Fadell et al. | Jan 2008 | A1 |
20080007526 | Xu et al. | Jan 2008 | A1 |
20080048972 | Kakarala | Feb 2008 | A1 |
20080061219 | Lee et al. | Mar 2008 | A1 |
20080094356 | Ording et al. | Apr 2008 | A1 |
20080158158 | Cheah et al. | Jul 2008 | A1 |
20080167834 | Herz et al. | Jul 2008 | A1 |
20080174570 | Jobs et al. | Jul 2008 | A1 |
20090041476 | Ann et al. | Feb 2009 | A1 |
20090166411 | Kramer et al. | Jul 2009 | A1 |
20090244026 | Purdy et al. | Oct 2009 | A1 |
20100013771 | Cheng | Jan 2010 | A1 |
20100073328 | Lynch et al. | Mar 2010 | A1 |
20100302208 | Brosnan et al. | Dec 2010 | A1 |
20100315337 | Ferren et al. | Dec 2010 | A1 |
20100315413 | Izadi et al. | Dec 2010 | A1 |
20110018840 | Fann et al. | Jan 2011 | A1 |
20110141014 | Liao et al. | Jun 2011 | A1 |
20110234539 | Liao et al. | Sep 2011 | A1 |
20110248961 | Svajda et al. | Oct 2011 | A1 |
20110310018 | Song et al. | Dec 2011 | A1 |
20110316790 | Ollila et al. | Dec 2011 | A1 |
20120002693 | Hasegawa et al. | Jan 2012 | A1 |
20120127076 | Song et al. | May 2012 | A1 |
20120176339 | Chen et al. | Jul 2012 | A1 |
20120320385 | Mu et al. | Dec 2012 | A1 |
Number | Date | Country |
---|---|---|
101593785 | Dec 2009 | CN |
2009031666 | Feb 2009 | JP |
20070008743 | Jan 2007 | KR |
20110051165 | May 2011 | KR |
2010081652 | Jul 2010 | WO |
Entry |
---|
ADNB-6011-EV and ADNB-6012-EV High Performance Laser Mouse Bundles Data Sheet, Aug. 17, 2007, Avago Technologies, 52 pages. |
ADNB-6532 Small Form Factor Laser Stream Mouse Bundle Data Sheet, Nov. 10, 2006, Avago Technologies, 44 pages. |
ADNK-6013 Optical Mouse Designer's Kit Design Guide, Mar. 30, 2006, Avago Technologies, 24 pages. |
Angelopoulou, Elli; “The Reflectance Spectrum of Human Skin,” 1999; Department of Computer & Information Science, University of Pennsylvania; pp. 1-14. |
U.S. Appl. No. 12/009,863: “Optical Navigation System Using a Single-Package Motion Sensor” Steven Sanders et al., filed on Jan. 22, 2008: 27 pages. |
U.S. Appl. No. 12/218,021: “User Interface Devices, Methods, and Computer Readable Media for Sensing Movement of an Actuator Across a Surface of a Window,” Mark Francis, filed on Jul. 10, 2008; 26 pages. |
U.S. Appl. No. 13/231,176: Optical Navigation Module With Programmable Lift-Cut-off Ke-Cai Zeng et al., filed on Sep. 13, 2011; 33 pages. |
U.S. Appl. No. 13/248,355 “Optical Navigation Module With Capacitive Sensor,” Jinghui Mu et al., Filed on Sep. 29, 2011; 33 pages. |
U.S. Appl. No. 13/252,768: “Optical Navigation System Using a Single-Package Motion Sensor” Steven Sanders et al., filed on Oct. 4, 2011; 28 pages. |
Application No. PCT/US11/53552 “An Optical Navigation Module With Capacitive Sensor,” Filed on Sep. 27, 2011; 31 pages. |
International Search Report for International Application No. PCT/US2011/053552 dated May 23, 2012; 3 pages. |
KIPO ISR/WO for Application No. PCT/US2011/053552 dated May 23, 2012; 10 pages. |
USPTO Advisory Action for U.S. Appl. No. 12/218,021 dated Nov. 7, 2011; 3 pages. |
USPTO Advisory Action for U.S. Appl. No. 12/218,021 dated May 9, 2012; 3 pages. |
USPTO Advisory Action for U.S. Appl. No. 13/231,176 dated Dec. 6, 2013; 3 pages. |
USPTO Final Office Action for U.S. Appl. No. 13/231,176 dated Oct. 8, 2013; 18 pages. |
USPTO Final Rejection for U.S. Appl. No. 12/218,021 dated Feb. 22, 2012; 45 pages. |
USPTO Final Rejection for U.S. Appl. No. 12/218,021 dated Aug. 29, 2011; 41 pages. |
USPTO Final Rejection for U.S. Appl. No. 13/252,768 dated Dec. 7, 2012; 7 pages. |
USPTO Non Final rejection for U.S. Appl. No. 13/231,176 dated Mar. 26, 2014; 15 pages. |
USPTO Non Final Rejection for U.S. Appl. No. 13/248,355 dated May 7, 2014; 27 pages. |
USPTO Non Final Rejection for U.S. Appl. No. 13/252,768 dated Jun. 25, 2012; 12 pages. |
USPTO Non-Final Rejection for U.S. Appl. No. 12/009,863 dated Dec. 8, 2010; 13 pages. |
USPTO Non-Final Rejection for U.S. Appl. No. 12/218,021 dated Apr. 27, 2011; 34 pages. |
USPTO Non-Final Rejection for U.S. Appl. No. 13/231,176 dated Jun. 21, 2013; 22 pages. |
USPTO Notice of Allowance for U.S. Appl. No. 12/009,863 dated May 26, 2011; 5 pages. |
USPTO Notice of Allowance for U.S. Appl. No. 12/218,021 dated Jan. 30, 2013; 9 pages. |
USPTO Notice of Allowance for U.S. Appl. No. 12/218,021 dated Oct. 5, 2012; 9 pages. |
USPTO Notice of Allowance for U.S. Appl. No. 13/252,768 dated Jan. 31, 2013; 8 pages. |
USPTO Notice of Allowance for U.S. Appl. No. 13/252,768 dated May 22, 2013; 8 pages. |
USPTO Notice of Allowance for U.S. Appl. No. 13/252,768 dated Sep. 6, 2013; 6 pages. |
USPTO Notice of Allowance for U.S. Appl. No. 13/252,768 dated Dec. 10, 2013; 7 pages. |
USPTO Requirement for Restriction for U.S. Appl. No. 13/231,176 dated May 21, 2013; 5 pages. |
Written Opinion of the International Searching Authority for International Application No. PCT/US2011/053552 dated May 23, 2012; 5 pages. |