This invention relates generally to a method and apparatus for determining push interaction, and more particularly to an apparatus and method for determining a user intended push interaction employing only a standard webcam.
Many gesture recognition systems typically employ complex and complicated time of flight cameras, stereo cameras, or other such sophisticated systems. In order to determine action in a z-axis direction relative to a camera, time of flight cameras are equipped with a system for determining a distance from the camera at any time. Similarly, a stereo camera system is able to employ a triangulation or other stereo distance determination. Each of these systems is a complex system not typically employed in a standard computer laptop or other mobile device. Therefore, these systems are not available for use in a more typical computer or mobile device environment.
In order to overcome these drawbacks, various systems employ complex hybrid systems to overcome the z-axis. For example, U.S. Pat. No. 7,774,075 presents a method and apparatus for determining an object's three-dimensional location (i.e. real world coordinates) using the audio-video infrastructure of a 3G cellular phone or a 3C (Computer, Communications, Consumer) electronic device. A first detection device (e.g. a camera) is used to capture images of the objects. The captured image data is used to compute location data of the object in a first two-dimensional plane. A second detection device (e.g. microphone or infrared detector) may be used to collect additional location data in a second plane, which when combined with image data from the captured images allows the determination of the real world coordinates (x, y, z) of the object. The real-world coordinate data may be used in various applications.
In an alternative embodiment of this '075 patent, if the size of an object of interest is known or can be calculated, and the size of the projected image does not vary due to rotation of the object, a single camera (e.g. the camera in a 3G or 3C mobile device) may be used to obtain three-dimensional coordinate data for the applications.
For use in mobile or other unstable environments, however, calculation of a precise 3D coordinate may be difficult because of various factors. Furthermore, there is no way to guarantee that the object will not rotate. Furthermore, a mobile or other portable platform may result in a webcam that is likely to be moved about quite a bit, thus potentially requiring frequent recalibration of the system. Furthermore, movement may be improperly interpreted as deliberate action, thus resulting in false actions being taken by the system. Therefore, it would be beneficial to provide a method and apparatus that overcome the drawbacks of the prior art.
In accordance with various embodiments of the present invention, a push interaction useful for determining movement in a z-axis direction relative to a webcam is provided. Rather than relying upon complex systems for making such a determination, a simple system for determining an initial size of an object, and then later comparing a size of an object to that initial size to determine intentional motion in the z-axis direction relative to the camera is provided. Further object tracking in the x and y axis directions places the object at a particular location in a field of view, allowing for selection between one or more objects included in that field of view.
Depth or proximity indications may be provided as feedback to a user of the system. Such indications may be provided in a visual, audible, or other appropriate fashion, allowing the user to determine when a particular selection or other indication has been made, and aiding the user in properly positioning an object. Through the use of such selection mechanism, a user may simulate the selection mechanism employed with a touch screen or conventional keyboard without making contact with any particular surface. Such application may be desirable when considering publically available keypads or other input devices to avoid germs or the like. Furthermore, by receiving such a selection indicators, a user may be encouraged or enabled to change their selection before such selection is confirmed.
While the invention is generally described as applicable to a simple webcam provided in a computer, mobile device or other computing apparatus, the invention may also be applied to any single camera that does not otherwise have the ability to independently determine distance, such as with a single infrared or other camera.
Still other objects and advantages of the invention will in part be obvious and will in part be apparent from the specification and drawings.
The invention accordingly comprises the several steps and the relation of one or more of such steps with respect to each of the others, and the apparatus embodying features of construction, combinations of elements and arrangement of parts that are adapted to affect such steps, all as exemplified in the following detailed disclosure, and the scope of the invention will be indicated in the claims.
For a more complete understanding of the invention, reference is made to the following description and accompanying drawings, in which:
Detailed embodiments of the present invention will now be described, making reference to the following figures in which like reference number denote like structure or steps. Referring first to
If the inquiry at step 125 is answered in the affirmative, and it is therefore determined that an object has been detected in the region of interest, processing continues at step 130 where an initial size of the object (S0) is determined. This initial size is one that is not necessarily related to the absolute size of the object, but is rather related to the perceived size of the object when placed within the region of interest. Thus, when shown to a webcam, the perceived size is determined, which will be based upon the actual size of the object, as well as distance from the webcam.
Once the initial size of the object has been determined, the object is tracked within the region of interest at step 135, and at step 140, it is confirmed that the object is able to be tracked. If at step 140 it is determined that the object cannot be tracked, processing once again returns to step 120, and the object is detected anew. If on the other hand, it is determined at step 140 that the object is being properly tracked, processing passes to step 145 where a current size of the object (Snew) is measured.
After such a measurement, at step 150 it is determined whether a ratio of the new size to the old size (Snew/S0) is greater than a predetermined threshold. If the inquiry at step 150 is answered in the negative, and therefore the ratio is not greater than the predetermined threshold, processing returns to step 135, and tracking of the object within the region of interest continues. If on the other hand, if the inquiry at step 150 is answered in the affirmative, and it is therefore determined that the ratio is greater than a predetermined threshold, processing continues at step 160 where it is confirmed that a push has been observed. Processing then ends at step 170.
In an alternative embodiment of the invention, in order to improve the accuracy of determining whether a push has been observed, a multi-level threshold scheme may be employed. Thus, as is shown in
If the inquiry at step 250 is answered in the affirmative, and it is therefore determined that the size ratio (Snew/S0) is greater than the threshold value associated with the first threshold number (TN=1), processing continues with step 255 where it is inquired as to whether the threshold number is equal to a predetermined number, indicative of the number of threshold tests are to be performed. If this inquiry at step 255 is answered in the negative, and it is therefore determined that the current threshold number is not equal to the desired threshold number, then processing passes to step 257 where the threshold number is incremented by one (in this case to two), and processing returns to step 135.
This processing continues for the predetermined number of threshold tests desired, each sequence using a threshold associated with the then-current threshold number. It is anticipated that each such threshold is larger than the prior threshold, thus allowing for a determination of continued motion forward towards the webcam, but of course any such threshold sequence may be employed to test for any simple or complex desired motion. Once the threshold has been met for the desired number of thresholds, the inquiry at step 255 will be answered in the affirmative, and processing will continue at step 160 of
In accordance with various embodiments of the present invention, the above described push interaction may act as a replacement to touchscreen technology where users are otherwise unwilling to make physical contact with a screen (which can be dirty or otherwise inconvenient) and where the screens and selection choices are preferably limited. Embodiments of the present invention may have numerous applications in conjunction with mobile and other devices including small screens since the object, which may comprise the finger of a user, may be much closer to the camera and hence, determining size changes and tracking the object may be enhanced. Applications may include gaming and other interactive programming. Such a push interaction system as set forth in accordance with embodiments of the present invention allows a webcam system to recognize if an action is deliberate—thus avoiding indicating a selection based upon accidental motion of erroneous object tracking. Similarly, the user may not be required to hold their hand or finger (object) still in space for a relatively long period of time.
In a preferred embodiment of the invention, the object described above may comprise the finger of a user. The system may therefore involve tracking the finger and recognizing if the finger moves closer towards a camera or not. Furthermore, a plurality of regions of interest may be provided, thus allowing for the determination of selection of one choice from among many. Therefore, as is shown in
Implementing the flowchart diagram as set forth in
Thereafter, as is set forth in
Thus, in order to offer a deliberate method to select an onscreen button or other selection, the user will be able to push through space and see a confirmatory change in color of the selection, or other appropriate indicator. Completion of the push may be highlighted by a final color change or other indicator. If the user changes their mind mid-way, the user may reset and stop the motion before the final indicator has been displayed or otherwise invoked. In an alternative embodiment of the invention, the user may be allowed to swipe their hand across the screen to reset. Thus, the system is adapted to link a visual size scale to finger proximity thereby providing feedback to the user that their action/selection is deliberate and registered with the system.
In an alternative embodiment, rather than simply tracking a particular object, such as a user's finger, it may be desirable to track various portions of such an object, thus determining relative movement between these various elements. As is shown in
It has further been determined by the inventors of the present invention that users whose finger or the like is in close proximity to a webcam interacting in a push situation may encounter a parallax problem in that as they move their finger closer to the webcam, in a straight line to the display, if they are not pointing directly at the camera, their finger will appear to be moving off in a direction away from the webcam. This “x,y creep” may be compensated for, and may be considered another method for determining whether a push interaction has been encountered. Thus, if the size of an object, such as the user's finger, remains a constant size, but appears to be moving away from the display, a push interaction may be determined. Such motion will not be confused with motion in the x,y direction, as any such movement will result in a reduction of the size of the object, due to the same parallax issues. Preferably, a combination of such parallax and change in size considerations allow for an intelligent determination of user intent and desired interaction.
In an additional embodiment of the invention, it may be desirable to display a visual image captured by the webcam. When doing so, in accordance with an alternative embodiment of the invention, it may be desirable to provide a depth or proximity indicator to provide real-time feedback to the user regarding the positioning of an object by the user. The system may be adapted to track and display one or more dynamic reference points to provide adequate feedback to the user. Thus, upon positioning an object within a field of view by a user, an indication may be displayed along with the display of the webcam captured image view providing an indication of a target location for the object, or an indication of requested direction of movement, in either the x, y or z axis direction. Such indicator in the x and y axis directions may comprise simple arrows, while movement in the z-axis direction may be indicated by an arrow, an indicator employing sequential different size boxes, circles or the like, these indicators changing appearance as the user approaches the desired z-axis location. Additional audible or other indications may be provided to indicate to the user when movement in the correct direction is made, while ultimately a final indication may be provided when the user has reached the desired location.
Therefore, as is shown by way of example in
Various embodiments in accordance with the method and apparatus of the present invention may be desirable in unstable environments, such as the use of a mobile device, laptop, or other non-fixed computing apparatus or webcam. Thus, while the user is able to select between one or more objects on a display without contact thereof employing only a single webcam, the user may also be guided to placement of a particular object in a particular location, including a desired distance from a webcam. Furthermore, relative motion of that object in the z-axis direction towards and away from the webcam may be determined with this simple hardware setup. Furthermore, as the system may employ various object markers, it is possible to account for rotation of the object, thus determining whether a size change is based upon absolute motion, or if the view of the object has changed, thus potentially causing a perceived change in size of the object.
Various embodiments of the present invention may be employed in a number of areas, including as a replacement to touchscreens, such as at an ATM, in a taxi, gaming, interactive billboards, or in other public kiosk locations. Embodiments of the invention may be employed to track a medication to determine whether a user has placed a pill or other medication in their mouth or the like. The system may first position the pill to determine the identity thereof, track the pill to the mouth of a user, and then determine that the pill has been placed therein. For one or more sales promotions, a user may Interactive Billboards (push interaction, probe requested to hold an object up to a simple webcam, and be guided in it positioning so that information may be obtained from the object, thus allowing for advertising, coupons, or other messaging to be provided to the user.
Additionally, a game may be provided where a user is requested to differentiate between a number of objects, locate these objects in the x, y plane in the field of view of a webcam, and then move the object in a z-axis direction towards or away from the webcam to indicate selection, or other interaction between the object and an on-screen indicator. The various embodiments described above relating to proximity detection and indication may also be employed. Such an object may include a user's finger or other desirable free-standing object or the like.
Other potential embodiments or uses of the embodiments of the invention include:
Dancing Games—User may be encouraged to move forward and backwards and to lift up left or right hand/arm based on proximity. Gains points for complying.
Fighting games—Important to understand position and proximity of the user.
Pingpong games fingerpong—move finger forward to use the paddle or hit the pingpong with your finger.
Drumming games—tap the air, to create a beat in sync with the music.
Air Guitar—recognizes you strumming an air guitar. Strum to the beat to win more points.
It will thus be seen that the objects set forth above, among those made apparent from the preceding description, are efficiently attained and, because certain changes may be made in carrying out the above method and in the construction(s) set forth without departing from the spirit and scope of the invention, it is intended that all matter contained in the above description and shown in the accompanying drawings shall be interpreted as illustrative and not in a limiting sense.
It is also to be understood that this description is intended to cover all of the generic and specific features of the invention herein described and all statements of the scope of the invention which, as a matter of language, might be said to fall there between.
This application is a continuation of, and claims priority to, U.S. application Ser. No. 15/374,334, filed Dec. 9, 2016, now U.S. Pat. No. 10,257,423, issued Apr. 9, 2019, which is a continuation of U.S. application Ser. No. 14/799,869, filed Jul. 15, 2015, now U.S. Pat. No. 9,538,147, issued Jan. 3, 2017, which is a continuation of U.S. application Ser. No. 13/036,210, filed on Feb. 28, 2011, now U.S. Pat. No. 9,116,553, issued Aug. 25, 2015. The contents of all of the prior applications are incorporated herein by reference in their entirety.
Number | Name | Date | Kind |
---|---|---|---|
3814845 | Hurlbrink et al. | Jun 1974 | A |
5065447 | Barnsley et al. | Nov 1991 | A |
5441047 | David et al. | Aug 1995 | A |
5544649 | David et al. | Aug 1996 | A |
5619991 | Sloane | Apr 1997 | A |
5646912 | Cousin | Jul 1997 | A |
5752621 | Passamante | May 1998 | A |
5764296 | Shin | Jun 1998 | A |
5810747 | Brundy et al. | Sep 1998 | A |
5911132 | Sloane | Jun 1999 | A |
5961446 | Beller et al. | Oct 1999 | A |
6151521 | Guo et al. | Nov 2000 | A |
6154558 | Hsieh | Nov 2000 | A |
6233428 | Fryer | May 2001 | B1 |
6283761 | Joao | Sep 2001 | B1 |
6380858 | Yarin et al. | Apr 2002 | B1 |
6409661 | Murphy | Jun 2002 | B1 |
6421650 | Goetz et al. | Jul 2002 | B1 |
6483993 | Misumi et al. | Nov 2002 | B1 |
6484144 | Martin et al. | Nov 2002 | B2 |
6535637 | Wootton et al. | Mar 2003 | B1 |
6611206 | Eshelman et al. | Aug 2003 | B2 |
6705991 | Bardy | Mar 2004 | B2 |
6879970 | Shiffman et al. | Nov 2005 | B2 |
6988075 | Hacker | Jan 2006 | B1 |
7184047 | Crampton | Feb 2007 | B1 |
7184075 | Reiffel | Feb 2007 | B2 |
7256708 | Rosenfeld et al. | Aug 2007 | B2 |
7277752 | Matos | Oct 2007 | B2 |
7304228 | Bryden et al. | Dec 2007 | B2 |
7307543 | Rosenfeld et al. | Dec 2007 | B2 |
7317967 | DiGianfilippo et al. | Jan 2008 | B2 |
7340077 | Gokturk | Mar 2008 | B2 |
7395214 | Shillingburg | Jul 2008 | B2 |
7415447 | Shiffman et al. | Nov 2008 | B2 |
7448544 | Louie et al. | Nov 2008 | B1 |
7562121 | Berisford et al. | Jul 2009 | B2 |
7627142 | Kurzweil et al. | Dec 2009 | B2 |
7657443 | Crass et al. | Feb 2010 | B2 |
7692625 | Morrison et al. | Apr 2010 | B2 |
7747454 | Bartfeld et al. | Jun 2010 | B2 |
7761311 | Clements et al. | Jul 2010 | B2 |
7769465 | Matos | Aug 2010 | B2 |
7774075 | Lin et al. | Aug 2010 | B2 |
7874984 | Elsayed et al. | Jan 2011 | B2 |
7881537 | Ma et al. | Feb 2011 | B2 |
7908155 | Fuerst et al. | Mar 2011 | B2 |
7912733 | Clements et al. | Mar 2011 | B2 |
7956727 | Loncar | Jun 2011 | B2 |
7983933 | Karkanias et al. | Jul 2011 | B2 |
8321284 | Clements et al. | Nov 2012 | B2 |
20010049673 | Dulong et al. | Dec 2001 | A1 |
20010056358 | Dulong et al. | Dec 2001 | A1 |
20020026330 | Klein | Feb 2002 | A1 |
20020093429 | Matsushita et al. | Jul 2002 | A1 |
20020143563 | Hufford et al. | Oct 2002 | A1 |
20030164172 | Chumas et al. | Sep 2003 | A1 |
20030190076 | Delean | Oct 2003 | A1 |
20030225325 | Kagermeier et al. | Dec 2003 | A1 |
20040100572 | Kim | May 2004 | A1 |
20040107116 | Brown | Jun 2004 | A1 |
20040155780 | Rapchak | Aug 2004 | A1 |
20050144150 | Ramamurthy et al. | Jun 2005 | A1 |
20050149361 | Saus et al. | Jul 2005 | A1 |
20050180610 | Kato et al. | Aug 2005 | A1 |
20050182664 | Abraham-Fuchs et al. | Aug 2005 | A1 |
20050234381 | Niemetz et al. | Oct 2005 | A1 |
20050267356 | Ramasubramanian et al. | Dec 2005 | A1 |
20060066584 | Barkan | Mar 2006 | A1 |
20060218011 | Walker et al. | Sep 2006 | A1 |
20060238549 | Marks | Oct 2006 | A1 |
20070030363 | Cheatle et al. | Feb 2007 | A1 |
20070118389 | Shipon | May 2007 | A1 |
20070194034 | Vasiadis | Aug 2007 | A1 |
20070233035 | Wehba et al. | Oct 2007 | A1 |
20070233049 | Wehba et al. | Oct 2007 | A1 |
20070233050 | Wehba et al. | Oct 2007 | A1 |
20070233281 | Wehba et al. | Oct 2007 | A1 |
20070233520 | Wehba et al. | Oct 2007 | A1 |
20070233521 | Wehba et al. | Oct 2007 | A1 |
20070008112 | Covannon et al. | Nov 2007 | A1 |
20070008113 | Spoonhower et al. | Nov 2007 | A1 |
20070265880 | Bartfeld et al. | Nov 2007 | A1 |
20070273504 | Tran | Nov 2007 | A1 |
20080000979 | Poisner | Jan 2008 | A1 |
20080093447 | Johnson et al. | Apr 2008 | A1 |
20080114226 | Music et al. | May 2008 | A1 |
20080114490 | Jean-Pierre | May 2008 | A1 |
20080138604 | Kenney et al. | Jun 2008 | A1 |
20080140444 | Karkanias et al. | Jun 2008 | A1 |
20080162192 | Vonk et al. | Jul 2008 | A1 |
20080178126 | Beeck et al. | Jul 2008 | A1 |
20080201174 | Ramasubramanian et al. | Aug 2008 | A1 |
20080219493 | Tadmor | Sep 2008 | A1 |
20080275738 | Shillingburg | Nov 2008 | A1 |
20080290168 | Sullivan et al. | Nov 2008 | A1 |
20080297589 | Kurtz et al. | Dec 2008 | A1 |
20080303638 | Nguyen et al. | Dec 2008 | A1 |
20090012818 | Rodgers | Jan 2009 | A1 |
20090018867 | Reiner | Jan 2009 | A1 |
20090043610 | Nadas et al. | Feb 2009 | A1 |
20090048871 | Skomra | Feb 2009 | A1 |
20090095837 | Lindgren | Apr 2009 | A1 |
20090127339 | Needhan | May 2009 | A1 |
20090128330 | Monroe | May 2009 | A1 |
20090159714 | Coyne, III et al. | Jun 2009 | A1 |
20090217194 | Martin et al. | Aug 2009 | A1 |
20090245655 | Matsuzaka | Oct 2009 | A1 |
20100042430 | Bartfeld | Feb 2010 | A1 |
20100050134 | Clarkson | Feb 2010 | A1 |
20100057646 | Martin et al. | Mar 2010 | A1 |
20100092093 | Akatsuka et al. | Apr 2010 | A1 |
20100136509 | Mejer et al. | Jun 2010 | A1 |
20100138154 | Kon | Jun 2010 | A1 |
20100255598 | Melker | Oct 2010 | A1 |
20100262436 | Chen et al. | Oct 2010 | A1 |
20100316979 | Von Bismarck | Dec 2010 | A1 |
20110021952 | Vallone | Jan 2011 | A1 |
20110119073 | Hanina et al. | May 2011 | A1 |
20110141009 | Izumi | Jun 2011 | A1 |
20110153360 | Hanina et al. | Jun 2011 | A1 |
20110161109 | Pinsonneault et al. | Jun 2011 | A1 |
20110195520 | Leider et al. | Aug 2011 | A1 |
20110275051 | Hanina et al. | Nov 2011 | A1 |
20120075464 | Derenne et al. | Mar 2012 | A1 |
Entry |
---|
Ammouri, S.; Biloduau, G. -A, “Face and Hands Detection and Tracking Applied to the Monitoring of Medication Intake,” Computer and Robot Vision, 2008. CRV '08. Canadian Conference, 147(154):May 28-30, 2008. |
Batz, et al. “A computer Vision System for Monitoring Medication Intake,” in Proc. IEEE 2nd Canadian Conf. on Computer and Robot Vision, Victoria, BC, Canada, 2005, pp. 362-369. |
Bilodeau et al. Monitoring of Medication Intake Using a Camera System. Journal of Medical Systems 2011. [retrieved on Feb. 18, 2013] Retrieved from ProQuest Technology Collection. |
Chen, Pauline W., “Texting as a Health Tool for Teenagers”, The New York Times, Nov. 5, 2009, http://www.nytimes.com/2009/11/05/health/0512/899,510. |
Danya International, Inc., “Pilot Study Using Cell Phones for Mobile Direct Observation Treatment to Monitor Medication Compliance of TB Patients”, Mar. 20, 2009, www.danya.com/MDOT.asp. |
Global Tuberculosis Control: A short update to the 2009 report, World Health Organization, (2009). |
Huynh et al., “Real time detection, tracking and recognition of medication intake.” World Academy of Science, Engineering and Technology 60 (2009), 280-287. |
Mintchell, “Exploring the Limits of Machine Vision”, Automating World, Oct. 1, 2011. |
Osterberg, Lars and Blaschke, Terrence, “Adherence to Medication”, New England Journal of Medicine 2005; 353:487-97, Aug. 4, 2005. |
Super-Resolution, Wikipedia, (Oct. 5, 2010). |
University of Texas, GuideView, Mar. 15, 2007, http://www.sahs.uth.tmc.edu/MSriram/GuideView. |
Valin, et al. “Video Surveillance of Medication intake”, Int. Conf. of the IEEE Engineering in Medicine and Biology Society, New York City, USA, Aug. 2006. |
Wang et al. “Recent Developments in human motion analysis.” Pattern Recognition 36 (220) 585-601 (Nov. 2001). |
Whitecup, Morris S., “2008 Patient Adherence Update: New Approaches for Success”, www.guideline.com, The Trend Report Series, (Oct. 1, 2008). |
Final Office Action from PTO, Cited in AI-0001-U1 (U.S. Appl. No. 12/620,686), (dated May 8, 2012), 1-24. |
Final Office Action from PTO, Cited in AI-0001-U2 (U.S. Appl. No. 13/558,377), dated May 7, 2013, 1-29. |
Final Office Action from PTO, Cited in AI-0002-U1 (U.S. Appl. No. 12/646,383), (dated May 8, 2012), 1-31. |
Final Office Action from PTO, Cited in AI-0002-U2 (U.S. Appl. No. 13/588,380), (dated Mar. 1, 2013), 1-27. |
Final Office Action from PTO, Cited in AI-0003-U1 (U.S. Appl. No. 12/646,603), (dated Feb. 1, 2012), 1-17. |
Final Office Action from PTO, Cited in AI-0004-U1 (U.S. Appl. No. 12/728,721), (dated Apr. 12, 2012), 1-31. |
Final Office Action from PTO, Cited in AI-0005-U1 (U.S. Appl. No. 12/815,037), (dated Sep. 13, 2012), 1-15. |
Final Office Action from PTO, Cited in AI-0006-U1 (U.S. Appl. No. 12/899,510), (dated Aug. 20, 2013). |
Final Office Action from PTO, Cited in AI-0008-U1 (U.S. Appl. No. 12/898,338), dated Nov. 9, 2012), 1-12. |
Final Office Action from PTO, Cited in AI-0012-U1 (U.S. Appl. No. 13/189,518), (dated Jul. 23, 2013), 1-16. |
International Preliminary Report on Patentability, cited in AI-0001-PCT1 (PCT/US2010/056935) (dated May 31, 2012), 1-8. |
Non-Final Office Action from PTO, Cited in AI-0001-U1 (U.S. Appl. No. 12/620,686), (dated Dec. 21, 2011),1-78. |
Non-Final Office Action from PTO, Cited in AI-0001-U2 (U.S. Appl. No. 13/558,377), (dated Oct. 22, 2012), 1-21. |
Non-Final Office Action from PTO, Cited in AI-0002-U1 (U.S. Appl. No. 12/646,383), (dated Dec. 22, 2011),1-78. |
Non-Final Office Action from PTO, Cited in AI-0002-U2 (U.S. Appl. No. 13/558,380), (dated Oct. 4, 2012), 1-20. |
Non-Final Office Action from PTO, Cited in AI-0003-U1 (U.S. Appl. No. 12/646,603), (dated Oct. 13, 2011),1-74. |
Non-Final Office Action from PTO, Cited in AI-0003-U1 (U.S. Appl. No. 12/646,603), (dated Jun. 13, 2013), 1-16. |
Non-Final Office Action from PTO, Cited in AI-0004-U1 (U.S. Appl. No. 12/728,721), (dated Jan. 6, 2012), 1-31. |
Non-Final Office Action from PTO, Cited in AI-0004-U1 (U.S. Appl. No. 12/728,721), (dated May 9, 2013), 1-25. |
Non-Final Office Action from PTO, Cited in AI-0005-U1 (U.S. Appl. No. 12/815,037), (dated Mar. 28, 2012),1-17. |
Non-Final Office Action from PTO, Cited in AI-0005-U1 (U.S. Appl. No. 12/815,037), (dated Jul. 18, 2013), 1-19. |
Non-Final Office Action from PTO, Cited in AI-0006-U1 (U.S. Appl. No. 12/899,510), (dated Jan. 23, 2013), 1-20. |
Non-Final Office Action from PTO, Cited in AI-0008-U1 (U.S. Appl. No. 12/898,338), (dated Jun. 19, 2012), 1-16. |
Non-Final Office Action from PTO, Cited in AI-0012-U1 (U.S. Appl. No. 13/189,518), (dated Dec. 21, 2012), 1-10. |
Non-Final Office Action from PTO, Cited in AI-0013-U1 (U.S. Appl. No. 13/235,387), dated Sep. 12, 2013), 1-16. |
PCT Search report and written opinion, Cited in AI-0001-PCT1 (PCT/US2010/56935, (dated Dec. 1, 2011),1-9. |
PCT Search report and written opinion, Cited in AI-0005-PCT1 (PCT/US2011/35093, (dated Dec. 9, 2011),1-8. |
PCT Search report and written opinion, Cited in AI-0006-PCT1 (PCT/US11/54666), (dated Feb. 28, 2012), 1-13. |
PCT Search report and written opinion, Cited in AI-0008-PCT1 (PCT/US11/54668), dated Feb. 28, 2012, 1-12. |
PCT Search report and written opinion, Cited in AI-0012-PCT1 (PCT/US12/41785), (dated Aug. 17, 2012),1-10. |
PCT Search report and written opinion, Cited in AI-0013-PCT1 (PCT/US12/42843), (dated Aug. 31, 2012), 1-8. |
PCT Search report and written opinion, Cited in AI-0018-PCT1 (PCT/US2012/051554), (dated Oct. 19, 2012), 1-12. |
PCT Search report and written opinion, Cited in AI-0019-PCT (PCT/US12/59139), (dated Dec. 18, 2012), 1-15. |
PCT Search report and written Opinion, Cited in AI-0020-PCT1 (PCT/US13/20026), (dated May 8, 2013), 1-14. |
Number | Date | Country | |
---|---|---|---|
20190238755 A1 | Aug 2019 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 15374334 | Dec 2016 | US |
Child | 16376195 | US | |
Parent | 14799869 | Jul 2015 | US |
Child | 15374334 | US | |
Parent | 13036210 | Feb 2011 | US |
Child | 14799869 | US |