Aiming feedback system with inertial sensors

Information

  • Patent Grant
  • 10265602
  • Patent Number
    10,265,602
  • Date Filed
    Thursday, March 3, 2016
    8 years ago
  • Date Issued
    Tuesday, April 23, 2019
    5 years ago
Abstract
An aiming system that provides feedback on how closely the aim of an object is aligned with a direction to a target. An inertial sensor on the object provides data on the object's position and orientation; this data is combined with target direction information to determine how to correct the aim. An illustrative application is a golf club aiming system that measures whether the clubface normal is aligned horizontally with the direction to the hole. The system sends feedback signals to the user to help the user adjust the aim. These signals may include for example audible tones or haptic vibrations that vary in frequency and amplitude to instruct the user to adjust the aim. For example, haptic signals may be sent to a smart watch worn by the user; the user may therefore obtain aiming feedback without having to look at a screen.
Description
BACKGROUND OF THE INVENTION

Field of the Invention


One or more embodiments of the invention are related to the field of motion capture data analysis using sensor data. More particularly, but not by way of limitation, one or more embodiments of the invention enable an aiming feedback system that uses inertial sensor data to measure the direction an object is aimed and to provide feedback to improve or correct the aim.


Description of the Related Art


Systems to measure the position and orientation of an object using inertial sensors are known in the art. For example, inertial guidance systems are widely used in navigation. The increasing availability of low-cost sensors has led to consumer applications that track position, orientation, and motion of users and of equipment. For example, inertial sensors to measure the position, orientation, and motion of a golf club are known in the art.


In many applications, the orientation of an object relative to a target position is of major importance. For example, in golf, a golfer must align the golf club so that the face of the club is pointing towards the hole. While an inertial sensor on the club may provide information about the club's orientation, this sensor alone cannot determine whether the club is aligned correctly relative to the target. It is therefore necessary to supplement inertial sensor information with other data to determine alignment to a target. In addition, it is desirable to have a convenient feedback mechanism to provide feedback to a user so that the user can adjust the aim to align it with the target. There are no known systems based on inertial sensors that provide aiming feedback to a user to assist the user in aligning the aim of an object to a target.


For at least the limitations described above there is a need for an aiming feedback system with inertial sensors.


BRIEF SUMMARY OF THE INVENTION

One or more embodiments described in the specification are related to an aiming feedback system with inertial sensors. Embodiments of the invention use inertial sensor data, possibly supplemented with other information, to determine the relative orientation of an object to a target direction. Feedback on the alignment of the object to the target direction is provided to the user aiming the object. Illustrative applications include for example aiming of sporting equipment, such as clubs, bats, and rackets. In golf, in particular, a golfer may need to align the aim of the clubface with the direction to the hole; embodiments of the system may for example use inertial sensors on the club to determine alignment to the hole and to help the golfer adjust the aim accordingly.


One or more embodiments of the invention enable an aiming feedback system with inertial sensors. Inertial sensors are coupled to an object that may be aimed at a target. The object may have for example a forward direction vector that may depend on the geometry of the object; for example, the forward direction vector for a golf club may be the normal vector to the clubface. Inertial sensor data from the object is transmitted to a processor, which uses this data to determine the orientation of the object's forward direction vector. The processor also receives, obtains, or calculates a target direction vector that indicates the direction to the target. For example, in a golf application the target may be a golf hole, and the target direction vector may point from the club towards the hole. The processor calculates a rotation between the object's forward direction vector and the target direction vector, and it generates an aim alignment signal that provides information on how to align the forward direction vector with the target direction vector.


The aim alignment signal is transmitted to one or more feedback elements that generate feedback signals for a user to assist the user in correcting the aim. One or more embodiments may use feedback elements that include for example, without limitation, displays, optical devices, speakers, and haptic feedback systems. Display feedback elements may for example display text or graphics that indicate the current aim direction, the target direction, the rotation between them, or how to modify the aim to point at the target. Speaker feedback elements may for example generate audio signals that vary with the aim, such as for example a tone or beat that varies as the aim gets closer to the target direction. Speakers may also generate spoken messages that tell the user how to rotate the aim to align it with the target direction; these messages may for example include both a desired direction of rotation (such as left or right) and an angular magnitude. Haptic feedback systems may provide for example taps or vibrations that vary with the aim, or that encode instructions on how to modify the aim to point at the target. Haptic feedback may be integrated for example into a smart watch or a wrist band worn by the user aiming the object. Feedback signals, such as for example audio or haptic signals, may have frequencies or amplitudes that vary with the desired rotation to align the aim with the target direction. For example, the amplitude or frequency of the signal may vary monotonically with the angular magnitude of the rotation between the target direction and the current aim. In addition, the amplitude or frequency may vary to indicate the direction of the desired rotation to align the aim with the target direction.


In some applications the user's goal may be align the projections of the forward direction vector and the target direction vector onto a plane, such as the horizontal plane, rather than to align the vectors themselves. For example, in golf, the forward direction vector for a club typically points upwards because of the loft of the club face; however, the golfer wants to align the horizontal component of this vector with the horizontal vector pointing towards the hole. Therefore, in one or more embodiments the processor may project one or both of the forward direction vector and the target direction vector onto a plane before calculating the rotation between them.


One or more embodiments may use any convenient method to obtain, receive, or calculate the target direction vector. In particular, one or more embodiments may use the object itself to determine the direction to the target. An object may have a designated targeting axis that may be used to point at the target. For example, a user may point the shaft of a golf club at the hole to determine the direction to the hole. When the object's targeting axis is pointed at the target, the inertial sensor data received from the object may be used to calculate the target direction vector. One or more embodiments may include a laser pointer aligned with the targeting axis to assist the user in pointing the targeting axis at the target. The object may then be placed in its normal use orientation and aimed by the user. The system obtains inertial sensor data from the object again, while it is in the normal use orientation, and calculates the forward direction vector that represents the current aim. In one or more embodiments the inertial sensor on the object may also be used to determine the slope of the ground; this information may be useful for example for a golfer who adjusts the swing based on the slope. The object may for example be placed on the ground with its targeting axis pointing horizontally towards the target; the inertial sensor data received in this orientation may be used to calculate the slope of the ground. The system may then provide slope information to the user along with feedback information on the aim relative to the target.





BRIEF DESCRIPTION OF THE DRAWINGS

The above and other aspects, features and advantages of the invention will be more apparent from the following more particular description thereof, presented in conjunction with the following drawings wherein:



FIG. 1 illustrates an embodiment of the system that provides aim correction feedback to a golfer who aims a golf club at a golf hole.



FIG. 2 illustrates forward direction vectors associated with a golf club and a tennis racket, corresponding to the direction in which a ball will be launched when hit, and shows illustrative reference axes of inertial sensors attached to the club and the racket.



FIG. 3 shows an architectural block diagram of the components of the system illustrated in FIG. 1.



FIG. 4 illustrates an embodiment that provides haptic aiming feedback using a smart watch with a haptic subsystem.



FIG. 5 illustrates an embodiment that encodes aim correction magnitude and direction in the amplitude and frequency, respectively, of a feedback signal generated by the system.



FIG. 6 illustrates an embodiment that projects the forward direction vector, in this case the normal to a golf clubface, onto the horizontal plane, and compares this horizontal projection to the direction to the target.



FIG. 7 shows an embodiment that uses the shaft of a golf club to determine the direction to the target.



FIG. 8 extends the example of FIG. 7 with an embodiment that has a laser pointer attached to or embedded in the shaft of the golf club, to assist in pointing at the target.



FIG. 9 shows an embodiment that provides feedback on the slope the ground when the shaft of a golf club is placed on the ground.





DETAILED DESCRIPTION OF THE INVENTION

An aiming feedback system with inertial sensors will now be described. In the following exemplary description numerous specific details are set forth in order to provide a more thorough understanding of embodiments of the invention. It will be apparent, however, to an artisan of ordinary skill that the present invention may be practiced without incorporating all aspects of the specific details described herein. In other instances, specific features, quantities, or measurements well known to those of ordinary skill in the art have not been described in detail so as not to obscure the invention. Readers should note that although examples of the invention are set forth herein, the claims, and the full scope of any equivalents, are what define the metes and bounds of the invention.



FIG. 1 illustrates an embodiment of the system that provides aiming feedback to a golfer. This example is illustrative; one or more embodiments may provide aiming feedback for any object, including but not limited to objects used for sports. In the embodiment shown, user 101 wants to aim the golf club 102 so that the ball travels in the target direction 110 towards the hole 103. An inertial sensor 104 is attached to the golf club 102, providing data on the position and orientation of the club. This inertial sensor data is transmitted to mobile device 120. In this example, the inertial sensor 104 contains a wireless transmitter to transmit data to mobile device 120. One or more embodiments may use any methods to transmit sensor data to any other device, over wired or wireless networks or combinations thereof. Device 120 calculates the current aim direction 111 of the golf club 102 from the inertial sensor data generated by sensor 104. Device 120 also receives, generates, calculates, or estimates the target direction vector 110. Embodiments may use various methods to determine, receive, calculate, estimate, or obtain this target direction vector. For example, in some situations the location of the target 103 may be known in advance, in which case knowledge of the position of the object to be aimed (such as golf club 102) is sufficient to calculate the target direction vector 110. In other situations, sensors may be used to determine the direction to the target. As described below, in one or more embodiments the object to be aimed may also serve as a sensor to acquire the target direction vector 110.


Device 120 determines the rotation 112 between the desired aim (target direction vector 110) and the actual aim (forward direction vector 111). It uses this rotation to provide feedback to the user 101 on how closely the aim is aligned with the target, or on specific corrections to the aim needed to align it with the target. In the example of FIG. 1, device 120 displays message 121 on the screen providing specific aim correction instructions. The device may also provide audio feedback, haptic feedback, or more generally feedback using any type of signal or signals. Audio or haptic feedback signals provide the potential benefit that the user 101 does not need to look away from the object he or she is aiming to receive the feedback. In FIG. 1, device 120 also provides audio feedback. Two types of audio feedback are shown: the device may provide a voice command 123, similar to the display message 121; the device may also or in addition provide signals such as tones or beeps 122 that indicate how close the aim is to the target direction. In the example shown, the beeps are infrequent when the aim is far away from the target direction, and they grow more frequent as the aim approaches the target direction. This scheme is illustrative; one or more embodiments may use any feedback signals to compare the aim to the target direction or to suggest aim changes.


As the user changes the aim of the golf club 102, the device 120 changes the feedback signals to reflect the updated aim. The user can therefore continue adjusting the aim until it is aligned with the target. In the right side of FIG. 1, the user has shifted the aim of the golf club so that the new forward direction vector 131 is closer to the target direction vector 110. The inertial sensor sends updated sensor data to the device 120, and the device calculates the updated rotation 132 between the vectors 131 and 110. The device then updates the feedback signals 141, 143, and 142. In particular, the audio feedback 142 increases the frequency of the beeps, telling the golfer that the aim is closer to the target. In one or more embodiments the frequent beeps may be replaced for example with a continuous tone when the aim is fully aligned with the target (or within a specified tolerance of the target direction vector). In one or more embodiments beeps or other signals may instead decrease in frequency as the aim approaches the target direction, for example with the signal becoming silent when the aim is perfectly aligned with the target.


As mentioned above, one or more embodiments of the invention may be used to provide feedback for the aiming of any object. FIG. 2 shows two objects that are equipped for aiming feedback: golf club 102 (as in FIG. 1), and tennis racket 210. In each case the object has a forward direction vector associated with the geometry of the object. For golf club 102, forward direction vector 202 is the normal vector to the clubface 201. For tennis racket 210, forward direction vector 212 is the normal vector to the racket face 211. Any vector may be selected as a forward direction vector for an object. The forward direction vector may in some cases correspond to the direction that a ball is launched, but in general it may be any desired direction associated with an object. In one or more embodiments there may be multiple forward direction vectors associated with the object, and a specific forward direction vector may be selected for aiming feedback based on context or on other data. For example, racket 210 may be used in either forehand or backhand direction, with opposing forward direction vectors.


In one or more embodiments one or more inertial sensors or other types of sensors may be coupled to or embedded in an object to be aimed. For example, inertial sensor 104 is attached to golf club 102, and inertial sensor 213 is attached to racket 210. Inertial sensors may be attached to any location of an object and in any orientation. In one or more embodiments inertial sensors or other sensors may be attached to an object at multiple different locations. To support aiming feedback, it is generally necessary to obtain or determine the relationship between the position and orientation of the inertial sensor or sensors and that of the forward direction vector for the object. Presuming that the object being aimed is a rigid body or an approximately rigid body, there is a fixed relationship between the position and orientation of the inertial sensors on the object and the position and orientation of the forward direction vector. This relationship may be determined for example via calibration methods when the inertial sensor is installed onto the object. In general, there is a transformation between the reference frame of the inertial sensor and the reference frame associated with the forward direction vector. For example, for golf club 102, the sensor reference frame 204 may be at or near the grip of the golf club, while the forward direction reference frame 203 may be associated with the clubface. As is known in the art, a transformation comprising a rotation and a translation may be performed to transform the sensor reference frame into the forward direction reference frame, or vice versa. The transformation is a characteristic of the object's geometry, the selected forward direction vector, and the installed location and orientation of the sensors.



FIG. 3 shows an architectural block diagram of the components of the embodiment shown in FIG. 1. Inertial sensor module 104 may contain any types of sensors. For illustration, FIG. 3 shows possible sensors that include a 3 axis accelerometer 310, a 3 axis gyroscope 311, a 3 axis magnetometer 312, and a GPS 313. Any or all of these sensors are optional. Any or all of these sensors may have any number of axes, including but not limited to three axes. One or more embodiments may use additional sensors that are not illustrated. Any sensor or sensors that provide information on the position or orientation of the object, or on any changes to position or orientation, may be used by one or more embodiments.


Data 315 from the inertial sensors (or other sensors) 310, 311, 312, 313 are transmitted by transmitter 314 to a receiver 321, which is coupled to a processor 330 that analyzes the sensor data. The connection between the inertial sensor module transmitter 314 and the receiver 321 may be wired or wireless, using any desired network protocols and media. In one or more embodiments the processor that analyzes the sensor data may be integrated into the same physical device as the inertial sensors. In one or more embodiments the processor may be remote from the inertial sensors.


Device 120 in FIG. 1 is a mobile device that contains a processor, a network receiver, and feedback devices such as a display and speakers. In one or more embodiments these functions may be separated into different devices. For example, a processor that analyzes sensor data may be in a different device than the device that generates feedback signals. In one or more embodiments the inertial sensor or sensors, the processor, and the feedback elements may all be integrated into a single device. For illustration, in FIG. 3 device 120 contains multiple components: a receiver 321 that receives inertial sensor data 315, a receiver 320 that receives the target direction vector 110, a processor 330 that analyzes the data, and feedback elements 341, 342, and 343. The receivers 321 and 320 may be identical elements or they may be different. In one or more embodiments the processor 330 may be a single processor, a multiprocessor, a network of processors, or a specialized circuit. In one or more embodiments the processor or processors may be for example a microprocessor, a mobile device, a personal computer, a laptop computer, a desktop computer, a tablet computer, a smart phone, a smart watch, a server, or a network of any of these devices.


Target Direction Vector 110 is obtained via process 301, and is transmitted to receiver 320. Transmission of target direction vector 110 may be over a wired or wireless network, using any desired protocols and media. One or more embodiments may use any methods, hardware, and software to implement the Measure Target Direction process 301. For example, in one or more embodiments the target direction to a target may be obtained from a database containing the location of known targets. For instance, in a golf application, targets at a golf course or a driving range may be in fixed, known locations, and device 120 may obtain these locations directly from a database when the golfer is at that site. In one or more embodiments the user may use one or more sensors to perform the step 301. In one or more embodiments the targets themselves may be instrumented with sensors, and the step 301 may comprise retrieving this target sensor data to obtain the target locations.


Processor 330 analyzes the inertial sensor data 315 and the target direction vector 110 to determine how closely the object aim is aligned with the target direction. It calculates the forward direction 331 of the object from the inertial sensor data 315 to determine the direction in which the object is currently aimed. The specific calculations of this step 331 depend on the types of sensors included in sensors 104. For example, if sensors 104 include accelerometer 310 and magnetometer 312, and if the object 102 is not accelerating, then the orientation of the object relative to an Earth reference frame may be calculated directly, as is known in the art. (The accelerometer provides the orientation of the sensors relative to the gravity vector, and the magnetometer provides the orientation of the sensors relative to the magnetic field vector.) If for example sensors 104 include accelerometer 310 and gyro 311, then the inertial sensor data 315 may be integrated to obtain the orientation of the sensors relative to any fixed reference frame, provided that an initial orientation is known, as is known in the art. Having obtained the orientation of the sensors, the orientation of the object's forward direction vector may be calculated directly using the fixed transformation between the sensor reference frame and the forward direction reference frame, as described above.


The processor then calculates the rotation 332 between the object's forward direction vector and the target direction vector 110. The angular magnitude of this rotation determines how closely aligned the object's aim is with the target direction. The axis of rotation determines the direction in which the object's aim should be rotated to align the aim with the target direction. Based on the rotation, the processor calculates one or more aim alignment signals 333 that describe the rotation or that provide information on how to align the aim with the target. The processor then transmits these aim alignment signals to one or more feedback elements that generate feedback signals to the user. FIG. 3 shows three feedback elements: screen 341, haptic feedback system 342, and speakers 343. These feedback elements are illustrative; one or more embodiments may use any device or devices to generate feedback. The feedback element or elements generate feedback signals 350 (such as for example sound or vibrations) that are transmitted to the user to assist the user in correcting the aim.



FIG. 4 illustrates an embodiment that includes a haptic feedback system. As in FIG. 1, device 120 receives inertial sensor data from golf club 102 and calculates the alignment between the club's forward direction vector and the target direction. Device 120 displays an aim correction message on the screen. In addition, device 120 transmits a feedback signal 402 to a smart watch 401 worn by the user. The watch 401 has a haptic feedback system that generates taps and vibrations 403 that the user can feel. The watch 401 also displays a message 404 with aim correction information like that shown on device 120. Table 405 shows illustrative haptic feedback signals that indicate the magnitude and direction of the aim corrections needed to align the golf club's aim with the target. These signals are illustrative; one or more embodiments may encode aim correction or aim alignment information into any desired set of haptic feedback signals.


In the embodiment shown in FIG. 4, the processor receiving inertial sensor data is embedded in mobile device 120, which then transmits feedback information to smart watch 401. In one or more embodiments the smart watch 401 may instead (or in addition) receive inertial sensor data directly, and calculate the aim alignment information without requiring the mobile device 120. More generally in one or more embodiments the functions of receiving sensor data, calculating aim alignment, and providing feedback may be distributed across physical devices in any desired manner. In one or more embodiments feedback signals may be provided by multiple devices simultaneously; for example, a user may obtain haptic feedback from a smart watch and audio feedback from a mobile phone at the same time.


One or more embodiments may use any desired functions to encode aim alignment information into feedback signals. FIG. 5 illustrates an embodiment that encodes the magnitude and direction of aim corrections into the amplitude and frequency, respectively, of a feedback signal. The feedback signal may be, for example, an audio signal, a vibration or tapping haptic signal, or a flashing light signal. The signal may vary in magnitude and frequency to indicate the required aim corrections. In the example of FIG. 5, the amplitude 501 increases monotonically with the magnitude of the aim correction. In this example a higher magnitude indicates a larger aim correction. One or more embodiments may employ the opposite convention of associating a larger signal amplitude with a smaller aim correction. The frequency 502 in this example has two different values that indicate whether the aim needs to be corrected to the left or to the right. In this example a high frequency signal indicates that the aim should be corrected to the right; a low frequency signal indicates that the aim should be corrected to the left. One or more embodiments may use other encoding schemes to indicate an arbitrary axis around which the forward direction vector should be rotated to bring it into alignment with the target direction vector. FIG. 5 shows examples of the feedback signal scheme defined by 501 and 502: signal 503 is a high frequency, lower amplitude signal indicating an adjustment to the right by 5 degrees; signal 504 is a low frequency, higher amplitude signal indicating an adjustment to the left by 10 degrees.


In one or more embodiments an aim adjustment may be based for example on the components or projections of the forward direction vector and the target direction vector onto a plane or a line, rather than on the original vectors themselves. FIG. 6 illustrates such an embodiment for a golf application. In golf the forward direction vector of a golf club, which is for example normal to the clubface, typically has an upward vertical component as well as a horizontal component since the club often is designed to launch the ball into the air. However, the aim adjustment to aim towards the hole may primarily involve adjusting the horizontal component of the aim. In FIG. 6, club 102 has a forward direction vector (normal to the clubface, for example) 202, which sends ball 601 into the air. The projection of this vector onto the horizontal ground plane 602 is vector 603. The aim adjustment may be based for example on angle 604 between the horizontal forward direction vector 603 and the target direction vector 110. In this example the target direction vector 110 is assumed to lie on the plane 602. In one or more embodiments the target direction vector may also be projected onto a plane to determine an aim correction.


We now describe a method for obtaining the target direction vector using the same object whose aim is adjusted towards the target. In one or more embodiments the object may have targeting axis that can be pointed at the target. This targeting axis may be different from the forward direction vector of the object. FIG. 7 illustrates an example using a golf application. Golf club 102 has forward direction vector 202, which is normal to the clubface. The shaft of the club defines a targeting axis 701. When a user points the shaft at the target, the system can obtain the target direction vector using the inertial sensor attached to the club. In FIG. 7, the z axis 702 of the inertial sensor 104 is aligned with the shaft axis 701; therefore, the orientation of this z-axis provides the target direction vector. This configuration is illustrative; one or more embodiments may use any position and orientation of an inertial sensor and translate from this position and orientation to the targeting axis in order to obtain the target direction vector. As described above, the specific calculations required to obtain the target direction vector from the inertial sensor data may depend on the types of sensors available. For example, if the sensors 104 include a 3 axis accelerometer and a 3 axis magnetometer, the orientation of the inertial sensor relative to the Earth reference frame may be obtained directly; hence the target direction vector may be calculated directly using the known relationship between the sensor reference frame and the targeting axis.


The right side of FIG. 7 illustrates a targeting procedure that may be used in one or more embodiments. The user places the golf club 102 on the ground in orientation 710, with the shaft pointing towards the target 103. The system obtains inertial sensor data while the club is in this orientation, and calculates the target direction vector using the orientation of the z-axis 702, which is aligned (in this example) with the targeting axis. The user then rotates (and possibly translates) the club to orientation 711 to make the shot. By tracking and integrating the inertial sensor data during this motion, the system can calculate the transformation 712 between the targeting orientation 710 and the aiming orientation 711. Thus the orientation of the forward direction vector 202 relative to the target direction can be calculated. Specifically, in the sensor reference frame 204 in orientation 710, the target direction vector 110 is aligned with the z axis 702; thus the vector 110 in this frame is k (the unit vector in the z direction). If we represent the transformation 712 between 710 and 711 as Tta (ta=targeting to aiming), and the transformation 703 between the sensor reference frame 204 and the clubface reference frame 203 as Tsn (sn=sensor to normal), then the target direction vector relative to the clubface frame 203 is TsnTtak. The forward direction vector in the clubface frame is simply k. Thus the rotation between the forward direction vector 202 and the target direction vector 110 is the rotation that rotates k into TsnTtak, which can be calculated directly.


In one or more embodiments a laser pointer or other pointing device may be attached to the object and aligned with the object's targeting axis, to assist the user in pointing at the target. FIG. 8 illustrates an example with a laser pointer 801 installed in the grip of the golf club, so that a laser beam 802 is emitted along the targeting axis 702. The user may place the club in orientation 710 so that the laser beam 802 hits at or near the target, for example on the flag at location 804. The target direction calculation procedure described above may then be performed to acquire the target direction vector.


One or more embodiments may also use the procedures described in FIGS. 7 and 8 to determine the slope of the ground in an area where the object is being aimed, or in any other area where a slope reading is desired. This information may be useful for example in golf, since the slope may affect how the golfer hits a shot. FIG. 9 illustrates an example with club 102 placed on a sloped region 901, such that the targeting vector 702 (aligned with the shaft of the club) points towards the target horizontally, but points upwards because of the slope. Inertial sensor data from sensor 104 can provide a measurement of the slope angle 902, for example by using accelerometer values on three axes to determine the orientation of sensor 104 relative to the vertical gravity vector. Specifically, presuming that the z-axis of the accelerometer is aligned with targeting axis 702, the angle 902 may be calculated as θ=a sin (αz/√{square root over (αx2y2z2)}), where the accelerometer vector is (αx, αy, αt). This information may be displayed for example as message 903 on mobile device 120. Alternatively, or in addition, any of the feedback devices described above may also generate feedback signals that depend on the slope.


While the invention herein disclosed has been described by means of specific embodiments and applications thereof, numerous modifications and variations could be made thereto by those skilled in the art without departing from the scope of the invention set forth in the claims.

Claims
  • 1. An aiming feedback system with inertial sensors, comprising: an inertial sensor configured to be attached to a location on or embedded within an object that is aimed at a target, wherein said object comprises a forward direction vector that defines a current aim direction in which said object is currently aimed, wherein said forward direction vector depends on a geometry of said object;wherein said object is a piece of sporting equipment with a shaft; and,wherein said object further comprises a targeting axis parallel to said shaft, wherein said targeting axis is used to orient said object at said target;a feedback element configured to receive an aim alignment signal; and,provide a feedback signal based on said aim alignment signal to a user who uses said object while said object is aimed at said target and while said targeting axis is pointed at said target, such that when said user changes said current aim of said object, said feedback element changes said feedback signal as said user changes said current aim of the object to reflect an updated aim; and,a processor configured to receive an initial inertial sensor data from said inertial sensor, wherein said initial inertial sensor data is captured when said object is oriented such that said targeting axis is oriented at said target;calculate a target direction vector from said initial inertial sensor data when said targeting axis is oriented at said target, wherein said target direction vector indicates a desired aim direction to said target;receive additional inertial sensor data from said inertial sensor when said object is placed in its normal use orientation and aimed by said user;calculate said forward direction vector that defines said current aim direction from said additional inertial sensor data, wherein said current aim direction indicates an actual aim direction in which said object is currently aimed to said target;calculate a rotation between said forward direction vector of said object and said target direction vector based on said additional inertial sensor data;calculate said aim alignment signal based on said rotation, wherein said aim alignment signal comprises information on how to align said forward direction vector with said target direction vector; and,send said aim alignment signal to said feedback element to assist said user in modifying or correcting said current aim direction of said object to said target while said object is aimed at said target.
  • 2. The system of claim 1, wherein said piece of sporting equipment is a golf club;said forward direction vector is a normal vector to a clubface of said golf club; and,said target direction vector points towards a golf hole.
  • 3. The system of claim 1, wherein said feedback element is a display; and,said feedback signal is one or more of text and graphics shown on said display.
  • 4. The system of claim 1, wherein said feedback signal is an audio signal.
  • 5. The system of claim 4, wherein said audio signal is a spoken message.
  • 6. The system of claim 5, wherein said spoken message tells said user one or both of a direction in which to rotate said object to align said forward direction vector and said target direction vector; and,an angular magnitude by which to rotate said object to align said forward direction vector and said target direction vector.
  • 7. The system of claim 1, wherein said feedback signal is a haptic feedback signal.
  • 8. The system of claim 7, wherein said feedback element is a watch or a wrist band.
  • 9. The system of claim 1, wherein said feedback signal comprises a magnitude attribute that increases monotonically or decreases monotonically as a function of an angular magnitude of said rotation between said desired aim direction to said target and said current aim direction.
  • 10. The system of claim 9, wherein said magnitude attribute is one or both of an amplitude of said feedback signal; and,a frequency of said feedback signal.
  • 11. The system of claim 1, wherein said feedback signal comprises a direction attribute that indicates a direction of an axis of said rotation to bring said forward direction vector into alignment with said target direction vector.
  • 12. The system of claim 11, wherein said direction attribute is one or both of an amplitude of said feedback signal; and,a frequency of said feedback signal.
  • 13. The system of claim 1, wherein said processor further projects said target direction vector onto a horizontal plane, yielding a horizontal target direction vector, before said rotation is calculated;projects said forward direction vector onto said horizontal plane, yielding a horizontal forward direction vector, before said rotation is calculated; and,calculates said rotation between said horizontal forward direction vector and said horizontal target direction vector.
  • 14. The system of claim 1, further comprising a laser coupled to said object and configured to project a beam of light in a direction of said targeting axis.
  • 15. The system of claim 1, wherein said piece of sporting equipment is a golf club;said forward direction vector is a normal vector to a clubface of said golf club; and,said targeting axis is an axis along said shaft of said golf club.
  • 16. The system of claim 15, wherein said processor is further configured to calculate a slope of a ground from said initial inertial sensor data when said shaft of said golf club is placed on said ground; and,transmit said slope to said feedback element.
  • 17. The system of claim 1, wherein said calculate said target direction vector from said initial inertial sensor data further comprises tracking and integrating said initial inertial sensor data as said piece of sporting equipment is rotated.
  • 18. The system of claim 1, wherein said processor is further configured to perform a transformation between a reference frame of said inertial sensor and a reference frame associated with said forward direction vector.
  • 19. The system of claim 1, wherein said object being aimed serves as a sensor to acquire said target direction vector.
US Referenced Citations (365)
Number Name Date Kind
1712537 White May 1929 A
3182508 Varju May 1965 A
3226704 Petrash Dec 1965 A
3270564 Evans Sep 1966 A
3776556 McLaughlin Dec 1973 A
3788647 Evans Jan 1974 A
3792863 Evans Feb 1974 A
3806131 Evans Apr 1974 A
3945646 Hammond Mar 1976 A
4759219 Cobb et al. Jul 1988 A
4902014 Bontomase et al. Feb 1990 A
4910677 Remedio et al. Mar 1990 A
4898389 Plutt Jun 1990 A
4940236 Allen Oct 1990 A
4991850 Wilhlem Feb 1991 A
5056783 Matcovich et al. Oct 1991 A
5086390 Matthews Apr 1992 A
5111410 Nakayama et al. May 1992 A
5127044 Bonito et al. Jun 1992 A
5184295 Mann Feb 1993 A
5230512 Tattershall Jul 1993 A
5233544 Kobayashi Aug 1993 A
5249967 O'Leary et al. Oct 1993 A
5259620 Marocco Nov 1993 A
5283733 Colley Jan 1994 A
5298904 Olich Mar 1994 A
5332225 Ura Jul 1994 A
5333061 Nakashima et al. Jul 1994 A
5364093 Huston et al. Nov 1994 A
5372365 McTeigue et al. Dec 1994 A
5441256 Hackman Aug 1995 A
5441269 Henwood Aug 1995 A
5443260 Stewart et al. Aug 1995 A
5486001 Baker Jan 1996 A
5524081 Paul Apr 1996 A
5542676 Howe et al. Aug 1996 A
5592401 Kramer Jan 1997 A
5610590 Johnson et al. Mar 1997 A
5638300 Johnson Jun 1997 A
5665006 Pellegrini Sep 1997 A
5688183 Sabatino et al. Nov 1997 A
5694340 Kim Dec 1997 A
5707299 McKenna Jan 1998 A
5772522 Nesbit Jun 1998 A
5779555 Nomura et al. Jul 1998 A
5819206 Horton Oct 1998 A
5826578 Curchod Oct 1998 A
5792001 Henwood Nov 1998 A
5868578 Baum Feb 1999 A
5904484 Burns May 1999 A
5941779 Zeiner-Gundersen Aug 1999 A
5973596 French et al. Oct 1999 A
5993333 Heckaman Nov 1999 A
5998968 Pittman et al. Dec 1999 A
6012995 Martin Jan 2000 A
6030109 Lobsenz Feb 2000 A
6044704 Sacher Apr 2000 A
6073086 Marinelli Jun 2000 A
6224493 Lee et al. May 2001 B1
6248021 Ognjanovic Jun 2001 B1
6253159 Bett et al. Jun 2001 B1
6254492 Taggett Jul 2001 B1
6266623 Vock et al. Jul 2001 B1
6293802 Ahlgren Sep 2001 B1
6366205 Sutphen Apr 2002 B1
6441745 Gates Aug 2002 B1
6456938 Barnard Sep 2002 B1
6537076 McNitt Mar 2003 B2
6540620 Consiglio Apr 2003 B1
6567536 McNitt May 2003 B2
6582328 Kuta et al. Jun 2003 B2
6611141 Schulz Aug 2003 B1
6697820 Tarlie Feb 2004 B1
6705942 Crook et al. Mar 2004 B1
6746336 Brant et al. Jun 2004 B1
6757572 Forest Jun 2004 B1
6774932 Ewing et al. Aug 2004 B1
6802772 Kunzle et al. Dec 2004 B1
6923729 McGinty et al. Feb 2005 B2
6868338 Elliott Mar 2005 B1
6900759 Katayama May 2005 B1
6908404 Gard Jun 2005 B1
7004848 Konow Feb 2006 B2
7037198 Hameen-Antilla Feb 2006 B2
7021140 Perkins Apr 2006 B2
7034694 Yamaguchi et al. Apr 2006 B2
7143639 Gobush May 2006 B2
7092846 Vock et al. Aug 2006 B2
7118498 Meadows et al. Oct 2006 B2
7121962 Reeves Oct 2006 B2
7175177 Meifu et al. Feb 2007 B2
7205894 Savage Apr 2007 B1
7212943 Aoshima et al. May 2007 B2
7219033 Kolen May 2007 B2
7234351 Perkins Jun 2007 B2
7160200 Grober Sep 2007 B2
7264554 Bentley Sep 2007 B2
7283647 Mcnitt Oct 2007 B2
7433805 Vock et al. Jul 2008 B2
7421369 Clarkson Sep 2008 B2
7457439 Madsen Nov 2008 B1
7457724 Vock et al. Nov 2008 B2
7492367 Mahajan et al. Feb 2009 B2
7494236 Lim Feb 2009 B2
7499828 Barton Mar 2009 B2
7561989 Banks Jul 2009 B2
7623987 Vock et al. Nov 2009 B2
7627451 Vock et al. Dec 2009 B2
7689378 Kolen Mar 2010 B2
7713148 Sweeney May 2010 B2
7731598 Kim Jun 2010 B1
7736242 Stites et al. Jun 2010 B2
7771263 Telford Aug 2010 B2
7780450 Tarry Aug 2010 B2
7800480 Joseph et al. Sep 2010 B1
7813887 Vock et al. Oct 2010 B2
7831212 Balardeta et al. Nov 2010 B1
7871333 Davenport Jan 2011 B1
7966154 Vock et al. Jun 2011 B2
7983876 Vock et al. Jul 2011 B2
8036826 MacIntosh et al. Nov 2011 B2
8117888 Chan et al. Feb 2012 B2
8172722 Molyneux et al. May 2012 B2
8231506 Molyneux et al. Jul 2012 B2
8246050 Do Aug 2012 B2
8249831 Vock et al. Aug 2012 B2
8257191 Stites et al. Sep 2012 B2
8282487 Wilson et al. Oct 2012 B2
8314840 Funk Nov 2012 B1
8352211 Vock et al. Jan 2013 B2
8400548 Bilbrey et al. Mar 2013 B2
8425292 Lui et al. Apr 2013 B2
8477027 Givens Jul 2013 B2
8527228 Panagas Sep 2013 B2
8565483 Nakaoka Oct 2013 B2
8589114 Papadourakis Nov 2013 B2
8696482 Pedenko et al. Apr 2014 B1
8723986 Merrill May 2014 B1
8725452 Han May 2014 B2
8764576 Takasugi Jul 2014 B2
8781610 Han Jul 2014 B2
8831905 Papadourakis Sep 2014 B2
8876621 Shibuya Nov 2014 B2
8888603 Sato et al. Nov 2014 B2
8905856 Parke Dec 2014 B2
8929709 Lokshin Jan 2015 B2
8944932 Sato et al. Feb 2015 B2
8944939 Clark et al. Feb 2015 B2
8956238 Boyd et al. Feb 2015 B2
8988341 Lin et al. Mar 2015 B2
8989441 Han et al. Mar 2015 B2
9032794 Perkins et al. May 2015 B2
9060682 Lokshin Jun 2015 B2
9146134 Lokshin et al. Sep 2015 B2
9395385 Parke Jul 2016 B2
9656122 Papadourakis May 2017 B2
9694267 Thornbrue et al. Jul 2017 B1
20010035880 Musatov et al. Jan 2001 A
20010049636 Hudda et al. Jun 2001 A1
20010029207 Cameron et al. Oct 2001 A1
20010045904 Silzer, Jr. Nov 2001 A1
20020019677 Lee Feb 2002 A1
20020052750 Hirooka Feb 2002 A1
20020049507 Hameen-Anttila Apr 2002 A1
20020064764 Fishman May 2002 A1
20020072815 McDonough et al. Jun 2002 A1
20020077189 Tuer et al. Jun 2002 A1
20020082775 Meadows et al. Jun 2002 A1
20020115046 McNitt et al. Aug 2002 A1
20020126157 Farago et al. Sep 2002 A1
20020004723 Meifu et al. Oct 2002 A1
20020151994 Sisco Oct 2002 A1
20020173364 Boscha Nov 2002 A1
20020177490 Yong et al. Nov 2002 A1
20020188359 Morse Dec 2002 A1
20030073518 Marty Apr 2003 A1
20030074659 Louzoun Apr 2003 A1
20030109322 Funk et al. Jun 2003 A1
20030163287 Vock et al. Aug 2003 A1
20030008722 Konow Sep 2003 A1
20030191547 Morse Sep 2003 A1
20030208830 Marmaropoulos Nov 2003 A1
20040028258 Naimark et al. Feb 2004 A1
20040033843 Miller Feb 2004 A1
20040044493 Coulthard Mar 2004 A1
20040147329 Meadows et al. Jul 2004 A1
20040227676 Kim et al. Nov 2004 A1
20040248676 Taylor Dec 2004 A1
20050021292 Vock et al. Jan 2005 A1
20050023763 Richardson Feb 2005 A1
20050032582 Mahajan et al. Feb 2005 A1
20050054457 Eyestone et al. Mar 2005 A1
20050156068 Ivans Jul 2005 A1
20050268704 Bissonnette et al. Aug 2005 A1
20050272516 Gobush Aug 2005 A1
20050203430 Williams et al. Sep 2005 A1
20050213076 Saegusa Sep 2005 A1
20050215340 Stites et al. Sep 2005 A1
20050227775 Cassady et al. Oct 2005 A1
20050261073 Farrington, Jr. et al. Nov 2005 A1
20050282650 Miettinen et al. Dec 2005 A1
20050288119 Wang et al. Dec 2005 A1
20060020177 Seo et al. Jan 2006 A1
20060025229 Mahajan et al. Feb 2006 A1
20060038657 Denison et al. Feb 2006 A1
20060063600 Grober Mar 2006 A1
20060068928 Nagy Mar 2006 A1
20060084516 Eyestone et al. Apr 2006 A1
20060109116 Keays May 2006 A1
20060122002 Konow Jun 2006 A1
20060166738 Eyestone et al. Jul 2006 A1
20060199659 Caldwell Jul 2006 A1
20060276256 Storek Jul 2006 A1
20060189389 Hunter et al. Aug 2006 A1
20060247070 Funk et al. Nov 2006 A1
20060250745 Butler et al. Nov 2006 A1
20060270450 Garratt et al. Nov 2006 A1
20060284979 Clarkson Dec 2006 A1
20060293112 Yi Dec 2006 A1
20070062284 Machida Mar 2007 A1
20070081695 Foxlin et al. Apr 2007 A1
20070087866 Meadows et al. Apr 2007 A1
20070099715 Jones et al. May 2007 A1
20070111811 Grober May 2007 A1
20070135225 Nieminen Jun 2007 A1
20070135237 Reeves Jun 2007 A1
20070129178 Reeves Jul 2007 A1
20070052807 Zhou et al. Aug 2007 A1
20070219744 Kolen Sep 2007 A1
20070265105 Barton Nov 2007 A1
20070270214 Bentley Nov 2007 A1
20070298753 Tary et al. Dec 2007 A1
20070298896 Nusbaum Dec 2007 A1
20080027502 Ransom Jan 2008 A1
20080085778 Dugan Apr 2008 A1
20080090703 Rosenberg Apr 2008 A1
20080108456 Bonito May 2008 A1
20080164999 Otto Jul 2008 A1
20080182685 Marty Jul 2008 A1
20080190202 Kulach et al. Aug 2008 A1
20080234935 Wolf et al. Sep 2008 A1
20080280642 Coxhill et al. Nov 2008 A1
20080284979 Yee et al. Nov 2008 A1
20080285805 Luinge et al. Nov 2008 A1
20090002316 Rofougaran Jan 2009 A1
20090017944 Savarese et al. Jan 2009 A1
20090029754 Slocum et al. Jan 2009 A1
20090033741 Oh et al. Feb 2009 A1
20090036237 Nipper et al. Feb 2009 A1
20090048044 Oleson et al. Feb 2009 A1
20090055820 Huang Feb 2009 A1
20090088276 Solheim et al. Apr 2009 A1
20090111602 Savarese et al. Apr 2009 A1
20090131190 Kimber May 2009 A1
20090137333 Lin et al. May 2009 A1
20090174676 Westerman Jul 2009 A1
20090177097 Ma et al. Jul 2009 A1
20090191846 Shi Jul 2009 A1
20090209343 Foxlin et al. Aug 2009 A1
20090209358 Niegowski Aug 2009 A1
20090213134 Stephanick et al. Aug 2009 A1
20090222163 Plante Sep 2009 A1
20090233735 Savarese et al. Sep 2009 A1
20090254276 Faulkner et al. Oct 2009 A1
20090254971 Herz et al. Oct 2009 A1
20090299232 Lanfermann et al. Dec 2009 A1
20100049468 Papadourakis Feb 2010 A1
20100062869 Chung et al. Mar 2010 A1
20100091112 Veeser et al. Apr 2010 A1
20100093458 Davenport et al. Apr 2010 A1
20100099509 Ahem et al. Apr 2010 A1
20100103269 Wilson et al. Apr 2010 A1
20100113174 Ahern May 2010 A1
20100121227 Stirling et al. May 2010 A1
20100121228 Stirling et al. May 2010 A1
20100130298 Dugan et al. May 2010 A1
20100144414 Edis et al. Jun 2010 A1
20100144456 Ahern Jun 2010 A1
20100144457 Kim Jun 2010 A1
20100178994 Do Jul 2010 A1
20100201512 Stirling et al. Aug 2010 A1
20100204616 Shears et al. Aug 2010 A1
20100216564 Stites et al. Aug 2010 A1
20100222152 Jaekel et al. Sep 2010 A1
20100063778 Schrock et al. Nov 2010 A1
20100063779 Schrock et al. Nov 2010 A1
20100308105 Savarese et al. Dec 2010 A1
20100309097 Raviv Dec 2010 A1
20100323794 Su Dec 2010 A1
20110004871 Liu Jan 2011 A1
20110029235 Berry Feb 2011 A1
20110037778 Deng et al. Feb 2011 A1
20110050864 Bond Mar 2011 A1
20110052005 Selner Mar 2011 A1
20110053688 Crawford Mar 2011 A1
20110075341 Lau et al. Mar 2011 A1
20110081981 Okamoto Apr 2011 A1
20110126184 Lisboa May 2011 A1
20110165998 Lau et al. Jul 2011 A1
20110195780 Lu Aug 2011 A1
20110230273 Niegowski et al. Sep 2011 A1
20110230274 Lafortune et al. Sep 2011 A1
20110230985 Niegowski et al. Sep 2011 A1
20110230986 Lafortune Sep 2011 A1
20110238308 Miller et al. Sep 2011 A1
20110305369 Bentley Dec 2011 A1
20120023354 Chino Jan 2012 A1
20120052972 Bentley Mar 2012 A1
20120115626 Davenport May 2012 A1
20120115682 Homsi May 2012 A1
20120116548 Goree et al. May 2012 A1
20120120572 Bentley May 2012 A1
20120157241 Nomura et al. Jun 2012 A1
20120179418 Takasugi et al. Jul 2012 A1
20120179742 Acharya et al. Jul 2012 A1
20120191405 Molyneux et al. Jul 2012 A1
20120295726 Cherbini Nov 2012 A1
20120316004 Shibuya Dec 2012 A1
20130029791 Rose et al. Jan 2013 A1
20130095941 Bentley et al. Apr 2013 A1
20130110415 Davis et al. May 2013 A1
20130128022 Bose et al. May 2013 A1
20130173212 Saiki et al. Jul 2013 A1
20130178304 Chan Jul 2013 A1
20130191063 Nomura Jul 2013 A1
20130225309 Bentley et al. Aug 2013 A1
20130267335 Boyd Oct 2013 A1
20130271602 Bentley et al. Oct 2013 A1
20130298668 Sato Nov 2013 A1
20130319113 Mizuta Dec 2013 A1
20130330054 Lokshin Dec 2013 A1
20130332004 Gompert et al. Dec 2013 A1
20130346013 Lokshin Dec 2013 A1
20140019083 Nakaoka Jan 2014 A1
20140100048 Ota et al. Apr 2014 A1
20140100049 Ota et al. Apr 2014 A1
20140100050 Ota et al. Apr 2014 A1
20140135139 Shibuya et al. May 2014 A1
20140156214 Nomura Jun 2014 A1
20140172873 Varoglu et al. Jun 2014 A1
20140200092 Parke Jul 2014 A1
20140200094 Parke et al. Jul 2014 A1
20140229135 Nomura Aug 2014 A1
20140229138 Goree et al. Aug 2014 A1
20140257743 Lokshin et al. Sep 2014 A1
20140257744 Lokshin et al. Sep 2014 A1
20140295982 Shibuya Oct 2014 A1
20140376876 Bentley et al. Dec 2014 A1
20140378239 Sato et al. Dec 2014 A1
20140379293 Sato Dec 2014 A1
20140379294 Shibuya et al. Dec 2014 A1
20140379295 Sato et al. Dec 2014 A1
20150007658 Ishikawa et al. Jan 2015 A1
20150012240 Sato Jan 2015 A1
20150042481 Nomura Feb 2015 A1
20150098688 Lokshin Apr 2015 A1
20150107358 Parke Apr 2015 A1
20150124048 King May 2015 A1
20150126308 Penn May 2015 A1
20150131845 Forouhar et al. May 2015 A1
20150154452 Bentley et al. Jun 2015 A1
20150256689 Erkkila et al. Sep 2015 A1
20150348591 Kaps et al. Dec 2015 A1
20160354668 Cho Dec 2016 A1
20170061817 Mettler May Mar 2017 A1
Foreign Referenced Citations (16)
Number Date Country
2025369 Feb 2009 EP
2002210055 Jul 2002 JP
2004207985 Jul 2004 JP
10-2006-0041060 May 2006 JP
2011000367 Jun 2011 JP
2012196241 Oct 2012 JP
10-20030085275 Nov 2003 KR
10-2007-0119018 Dec 2007 KR
10-20100020131 Feb 2010 KR
10-20100074068 Jul 2010 KR
10-1079319 Apr 2011 KR
9427683 Aug 1994 WO
2007130057 Nov 2007 WO
2009056688 May 2009 WO
2011057194 May 2011 WO
2014085744 Jun 2014 WO
Non-Patent Literature Citations (80)
Entry
Patent Examination Report received in Australia Application No. 2011313952, dated Mar. 15, 2016, 5 pages.
“About Banjo” retrieved from the internet, dated 2015.
International Search Report received in PCT/US2016/042668, dated Oct. 4, 2016, 21 pages.
International Search Report received in PCT/US2016/042671, dated Oct. 13, 2016, 17 pages.
International Search Report and Written Opinion received in PCT/US2016/042676, dated Oct. 24, 2016 (12 pages).
International Preliminary Report on Patentability received in PCT/US2015/026917, dated Nov. 3, 2016 (5 pages).
International Preliminary Report on Patentability in PCTUS2015061695 (our matter 19007-P082-PCT), on Jun. 1, 2017, 5 pages.
European Search Report received in PCTUS2015026896 on May 11, 2017, 13 pages.
Zepp Labs, Inc. v. Blast Motion, Inc. Petition for Inter Partes Review of U.S. Pat. No. 8,903,521 filed on Feb. 24, 2016, as IPR2016-00672, and accompanying Declaration of Dr. Steven M. Nesbit.
Zepp Labs, Inc. v. Blast Motion, Inc. Petition for Inter Partes Review of U.S. Pat. No. 9,039,527 filed on Feb. 24, 2016, as IPR2016-00674, and accompanying Declaration of Dr. Steven M. Nesbit.
Zepp Labs, Inc. v. Blast Motion, Inc. Petition for Inter Partes Review of U.S. Pat. No. 8,941,723 filed on Feb. 24, 2016, as IPR2016-00675, and accompanying Declaration of Dr. Steven M. Nesbit.
Zepp Labs, Inc. v. Blast Motion, Inc. Petition for Inter Partes Review of U.S. Pat. No. 8,905,855 filed on Feb. 24, 2016, as IPR2016-00676, and accompanying Declaration of Dr. Steven M. Nesbit.
Zepp Labs, Inc. v. Blast Motion, Inc. Petition for Inter Partes Review of U.S. Pat. No. 8,944,928 filed on Feb. 24, 2016, as IPR2016-00677, and accompanying Declaration of Dr. Steven M. Nesbit.
Chris Otto, et al, “System Architecture of a Wireless Body Area Sensor Network for Ubiquitous Health Monitoring”, Journal of Mobile Multimedia, vol. 1, No. 4, Jan. 10, 2006, University of Alabama in Huntsville, 20 Pages.
Linx Technologies “High Performance RF Module: Hp3 Series Transmitter Module Data Guide Description”, Jul. 27, 2011, 13 pages.
Roger Allan, “Wireless Sensor Architectures Uses Bluetooth Standard”, www.electronicdesign.com/communications/wireless-sensor-architecture-uses-bluetooth-standard, Aug. 7, 2000, 5 pages.
Don Tuite, “Motion-Sensing MEMS Gyros and Accelerometers are Everywhere”, www.electronicdesign.com/print/analog/motion-sensing-mems-gyros-and-accelerometers-are-everywhere, Jul. 9, 2009, 6 pages.
InvenSense News Release, “InvenSense Unveils World's 1st IMU Solution for Consumer Applications”, ir.invensense.com, 2016, 2 Pages.
Dean Takahashi, “Facebook, Twitter, Last.fm coming to Xbox Live this Fall”, Jun. 1, 2009, Webpage printout, 5 pages.
The iClub System, Products pages, www.iclub.net, 2001-2005, 5 pages.
Websters New College Dictionary , Definition of “Virtual Reality”, Third Edition, 2005, 3 Pages.
SmartSwing, “SmartSwing Introduces Affordable Intelligent Golf Club”, www.smartswinggolf.com , Jan. 2006, 2 pages.
Henrick Arfwedson, et al., “Ericsson's Bluetooth modules”, Ericsson Review No. 4, 1999, 8 pages.
ZigBees, “Zigbee information”, www.zigbees.com , 2015, 4 pages.
SolidState Technology, “MEMS enable smart golf clubs”, www.electroiq.com , 2005, 3 pages.
IGN, “Japanese WII Price Release Date Revealed”, www.ign.com , 2006, 1 page.
First Annual Better Golf Through Technology Conference 2006 webpage, www.bettergolfthroughtechnology.com , Massachusetts Institute of Technology, Cambridge Massachusetts, Feb. 2006, 1 page.
Concept2Rowing, “Training” web content, www.concept2.com , 2009, 1 page.
Expresso, Products pages, www.expresso.com/products , 2009, 2 pages.
Manish Kalia, et al., “Efficient Policies for Increasing Capacity in Bluetooth: An Indoor Pico-Cellular Wireless System”, IBM India Research Laboratory, Indian Institute of Technology, 2000, 5 pages.
R.Rao, et al., “Demand-Based Bluetooth Scheduling”, Pennsylvania State University, 2001, 13 pages.
International Search Report received for PCT Application No. PCT/US2012/065716, dated Jan. 3, 2013, 10 pages.
MyCaddie, 2009, retrieved on Sep. 26, 2012 from http://www.iMakePars.com, 4 pages.
Swing it See it Fix it, Improve Gold Swing, SwingSmart Golf Analyzer, retrieved on Sep. 26, 2012 from http://www.SwingSmart.com, 2 pages.
Learn how Swingbyte can improve your game, retrieved on Sep. 26, 2012 from http://www.swingbyte.com, 2 pages.
International Search Report received for PCT Application No. PCT/US2011/055173, dated Mar. 6, 2012, 8 pages.
International Search Report received for PCT Application No. PCT/US2011/049461, dated Feb. 23, 2012, 14 pages, 2012.
PCT Search Report, PCT/US2012/029310, dated Sep. 28, 2012, 3 pages.
IPRP, PCT/US2011/049461, dated Mar. 7, 2013, 6 pages.
IPRP, PCT/US2011/058182, dated Apr. 30, 2013, 5 pages.
IPER, PCT/US2011/055173, dated Apr. 25, 2013, 5 pages, (2013).
IPRP, PCT/US2012/065716, dated May 20, 2014, 6 pages.
International Search Report for PCT Application No. PCT/US2013/021999, dated Apr. 30, 2013, 8 pages.
International Search Report for PCT Application No. PCT/US2012/066915, dated Mar. 29, 2013, 10 pages.
International Search Report for PCT Application No. PCT/US2015/26896, dated Jul. 28, 2015, 15 pages.
International Search Report for PCT Application No. PCTUS2015/26917, dated Jul. 30, 2015, 16 pages.
The Nike+FuelBand User's Guide, rev 14, 26 pages, 2012.
UP by Jawbone Extended User Guide, 10 pages, 2012.
Armour39, Under Armour Guarantee, Getting Started, retrieved from the Internet on Jul. 12, 2013, 7 pages.
Armour39 Module & Chest Strap, retrieved from the Internet on Jul. 12, 2013, 6 pages.
MiCoach Pacer User Manual, 31 pages, (2009).
Foreman et al. “A Comparative Analysis for the Measurement of Head Accelerations in Ice Hockey Helmets using Non-Accelerometer Based Systems,” Nov. 19, 2012, 13 pages.
Reebok-CCM and MC10 to Launch Revolutionary Sports Impact Indicator, MC10 News (http://www.mc10inc.com/news/), Oct. 24, 2012, 3 pages.
CheckLight MC10 Overview, Reebok International Limited, Nov. 20, 2012, 7 pages.
Reebok and MC10 Team Up to Build CheckLight, a Head Impact Indicator (Hands-on), MC10 News (http://www.mc10inc.com/news/), Jan. 11, 2013, 1 pg.
Trace—The Most Advanced Activity Monitor for Action Sports, webpage, retrieved on Aug. 6, 2013, 22 pages.
CheckLight, Sports/Activity Impact Indicator, User Manual, 13 pages, 2013, Reebok International Limited.
King, The Design and Application of Wireless Mems Inertial Measurement Units for the Measurement and Analysis of Golf Swings, 2008.
Grober, An Accelerometer Based Instrumentation of the Golf Club: Comparative Analysis of Golf Swings, 2009.
Gehrig et al, Visual Golf Club Tracking for Enhanced Swing Analysis, Computer Vision Lab, Lausanne, Switzerland, 2003.
PocketPro Golf Designs, PocketPro Full Swing Analysis in Your Pocket, www.PocketPro.org, (2011).
Clemson University, Golf Shot Tutorial, http://www.webnucleo.org/home/online_tools/newton/0.4/html/about_this_tool/tutorials/golf_1.shp.cgi, retrieved on Nov. 10, 2011.
MiCoach Speed_Cell TM, User Manual, 23 pages, (2011).
Nike+iPod, User Guide, 32 pages (2010).
SureShotGPS SS9000X, Intelligent Touch, Instruction Manual, 25 page, 2011.
ActiveReplay, “TRACE—The Most Advanced Activity Monitor for Action Sports”, http://www.kickstarter.com/projects/activereplay/trace-the-most-advanced-activity-monitor-for-actio, 13 pages, Oct. 1, 2013.
Zepp Golfsense@Launch2011, https://www.youtube.com/watch?v=VnOcu8szjIk (video), Mar. 14, 2011.
Epson US Newsroom, “Epson America Enters Sports Wearables Market with Introduction of M-Tracer MT500GII Golf Swing Analyzer”, www.news.epson.com, Jan. 5, 2015, 4 pages.
International Search Report and Written Opinion dated Dec. 22, 2015 received in PCTUS1561695, 7 pages.
Search Report Received in PCT2013021999 dated Jan. 21, 2016.
International Search Report and Written Opinion mailed in PCTUS1642674 dated Aug. 12, 2016, 9 pages.
International Search Report and Written Opinion received in PCT/US2017/52114, dated Oct. 3, 9 pages.
International Search Report and Written Opinion Received in PCT/US2017/37987, dated Nov. 9, 2017, 12 pages.
Supplementary Extended European Search Report received in 11820763.8 dated Nov. 13, 2017, 16 pages.
Supplementary Extended European Search Report received in 11833159.4 dated Nov. 6, 2017, 14 pages.
Supplementary Partial European Search Report received from EP Application Serial No. 11820763.8, dated Aug. 8, 2017, 15 pages.
Supplementary Partial European Search Report received from EP Application Serial No. 1833159.4, dated Aug. 8, 2017, 15 pages.
David E. Culler, Et al., “Smart Sensors to Network the World”, published in Scientific American Magazine, No. 06/2004, dated Jun. 1, 2004, pp. 85-91.
International Search Report and Written Opinion received in PCT/US2017/039209, dated Aug. 24, 2017, 7 pages.
Supplementary Extended European Search Report received in 15782595.1 dated Nov. 27, 2017, 5 pages.
Related Publications (1)
Number Date Country
20170252607 A1 Sep 2017 US