A variety of methods have been proposed to measure head impacts. One approach uses sensors in a helmet. This approach is flawed since the helmet may rotate on the head during an impact, or even become displaced.
Another approach uses tri-axial accelerometers embedded in patches attached to the head. This approach is has limited accuracy since the position and orientation of the patches on head is not known precisely.
Yet another approach uses a combination of a tri-axial linear accelerometer and a gyroscope. This approach yields rotations and linear acceleration at the sensor location. However, when the desire is to measure the motion of a rigid body, such as a human head, it is often impossible or impractical to place a sensor at the center of the rigid body.
A helmet is used to provide the head with a degree of protection from impacts. The performance of helmet is commonly assessed in the laboratory by subjecting an instrumented dummy head to known impacts. The linear and rotational acceleration of the dummy head is measured with and without the helmet to determine the degree of protection provided. However, in practice the degree of protection is dependent upon additional factors, such as how well the helmet fits the head, how the helmet is placed on the head, degradation of the helmet with use, the direction and intensity of the impact, etc. Accordingly, it would be useful to provide means for in-situ measurement of helmet performance. Such a means could also be used to simplify laboratory testing of helmets.
The accompanying figures, in which like reference numerals refer to identical or functionally similar elements throughout the separate views and which together with the detailed description below are incorporated in and form part of the specification, serve to further illustrate various embodiments and to explain various principles and advantages all in accordance with the present invention.
Skilled artisans will appreciate that elements in the figures are illustrated for simplicity and clarity and have not necessarily been drawn to scale. For example, the dimensions of some of the elements in the figures may be exaggerated relative to other elements to help to improve understanding of embodiments of the present invention.
Before describing in detail embodiments that are in accordance with the present invention, it should be observed that the embodiments reside primarily in combinations of method steps and apparatus components related to monitoring motion of a substantially rigid body, such as a head, and monitoring performance of a protective body, such as a helmet. Accordingly, the apparatus components and method steps have been represented where appropriate by conventional symbols in the drawings, showing only those specific details that are pertinent to understanding the embodiments of the present invention so as not to obscure the disclosure with details that will be readily apparent to those of ordinary skill in the art having the benefit of the description herein.
In this document, relational terms such as first and second, top and bottom, and the like may be used solely to distinguish one entity or action from another entity or action without necessarily requiring or implying any actual such relationship or order between such entities or actions. The terms “comprises,” “comprising,” or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or apparatus that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus. An element preceded by “comprises . . . a” does not, without more constraints, preclude the existence of additional identical elements in the process, method, article, or apparatus that comprises the element.
It will be appreciated that embodiments of the invention described herein may include the use of one or more conventional processors and unique stored program instructions that control the one or more processors to implement, in conjunction with certain non-processor circuits, some, most, or all of the functions of monitoring head and helmet accelerations described herein. The non-processor circuits may include, but are not limited to, a radio receiver, a radio transmitter, signal drivers, clock circuits, power source circuits, and user input devices. As such, these functions may be interpreted as a method to monitor head and helmet accelerations, or accelerations of other bodies. Alternatively, some or all functions could be implemented by a state machine that has no stored program instructions, or in one or more application specific integrated circuits (ASICs), in which each function or some combinations of certain of the functions are implemented as custom logic. Of course, a combination of the two approaches could be used. Thus, methods and means for these functions have been described herein. Further, it is expected that one of ordinary skill, notwithstanding possibly significant effort and many design choices motivated by, for example, available time, current technology, and economic considerations, when guided by the concepts and principles disclosed herein will be readily capable of generating such software instructions and programs and ICs with minimal experimentation.
In the foregoing specification, specific embodiments of the present invention have been described. However, one of ordinary skill in the art appreciates that various modifications and changes can be made without departing from the scope of the present invention as set forth in the claims below. Accordingly, the specification and figures are to be regarded in an illustrative rather than a restrictive sense, and all such modifications are intended to be included within the scope of the present invention. The benefits, advantages, solutions to problems, and any element(s) that may cause any benefit, advantage, or solution to occur or become more pronounced are not to be construed as a critical, required, or essential features or elements of any or all the claims. The invention is defined solely by the appended claims including any amendments made during the pendency of this application and all equivalents of those claims as issued.
The present disclosure relates to a method and apparatus for monitoring motion of a rigid body, such as a human head or a helmet, relative to a first location. Linear and rotational motions are sensed by one or more sensors attached to the rigid body at locations displaced from the first location. The sensed rotation is used to compensate for the angular and centripetal acceleration components in the sensed linear motion. In one embodiment, the angular and centripetal acceleration components are estimated explicitly from the sensed rotation. In a further embodiment, the sensed rotations are used to estimate the relative orientations of two or more sensors, enabling the linear motions measured by the two sensors to be combined so as to cancel the angular and centripetal accelerations. Monitored head motion may be compared with monitored helmet motion to determine the performance of the helmet.
In one embodiment, a six degree-of-freedom sensor comprises a three-axis linear motion sensor, such as a tri-axial accelerometer that senses local linear motion, and a rotational sensor that measures three components of a rotational motion. The rotational sensor maybe, for example, a three-axis gyroscope that senses angular velocity, or a three-axis rotational accelerometer that senses the rate of change of angular velocity with time, or a three axis angular displacement sensors such as a compass, or a combination thereof. The six degree-of freedom sensor may comprise more than six sensing elements. For example, both rotational rate and rotational acceleration could be sensed (or even rotational position). These signals are not independent, since they are related through their time histories. However, having both types of sensors may avoid the need for integration or differentiation.
The processor 100 receives the sensor signals 108 and 110 and from them generates angular acceleration signals 112 and linear acceleration signals 114 in a frame of reference that does not have its origin at a sensor position and may not have its axes aligned with the axes of the sensor.
In one embodiment, which uses two sensors, the origin of the frame of reference is at a midpoint of the line A-A between the sensors 102 and 104, denoted in
In a further embodiment, which uses a single sensor, the origin may be selected to be any point whose position is known relative to the single sensor.
In the selected frame of reference, the vector of angular velocities of the substantially rigid body is denoted as ω, the angular acceleration vector is denoted as {dot over (ω)}), and the linear acceleration vector is denoted as a.
It is noted that the angular acceleration may be obtained from angular velocity by differentiation with respect to time and, conversely, the angular velocity may be obtained from the angular acceleration by integration with respect to time. These integrations or differentiations may be performed using an analog circuit, a sampled data circuit or by digital signal processing. Thus, either type of rotation sensor could be used. Alternatively, or in addition, a rotation displacement sensor, such as a magnetic field sensor, may be used. Angular velocity and angular acceleration may then be obtained by single and double differentiation, respectively.
The response s of a linear accelerometer at a position r={r1,r2,r3}T in the selected frame of reference is given by
s=s
lin
[a+(K({dot over (ω)})+K2(ω))r]=Slin[a−K(r){dot over (ω)}+P(r)γ(ω)], (1)
where a is the linear acceleration vector at the origin of the frame of reference and γ(ω) is a vector of centripetal accelerations given by
Slin is the linear sensitivity matrix for the sensor (which is dependent upon the sensor orientation), the matrix function K is defined as the skew symmetric matrix given by
the matrix P is given by
In general, for a rotational sensor, the response vector is
w=S
rot(ω,{dot over (ω)}) (5)
where Srot is the angular sensitivity matrix of the sensor. From this we can get (using integration or differentiation as required)
ω=F(w)
{dot over (ω)}=G(w)′ (6)
where F and G are functions that depend upon the angular sensitivity matrix Srot of the sensor.
In accordance with a first aspect of the disclosure, the linear acceleration at the origin of the frame of reference may be derived from the sensed linear and rotation motion.
Rearranging equation (1) gives
a=S
lin
−1
s+K(r){dot over (ω)}−P(r)γ(ω), (7)
and estimating the rotational components from the rotation sensor signal w gives
a=S
lin
−1
s+K(r)G(w)−P(r)γ(F(w)), (8a)
or,
a=S
lin
−1
s−[K(G(w))+K2(F(w))]r, (8b)
Thus, the linear acceleration at the origin is obtained as a combination of the linear motion s, and rotational motion w sensed at the sensor location, the combination being dependent upon the position r of the sensor relative to the origin and the linear sensitivity and orientation of the sensor through the matrix Slin. The matrix parameters K (r) and P (r) used in the combination (8a) are dependent upon the position r.
For a rigid body, the rotational acceleration at the origin is the same as the rotational acceleration at the sensor location and is given by equation (6).
It is noted that the combination defined in equations (8a) and (8b) requires knowledge of the sensitivities of the sensor and knowledge of position of the sensor relative to the origin.
In equation (7), the matrix Slin is dependent upon the orientation of the sensor relative to the frame of reference.
In one embodiment the sensor is oriented in a known way on the rigid body. This is facilitated by marking the sensor (for example with an arrow).
In a further embodiment, the sensor is shaped to facilitate consistent positioning and/orientation on the body. For example, a behind-the-ear sensor may be shaped to conform to the profile of an ear, or a nose sensor is shaped to conform to the bridge of the nose.
In a still further embodiment, a measurement of the sensor orientation relative to the direction of gravity is made and the frame of reference is fixed relative to the direction of gravity.
In a still further embodiment, measurement of the sensor orientation is made relative to a reference sensor, shown as 118 in
A sensor may be attached using self-adhesive tape, for example. The sensor should be as light as possible, so that the resonance frequency of the sensor mass on the compliance of the skin is as high as possible (see, for example, ‘A Triaxial Accelerometer and Portable Data Processing Unit for the Assessment of Daily Physical Activity’, Carlijn V. C. Bouten et al., IEEE TRANSACTIONS ON BIOMEDICAL ENGINEERING, VOL. 44, NO. 3, MARCH 1997, page 145, column 2, and references therein). A self-adhesive, battery powered sensor may be used, the battery being activated when the sensor is attached to the head.
The sensor 102 may be calibrated with respect to the reference sensor 118.
The system 200 enables monitoring motion of a substantially rigid body relative to a first location in response to linear 108″ and rotational motion signals 108′ from a motion sensor 102 locatable on the substantially rigid body at a second location, displaced from the first location. The system comprises a processing module 202 responsive to the rotational motion signal 108′ and operable to produce a plurality of rotational components, 112 and 204. A memory 206 stores parameters dependent upon the first and second locations. A combiner 208 combines the plurality of rotational components with the linear motion signals 108″, dependent upon the parameters stored in the memory 206, to provide an estimate of the motion at the first location in the substantially rigid body. The signals 114 and/or 112, representative of the motion at the first location, are provided as outputs. The rotational components comprise first rotational components 112, dependent upon angular acceleration of the substantially rigid body and second rotational components 204 dependent upon angular velocity of the substantially rigid body.
The flow chart in
While the approach described above has the advantage of using a single sensor, one disadvantage is that, unless a reference sensor is used, the approach requires knowledge of position of the sensor relative to the origin. However, if a reference sensor is used, the position, orientation and sensitivity may be estimated.
In accordance with a second aspect of the present disclosure, the linear acceleration at the origin of the frame of reference may be derived from the sensed linear and rotation motion at two or more sensors. In one embodiment, two sensors are used, located on opposite sides of the desired monitoring position. For example, one sensor could be either side of a head to monitor motion relative to a location between the sensors. This approach avoids the need to know the sensor locations relative to the selected origin, and also avoids the need for differentiation or integration with respect to time, although more than one sensor is required.
To facilitate explanation, a two-sensor system is considered first. The first and second sensors are referred to as ‘left’ and ‘right’ sensors, however, it is to be understood that any pair of sensors may be used.
The origin is defined as the midpoint between the two sensors. Thus, the sensor positions are rL={r1,r2,r3}T for the left sensor and rR={−r1,−r2,−r3}T for the right sensor.
The accelerations are not necessarily the same, since, as discussed above, each measurement is in the frame of reference of the corresponding sensor. In the frame of reference of the left sensor,
S
L,lin
−1
s
L
=a+[K({dot over (ω)})+K2(ω)]rL, (9)
R
−1
S
R,lin
−1
s
R
=a+[K({dot over (ω)})+K2(ω)]rR, (10)
where R is a rotation matrix that is determined by the relative orientations of the two sensors and the sensitivity matrices are relative to the sensor's own frame of reference. R−1SR,lin−1sR is a vector of compensated and aligned right sensor signals and SL,lin−1sL is the vector of compensated left sensor signals.
Averaging (9) and (10) gives
where R is a rotation matrix that is determined by the relative orientations of the two sensors and the sensitivity matrices are relative to the sensor's own frame of reference. Here we have used rL+rR=0.
This allows the linear acceleration at the origin (the midpoint) to be estimated as the simple combination
In some applications, the left and right sensors may be orientated with sufficient accuracy that the rotation matrix can be assumed to be known. In other applications, the rotation matrix R may be estimated from a number of rotation measurements (rate or acceleration). The measurements may be collected as
W
R
=RW
L, (13)
where WL and WR are signal matrices given by
W
R
=[w
R,1
w
R,2
. . . w
R,N],
W
L
=[w
L,1
w
L,2
. . . w
L,N]. (14)
This equation may be solved by any of a variety of techniques known to those of ordinary skill in the art. For example, an unconstrained least squares solution is given by
R=W
R
W
L
T(WLWLT)−1. (15)
The solution may be constrained such that R is a pure rotation matrix. Further, the equations may be extended to enable sensor offsets to be determined if desired.
Alternatively, the rotation matrix may be found from the rotational motion signals using an iterative algorithm, such as least mean square or recursive least mean square algorithm.
The relative orientation may also be obtained by comparing gravitation vectors and magnetic field vectors, provided that the body is not rotating or is rotating only slowly.
More generally, a weighted average of the aligned signals from two or more sensors (adjusted for orientation and sensitivity) may be used to estimate the linear acceleration at a position given by a corresponding weighted average of the sensor positions when the sum of the weights is equal to one. If a is the linear motion at the position
the weighted average of aligned signals is
where Ri is the alignment matrix for sensor i, αi are weights that sum to unity, and Si is a sensitivity matrix. The vector ri−
Equation (16) is a generalization of equation (12) and describes operation of a system for monitoring motion of a substantially rigid body relative to a first location,
of the motion at the first location in the substantially rigid body. A signal representative of the motion of the substantially rigid body relative to the first location is output or saved n a memory.
The position vector r of the first location is a weighted average
of the position vectors of the plurality of second locations and the estimate of the motion at the first location comprises a corresponding weighted average
of the plurality of aligned motion vectors.
Measurements of the motion at the two sensors may be synchronized by means of a synchronization signal, such as a clock with an encoded synchronization pulse. The clock may be generated by the processor 100 or by one of the sensors. When identical sensors are used, a ‘handshake’ procedure may be used to establish which sensor will operate at the master and which will operate as the slave. Such procedures are well known to those of ordinary skill in the art, particularly in the field of wired and wireless communications.
The signals 112 and 116 together describe the motion of the rigid body and may be used to determine, for example, the direction and strength of an impact to the body. This has application to the monitoring of head impacts to predict brain injury.
In a still further embodiment, the alignment matrix R is found by comparing measurements of the gravity vector made at each sensor location. These measurements may be made by the linear elements of sensor or by integrated gravity sensors. In this embodiment one of the sensors does not require rotational sensing elements, although such elements may be included for convenience or to improve the accuracy of the rotation measurement.
In one embodiment the sensor is oriented in a known way on the rigid body. This is facilitated by marking the sensor (for example with an arrow).
Consistent positioning and orientation of the sensors may be facilitated by shaping or marking the sensor. For example, a behind-the-ear sensor may be shaped to conform to the profile of the back of an ear, or a nose sensor shaped to conform to the bridge of the nose.
In one embodiment, a portion of the patch is configured to be removable once the sensor has been positioned on the patch. This part of patch may be considered to be a template that allows accurate and consistent positioning of the patch. This may be advantageous, for example, to prevent the edge 608 from moving against the back of the ear, which may cause irritation.
More generally, the sensor elements are coupled to a mounting structure shaped for consistent orientation with respect to a characteristic feature of a substantially rigid body, and outputs linear and rotational motion signals. In a further embodiment, the mounting structure comprises a flexible band, such as 702 shown in
In a further embodiment of the invention, the head mounted sensing system is calibrated relative to a reference sensing system on a helmet, mouthguard or other reference structure. The position of the helmet on a head is relatively consistent. The positioning of mouthguard, such as protective mouth guard is very consistent, especially if custom molded to the wearer's teeth. While both a helmet and a mouthguard can be dislodged following an impact, they move with the head for low level linear and rotational accelerations. The calibration is not simple, however, since there is a non-linear relationship between the sensor signals due to the presence of centripetal accelerations. The method has application to head motion monitoring, for sports players and military personnel for example, but also has other applications. For example, the relative positions and orientations of two rigid objects that are coupled together, at least for a while, may be determined from sensors on the two bodies.
Self-calibration avoids the need to position and orient a sensor accurately on the head and also avoids the need to pre-calibrate the head sensors for sensitivity. This reduces the cost of the head sensors. A unique identifier may be associated with each helmet or mouthguard. This avoids the need for have a unique identifier associated with each head sensor, again reducing cost. Also, signals transmitted to a remote location (such as the edge of a sports field) are more easily associated with an individual person whose head is being monitored. That is, the helmet or mouthguard may be registered as belonging to a particular person, rather than registering each head sensor. Additionally, the helmet or mouthguard sensor may be used as a backup should the head sensor fail and may also detect such failure. Still further, head motion and helmet motion may be compared to each other to monitor the performance of the helmet, that is, the effectiveness of the helmet in reducing head accelerations.
The system 700 also comprises a reference sensor 118 of a reference sensing system 708 coupled to a helmet 704. The reference sensing system 708 may also include a processor, a transmitter and a receiver. A helmet 704 is shown in
In operation, the processing module operates to compute a rotation matrix R that describes the relative orientation of the head mounted sensor 102 relative to the helmet mounted sensor 118. The rotation matrix satisfies
W
H
=S
H,rot
RS
R,rot
−1
W
R, (17)
where WL and WR are signal matrices given by
W
R=[ωR,1 ωR,2 . . . ωR,N],
W
H=[ωH,1 ωH,2 . . . ωH,N]. (18)
The subscript ‘R’ denotes the reference (helmet or mouthpiece, for example) sensor and the subscript ‘H’ denotes the head mounted sensor. Since the inverse sensitivity matrix SR,rot−1 of the reference sensor is known, equation (17) may be solved in the processing module for the matrix product SH,rotR, the inverse of which is used to compute rotations relative to the frame of reference of the reference sensor. The matrix product may be estimated when the reference structure is first coupled to the head, or it may be continuously updated during operation whenever the rotations are below a threshold. Higher level rotations are not used, since they may cause the helmet to rotate relative to the head.
When a linear reference sensor is used, the gravitation vectors measured by the reference and head mounted sensors may be used to estimate the rotation matrix. The rotation matrix satisfies
G
H
=S
H,lin
RS
R,lin
−1
G
R, (19)
where GL and GR are matrices of gravity vectors given by
G
R
=[g
R,1
g
R,2
. . . g
R,N],
G
H
=[g
H,1
g
H,2
. . . g
H,N]. (20)
The gravity vectors are measured during periods where the head is stationary. Equation (19) may be solved for the matrix product SH,linR. Similarly, sensed magnetic field vectors may be used, alone or in combination with other signals, to determine orientation, as discussed below.
The acceleration at the head mounted sensor may be written as
s
H
=S
H,lin
R[a−K(rRH){dot over (ω)}+P(rRH)γ(ω)], (22)
where a is the acceleration vector at the reference sensor. Since the rotation vectors are known (from the head mounted sensor and/or the reference sensor) equation (22) may be solved in the processing module to estimate the position vector rRH of the head mounted sensor relative to the reference sensor. Additionally, if the position center of the head is known relative to the reference sensor on the helmet, the position of the head mounted sensor may be found relative to center of the head.
The orientation can be found from the rotational components. If the linear and rotation sensing elements are in a known alignment with one another, the orientation of the linear sensing elements can also be found. Once the orientation is known, either predetermined or measured, the sensitivity and positions of the linear elements can be found. The output from a sensing element is related to the rigid body motion {a,{dot over (ω)},ω} by
s
i
=g
i
−1ηiT[a+K({dot over (ω)})ri+K2(ω)ri], (23)
where ηiT is the orientation and gi−1 is the sensitivity. In matrix format, the relationship may be written as
An ensemble averaging over a number of sample points provides as estimate of the inverse sensitivity of the sensing element and the position of the sensing element as
where the matrix A is given by
A=[s
i−ηiT{K({dot over (ω)})+K2(ω)}]. (26)
Thus, the position and sensitivity of the sensing element may be determined from the sensor output si and the measured rotation, once the orientation is known.
The sensor orientation may be determined (a) by assumption (b) from gravity measurements (c) from rotation measurement and/or (d) from rigid body motion measurements, for example. Once the orientation is known, the sensitivity and position may be determined from equations (25) and (26) above.
If several sensing elements are positioned at the same location, their positions may be estimated jointly. Equation (24) can be modified as
or, in matrix form,
Once calibrated, the acceleration at the origin (the center of the head for example) may be found using
a=[S
H,lin
R]
−1
s
H
+K(rH){dot over (ω)}−P(rH)γ(ω), (29)
where rH is the position of the head mounted sensor relative to the origin. This computation uses the inverse of the matrix product SH,lin R, so separation of the two matrices, while possible, is not required.
Thus, a reference sensor on the mounted on a reference structure, such as a helmet or mouthguard, may be used to determine the orientation and position of the head mounted sensor, together with its sensitivity. This is important for practical applications, such as monitoring head impacts during sports games or for military personnel, where accurate positioning of a head mounted sensor is impractical, and calibration of the head mounted sensors may be expensive.
The helmet 704 may support one or more visual indicators such as light emitting diodes (LEDs) 706 or different colors. These indicators may be used to show the system state. States could include, for example, ‘power on’, ‘head sensors found’, ‘calibrating’, ‘calibration complete’ and ‘impact detected’. In one embodiment, an impact above a threshold is indicated by a flashing red light, with the level of the impact indicated by the speed of flashing.
In one embodiment, the head motion is only calculated or output when motion is above a threshold level and calibration is only performed when the motion is below a threshold.
The performance of a protective body, such as a military, sports or motorcycle helmet, or a shipping container, may be determined by comparing the motion of the protective body (e.g. the helmet or container) with the motion of the protected body (e.g. the head or shipped item). For example, linear and rotational accelerations may be compared. If the acceleration of the protected body is much lower than that of the protective body, the performance is good. If the acceleration of the protected body is similar to that of the protective body, the performance is poor.
The signal from a linear accelerometer is a combination of linear and rotational motion, as described in equation (1), and is dependent upon the chosen origin. In order to compare signals from a head accelerometer and a helmet accelerometer, the sensors should be referenced to a common frame of reference, including a common origin. This requires knowledge of their relative positions and orientations. However, the magnitude (not direction) of the signals can be compared if the relative positions are known. The signals from rotational sensor may be compared if their relative orientations are known. However, the magnitude of rotational motion may be compared directly.
In one embodiment performance of a helmet on a head is monitored by determining linear motion of the head relative to a common origin in response to sensed head motion, determining linear motion of the helmet relative to the common origin in response to sensed helmet motion, and comparing the determined linear motion of the head with the determined linear motion of the helmet to provide a measure of the performance of the helmet. The linear motion of the head relative to the common origin is determined by receiving first sensor signals from a head-mounted sensor, receiving second first sensor signals from a helmet-mounted sensor, determining a location of the head-mounted sensor in response to the first and second signals, and determining linear motion of the head relative to the common origin dependent upon sensed linear motion of the head, sensed rotational motion of the head and the determined location of the head-mounted sensor. Optionally, motion directions of the head and helmet may be compared in a common frame of reference by determining an orientation of the head-mounted sensor with respect to a common frame of reference in response to the first and second signals, determining linear motion of the head relative to a common frame of reference in response to the sensed head motion, determining linear motion of the helmet relative to the common frame of reference in response to the sensed helmet motion, and comparing the linear head motion in the common frame of reference with the linear helmet motion in the common frame of reference to determine performance of the helmet.
Rotational motions may also be compared by determining rotational motion of the head in response to the sensed head motion, determining rotational motion of the helmet in response to the sensed helmet motion, and comparing the rotational head motion with the linear helmet motion to determine performance of the helmet. The determined linear motion of the head, the determined rotational motion of the head, the determined linear motion of the helmet and the determined rotational motion of the helmet may have a common frame of reference.
The location and orientation of the head-mounted sensor may be found in response to the first second signals received from a head-mounted sensor and second sensor signals received from a helmet-mounted sensor. In addition, one or more sensitivities of the head-mounted sensor may be determined in response to the first and second sensor signals.
In one embodiment, performance of a helmet on a head is monitored by determining head motion relative to a common frame of reference in response to sensed motion of the head, determining helmet motion relative to the common frame of reference in response to sensed motion of the head, and comparing the head motion in the common frame of reference with the helmet motion in the common frame of reference to determine performance of the helmet. Determining motion in a common frame of reference from a sensor measurement uses knowledge of the position and orientation of the sensor in the common frame of reference.
The common frame of reference may be a helmet frame of reference, in which case determining the head and helmet motions relative to the common frame of reference comprises transforming sensed head and helmet motions to the helmet frame of reference. The location, orientation and sensitivity of the one or more helmet sensors may be predetermined, while the location, orientation and sensitivity of the one or more head sensors may be determined (as described above, for example) from measurements at lower acceleration levels, where the helmet motion is substantially the same as the head motion.
The common frame of reference may be a frame of reference of one or more helmet-mounted sensors, in which case determining head motion relative to the common frame of reference comprises transforming sensed head motion to the frame of reference of the one or more helmet-mounted sensors.
Head motion relative to the common frame of reference may be found by determining head rotational motion from one or more head-mounted sensors, determining helmet rotational motion from one or more helmet-mounted sensors, and determining an orientation of the one or more head-mounted sensors relative to the common frame of reference from the head rotational motion and the helmet rotational motion.
Head linear motion may be found from the one or more head-mounted sensors by determining helmet linear motion from the one or more helmet-mounted sensors, and determining a position of the one or more head-mounted sensors relative to the common frame of reference from the head linear motion and the helmet linear motion.
Helmet motion relative to the common frame of reference in response to sensed helmet motion may be found by sensing helmet rotation rate using a gyroscope to provide a first estimate of rotation rate. The gyroscope offset bias may be found by sensing helmet orientation using a magnetic field sensor and a gravitational field sensor to provide a second estimate of rotation rate from the sensed helmet orientation. The bias offset may be found by comparing the first and second estimates of rotation rate. The resulting bias offset may be subtracted from the first estimate of rotation rate to provide a rotational component of the helmet motion. In one embodiment, a correction to the bias offset is found by passing the difference between the first and second estimates of rotation rate through a lowpass filter.
In one embodiment, head motion, relative to the common frame of reference and in response to sensed head motion, is found by determining linear and rotational head motion at an origin of the common frame of reference in response to linear and rotational head motion sensed by a head-mounted motion sensor at a location displaced from the origin of the common frame of reference. The location of the head-mounted motion sensor may be estimated from the sensed head motion and the sensed helmet motion, as described above.
Helmet motion, relative to the common frame of reference and in response to sensed helmet motion, is found by determining linear and rotational helmet motion at an origin of the common frame of reference in response to linear and rotational helmet motion sensed at location displaced from the origin of the common frame of reference.
ω=F(w), {dot over (ω)}=G(w). (30)
This transformation depends upon the orientation of the sensor relative to the common frame of reference. A measured linear acceleration s, is transformed to a linear acceleration a in the common frame of reference according to
a=S
lin
−1
s+K(r){dot over (ω)}−P(r)γ(ω), (31)
As described above in equation (7) above, for example. These transformations depends upon the orientation and location (r) of the sensor relative to the common frame of reference, as well as the rotational motion. At block 1014, the head and helmet motions in the common frame of reference are compared to provide an indication of helmet performance. The helmet performance is reported at block 1016. For example, the helmet performance may be saved in a transient or non-transient computer memory, displayed on a screen, communication to a remote location or any combination thereof. At block 1018, a check is made to determine if the helmet has been removed. This may be done, for example, by monitoring a strength of a wireless communication signal between a transceiver in the helmet and a transceiver of head sensor. If the helmet has not been removed, as depicted by the negative branch from decision block 1018, flow returns to block 1010. Otherwise, the method terminates at block 1020.
The common frame of reference may be a helmet frame of reference or a frame of reference of a helmet mounted sensor. The orientation of a head mounted sensor in the frame of reference of a helmet sensor may be found by sensing orientation of both the helmet sensor and the head sensor to an Earth frame of reference (using a tilt sensor and an electronic compass, for example). However, location of the head sensor with respect to the original may also be needed.
The above method provides for in-situ monitoring of helmet performance and head impacts. The same method may be applied to a helmet mounted on a dummy head to facilitate laboratory testing of helmets, or to monitoring an item in a shipping container.
The system 700 of
The processor may be further operable to determine an orientation of the second motion sensor dependent upon the first and second motion signals or to determine a location of the second motion sensor dependent upon the first and second motion signals. The helmet sensing system and the processor may be located in a common housing. The first and second motion signals may be synchronized in time.
In one embodiment, the power source comprises a resonant electromagnetic circuit adapted to receive power from a remote source.
A view through the section 12-12 is shown in
An advantage of the embodiment shown in
Alternative embodiments may use re-chargeable batteries coupled to the sensing circuit or a circuit configured to receive power from a remote source, using resonant electromagnetic coupling for example. In the embodiment shown in FIG's 15 and 16, the cavity 1120 in the mounting structure 604 opens to the underside of the sensor patch 102. Connection pads 1116 and 1118 on the top of sensing circuit 602 electrically couple to the terminals 1108 and 1110 of battery 1106 when the sensing circuit is inserted into the cavity 1120. In this embodiment, an antenna 1122 is embedded in or coupled to the mounting structure 604 to enable wireless communication, wireless power reception or both.
The exemplary embodiments described above provide a sensor 102 that may be used for monitoring motion of a body. The sensor includes a flexible mounting structure 604 having a cavity 1120 and an underside 1102 configured to enable coupling to the body. The sensor also includes a power source, such as battery 1106, and a motion sensing circuit 602 sized to be easily inserted into and removed from the cavity 1120 of the mounting structure 604 through an opening 1112 in the flexible structure. The motion sensing circuit 602 is adapted to receive power from the power source and provides motion signals in response to motion of the body to which it is coupled.
The sensor 102 may include a removable template 604″, as shown in
The sensor may also include a wireless communication circuit that is physically coupled to the motion sensing circuit 602 and includes an antenna 1122 that extends from the cavity 1120 of the flexible mounting structure when the motion sensor circuit is inserted into the cavity of the mounting structure. Battery 1106 may be omitted when wireless power is provided.
Various other configurations which allow for reuse of the sensing circuit will be apparent of those of ordinary skill in the art. For example, the adhesive layer 1102 may be replaced with double-side tape.
Some aspects and features of the disclosed sensor are set out in the following numbered items.
1. A sensor for monitoring motion of a body, comprising:
Rotation rate sensors, such as gyroscopes, often have a bias offset. In the descriptions above it is assumed that the bias offsets have been removed. In accordance with one embodiment of the disclosure, the bias offsets may be remove by use of a magnetic field sensor (electronic compass) and accelerometer (gravitational filed sensor) in a known orientation to the gyroscope.
When dynamic acceleration is small, the absolute orientation of the sensor relative to the Earth can be found from the magnetic field vector m and the acceleration vector g. The rotation can denoted by the rotation matrix
R(t)=[nwu]T, (32)
where the vectors are calculated as, for example,
The rotation rate matrix is related to the rotation matrix R and its time derivative {dot over (R)} as
so the rotation rate vector may be calculated as, for example,
This estimate of the rotation rate is generally noisy, due in part to stray magnetic fields. The raw gyroscope signal ωg(n) at sample n is less noisy but has a variable bias offset. The vector of offsets, c(n), can be updated according to, for example
c(n)=(1−μ)c(n−1)+μL(q−1)(ωg(n)−ωc(n)), (36)
where q−1 is the delay operator, L(q−1) denotes a lowpass filter and μ is an update step size (which may be reduced when dynamic acceleration is detected). The corrected rotation rate may be computed as
{circumflex over (ω)}(n)=ωg(n)−c(n). (36)
This, in accordance with certain aspects of the disclosure, a gyroscope bias is removed by:
Other techniques for using an electronic compass and tilt sensor to determine absolute gyroscope bias offset will be apparent to those of ordinary skill in the art. For example, time periods during which there is no rotation may be detected to enable direct measurement of the offset vector.
The rotation rate vector may be integrated to obtain a dynamic estimate of head or helmet orientation. Such integration is subject to drift due integration of random noise and imperfect offset removal. The drift may be compensated for by using an electronic compass and tilt sensor. Compensation techniques are well known to those of skill in the art. Head and helmet orientations may be reported to enable graphical display of an impact. Further, head or helmet orientations, or both, and acceleration directions from different players may be combined to create, display or analyze a model of an impact between two or more players. The players may be identified by finding impacts that occur at the same time.
In one embodiment, sampling of the signals from head and helmet sensors is synchronized by transmitting a common timing signal to sampling circuits of both sensors. The timing signals may originate at one of the sensors or at a remote device.
One application of the disclosure relates to a method for monitoring performance of a shipping container in protecting a shipped item within the container from impacts. Motion of the shipped item is determined relative to a common frame of reference in response to sensed motion of the shipped item, and motion of the shipping container is sensed relative to the common frame of reference in response to sensed motion of the shipping container. The motion of the shipped item in the common frame of reference is compared with the motion of the shipping container in the common frame of reference to determine performance of the shipping container. The common frame of reference may be an Earth frame, a frame of the shipping container, or a frame of a motion sensor mounted on the shipping container.
In a further embodiment, the systems and methods described above may be used to monitor impacts between bodies. The bodies may be head or helmets of players on opposing sports teams for example. Impact motion of the first body relative to a common frame of reference is determined in response to sensed motion of the first body, and impact motion of the second body relative to the common frame of reference is determined in response to sensed motion of the second body. Simultaneous impact motion of the first and second bodies from the impact motions of the first and second bodies is detected so as to identify which players are involved in the impact. Finally, the simultaneous impact motions of the first and second bodies are compared. In addition, the orientations of the first and second bodies, relative to the common frame of reference, may be determined from rotation rate signals or orientation signals. This enables the relative positions of the first and second bodies during the detected simultaneous impact motion to be determined and facilitates analysis, display and reporting of the impact. The positions may be found, for example, by determining the linear accelerations of the two bodies (using equation (7) for example, with sensed linear and rotational motions), in a common frame of reference, relative to an origin, such as the approximate center of mass of each body, and aligning the linear acceleration vectors of the two bodies.
In an embodiment where the first and second bodies are helmets, head motion may also be monitored during the detected simultaneous impact to enable helmet performance to be monitored, as described above.
In the foregoing specification, specific embodiments of the present invention have been described. However, one of ordinary skill in the art appreciates that various modifications and changes can be made without departing from the scope of the present invention as set forth in the claims below. Accordingly, the specification and figures are to be regarded in an illustrative rather than a restrictive sense, and all such modifications are intended to be included within the scope of the present invention. The benefits, advantages, solutions to problems, and any element(s) that may cause any benefit, advantage, or solution to occur or become more pronounced are not to be construed as a critical, required, or essential features or elements of the invention.
This application claims priority from Provisional Application Ser. No. 61/519,354, filed May 20, 2011, titled “Method and Apparatus for Monitoring Rigid Body Motion in a Selected Frame of Reference”, which is hereby incorporated by reference herein, and application Ser. No. 13/506,766, filed May 16, 2012, titled “Method and Apparatus for Monitoring Motion of a Substantially Rigid Body in a Selected Frame of Reference”, which is hereby incorporated by reference herein.