Field of the Invention
The present invention relates generally to analyzing a person's movement, such as gait. More particularly, the present invention relates to a method and a system to record synchronized 3D kinematic and kinetic data as a person ambulates.
Related Art
Current gait lab configurations utilize a force plate that is built into the floor or otherwise requires a dedicated room and considerable time and expertise to set up. The force plate and set-up time can be expensive. In addition, the force plate constrains the patient to a straight line of motion or walking.
It has been recognized that it would be advantageous to develop a method and an apparatus, or a gait lab, for determining, documenting and improving the functional capability of a patient that is quickly and easily set up, portable, inexpensive, and that records full body kinematic (visual) and kinetic (load) data in 3D, and that does not constrain the patient.
The invention provides a mobile gait lab system comprising a pair of force sensors removably affixed to a person's ankles, feet, shoes or lower limb prostheses to measure kinetic data as the person ambulates. Each of the pair of force sensors comprises a four-axis ground reaction sensor configured to sense: 1) pressure force or vertical force, 2) anterior/posterior shear force, 3) medio/lateral shear force, and 4) torque or moment exerted between the person's ankles, feet, shoes or lower-limb prostheses and a support surface. In addition, the system comprises a plurality of video cameras to record markerless kinematic data as the person ambulates. The plurality of video cameras is removably disposed about a predetermined spatial volume. Furthermore, the system comprises a computer with one or more processors configured to temporally synchronize the kinetic data and the kinematic data together.
In addition, the invention provides a method for analyzing movement of a person, the method comprising: 1) measuring 3D kinetic data as the person ambulates with a pair of force sensors affixed to a person's ankles, feet, shoes or lower-limb prostheses with each of the pair of force sensors on a different one of the person's ankle, feet, shoes or lower-limb prostheses; 2) obtaining 3D kinematic data as the person ambulates with at least one video camera that video records the person ambulating without markers; and 3) temporally synchronizing the kinetic data and the kinematic data together.
Furthermore, the invention provides a system for analyzing movement of a person. The system comprises a pair of force sensors to be affixed to a person's ankles, feet, shoes or lower-limb prostheses to measure kinetic data as the person ambulates; at least one video camera configured to record markerless kinematic data as the person ambulates; and one or more processors configured to temporally synchronize the kinetic data and the kinematic data together.
Additional features and advantages of the invention will be apparent from the detailed description which follows, taken in conjunction with the accompanying drawings, which together illustrate, by way of example, features of the invention; and, wherein:
Reference will now be made to the exemplary embodiments illustrated, and specific language will be used herein to describe the same. It will nevertheless be understood that no limitation of the scope of the invention is thereby intended.
Definitions
Standard anatomical terminology is used herein, specifically with reference to
Ambulation and gait cycle terminology includes a stride and a step, as shown in
The term “transceiver” is used herein to refer to a transmitter for transmitting a signal, a receiver for receiving a signal, or both a transmitter and a receiver. The transceiver can both send and receive, or can include a transmitter for transmitting a signal, and a receiver for receiving a signal.
Description
The present invention comprises methods and systems for capturing 3D kinematic (visual) and kinetic (load) data about a moving person, combining those data streams with software, analyzing the movement for functional benefit, and producing a report which quantifies the physical dynamics over a variety of activities. In addition, the invention comprises methods and systems for determining, documenting, and improving the functional capability of a patient with a markerless motion capture and force sensing system. The methods and systems enable an investigator to quickly set up and record full body kinematic and kinetic data, and produce clinically relevant reports that aid in optimization and documentation of patient movement. The invention provides a comprehensive method to take full-body quantitative data, record the position and orientation of each limb in a multi-segment system and the forces that linked segment model exert on the surrounding environment. In one aspect, the markerless motion-capture system can be accompanied by force sensing modules in proximity to the foot. When kinematic (visual) and kinetic (load) data are captured together, highly accurate calculations that quantify movement and joint loading are possible. These calculations can be referred to as Inverse Dynamics. When a complete data set of patient movement is available, advanced analysis techniques can use this data with a forward simulation system to quantify the effect of each joint torque on every other limb segment. These mathematical techniques can be referred to as Intersegmental Dynamics Analysis (IDA). The IDA produces a result that details the accelerations produced by joint torques, which can be referred to as Acceleration Analysis (AA). In one aspect, the method and system can further include obtaining information about the patient's physical properties, such as height, weight, mass distribution, and body fat. The method and system can then adjust a patient-specific table that enables calculation of limb lengths, mass properties and joint centers, to enable more accurate Inverse Dynamics.
In one aspect, the present invention comprises a movement or gait lab to quantify and document the functional level of an amputee. When the patient is an amputee, the patient-specific table of properties can include information on the physical properties of the prosthetic device(s). Functional level can be categorized by multiple activities or standards, such as the capacity for variable cadence, capability to perform activities of daily living (ADLs), or time required to perform certain bench-mark tasks, such as the L Test or Timed Up and Go (TUG) Test.
The method and system can comprise Data Capture Software; Biomechanics Software; Joint Torque (& Power) Calculation; Forward Dynamics Simulation; Acceleration Analysis Software; and Capability Reporting Software. Prior work in gait labs that only use kinematic (visual) data, and that did not include capturing the forces a subject exerts on the environment, compute those forces from acceleration with methods that contribute error to the movement data. Prior work that only captured kinetic data, or forces associated with a limb do not completely depict a subject due to lack of knowledge of the position of limb segments, and most commonly only capture data from one side of the subject, given the example of lower extremity devices. Wrist- or ankle-worn accelerometer-based devices have difficulty distinguishing between activity, and are not able to quantify movement or loading of a specific limb segment, as would be needed for an amputee subject.
The method and system enables a new methodology and a new price point of data analysis and reporting on the capability of a person to perform movements while, at the same time, becomes a system that can hide the complexity from the user by turning large data sets into actionable information.
As illustrated in
The system 10 and/or the mobile movement/gait lab can comprise a plurality of markerless video cameras 18 to capture and record markerless video (kinematic) data of the person as the person ambulates. The plurality of markerless video cameras 18 can be removably disposed about a predetermined spatial volume 22, such as a room or predetermined space. The cameras can be disposed about a perimeter of the volume, and can define the volume or perimeter. The video cameras can be positioned and located to obtain different perspectives or angles of the volume or space. In one aspect, the video cameras can be disposed on stands. In another aspect, the video cameras can be mounted to existing walls. The kinematic data capture system or hardware, or the plurality of video cameras, can be single lens USB cameras, without sensors for sensing video or other markers on the person, or depth sensors (infrared or laser) as used in other types of the system. The person and the person's limbs can be distinguished using computer software, as discussed in greater detail below. Thus, the system and the gait lab can be set up and used with greater ease and speed.
In the preferred embodiment, a markerless motion capture system provides the visual, or kinematic, data to the system. The markerless video cameras do not require knowledgeable and expensive staff to undergo the time consuming process of placing identifying marker balls of LED lights on a subject.
The video cameras can be cheap, robust, high-quality and high-frame rate video cameras. For high-quality motion capture, the cameras can be 60 Hz and can provide a 640×480 capture video. For one preferred embodiment, the cameras can be single lens cameras and can be the USB connected SLEH-00448, or the SLEH-00201, both from Sony.
The cameras can be combined with software packages that are capable of capturing 2D video from two or more viewpoints, and rendering those recordings into a 3D stick figure that represents the long bones of the human body. For example, the iPiSoft system can be used that includes the iPi Recorder, iPi Studio, and iPi Biomechanics export. The general methodology is to take a background image of each viewpoint, record a multiple vantage video section with a subject starting in a known orientation, and then match the subject movement to an animated model by subtracting the background at each frame, and matching the model to the 2D image, from each viewpoint. The work flow typically follows such a progression: video of empty capture volume 22, subject movement video 14, lumped massed (visual hulls) 274, and multi-segment 3D virtual model/skeleton stick
In one aspect, each recording session can start with a patient wearing the sensors 30, and standing on a vertical force scale 150 connected to the computer 26 for baseline calibration 142. The user can be asked to stand momentarily with both feet on the scale (indicated at 144 in
The initial skeleton model 266 used in these systems can consist of a 16-segment human made of 1 Head, 1 Thorax, 1 Abdomen, 1 Pelvis, 2 Thigh, 2 Shank (Leg), 2 Feet, 2 Upper Arms, 2 Lower Arms, and 2 Hands. Multiple other initial models could be used in other embodiments, with more segments 270 added for more detail, at the potential cost of more time to calculate and convert the multiple 2D movement video streams to a 3D model movement.
The methods may further include, via the one or more computers, receiving physical information (indicated at 344 in
In addition, the system 10 and/or the mobile movement/gait lab can comprise a computer 26 with one or more processors. The computer 26 can comprise a laptop computer, a tablet computer, a desktop computer, a modified laptop computer with additional ports and/or wireless capabilities, a modified tablet computer with additional ports and/or wireless capabilities, multiple computers, etc. The computer can have one or more wireless transceivers (e.g. Bluetooth and/or wireless local area networks (LAN) transceivers), and/or one or more data ports (e.g. USB ports). The computer can have an internal and/or an external data storage device (e.g. disk drives or flash memory). The computer can have a visual display, such as a monitor, and data input, such as a keyboard or touch screen. The computer can have one or more (fixed or base) processors 28, as discussed in greater detail below. In one aspect, the plurality of video cameras can be wired-ly coupled to the computer, such as through USB ports (or a USB bus with a plurality of ports). In another aspect, the plurality of video cameras can be wirelessly coupled to the computer.
Furthermore, the system 10 and/or the mobile movement/gait lab can comprise a pair of force sensors 30 removably affixed to a person's ankles, feet, shoes or lower limb prostheses (all indicated at 34) to measure force and shear (kinetic) data as the person ambulates. As stated above, the term “a person's ankles, feet, shoes or lower-limb prosthesis” is intended to encompass any combination of ankle, foot, shoe or prosthesis. The sensors can be foot-borne sensors. The sensors can be affixed to each of the person's feet/foot, ankle(s), shoe(s), prosthetic(s), or combinations thereof depending upon the person. For example, a pair of sensors can be affixed to the person's respective ankles, shoes or feet. As another example, a pair of sensors can be affixed to the person's respective prosthetic and ankle, shoe or foot. Thus, each lower limb (foot, shoe, ankle or prosthetic) of the person has a sensor affixed thereto. The sensors 30 measure force and shear (kinetic) data as the person walks. In one aspect, the sensor can be a four-axis ground reaction sensor configured to sense: 1) pressure force or vertical force, 2) anterior/posterior shear force, 3) medio/lateral shear force, and 4) torque or moment exerted between the person's ankles, feet, shoes or lower limb prostheses and a support surface 38. The support surface can be the floor of the space. In another aspect, the sensor can be a three-axis ground reaction sensor configured to sense: 1) pressure force or vertical force, 2) anterior/posterior shear force, and 3) medio/lateral shear force.
Each sensor 30 can include a housing 42 and an attachment 46 coupled to the housing to attached the housing to the person's ankles, feet, shoes or lower limb prostheses. In addition, the housing or sensor can have a battery port 46 with an opening 50 through the housing and into the battery port. A rechargeable battery 54 can be insertable through the opening 50 and into the battery port 46. Thus, the battery 54 can be removed from the sensor or housing and disposed on a charger. The sensor 30 can include electronics housed in the housing 42. The sensor 30 and/or the electronics can comprise a sensor 58, the battery power source 54, a wireless transceiver 62, a digital memory device 66, and one or more (mobile) processors, which can all be disposed in the housing. In one aspect, the (mobile) processor 70 can be electrically and operatively coupled to the sensor 58, the power source 54, the wireless transceiver 62, and the digital memory device 66. In another aspect, the (mobile) processor 74 can include the wireless transceiver 62 and/or the digital memory device 66 and/or the sensor 58, and can be electrically and operatively coupled to the battery power source 54. The term “sensor” is used broadly herein to refer to both the sensor unit 30 including the housing, etc., and the sensor 58 itself as an electrical component. The one or more mobile processors 70 or 74 can be disposed in the housing 42, and coupled to one of the pair of force sensors 58. The one or more mobile processors can be configured to receive force signals, convert analog signals to digital signals, filter the signals, amplify the signals, condition the signals, compensate the signals and/or calibrate the signals. The wireless transceiver 62 is disposed in the housing, and coupled to the one or more mobile processors 70 or 74. The one or more mobile processors can be configured to produce a time-stamped, left- or right-side identified, seven-axis vector quantity. In addition, the one or more mobile processors and the wireless transceiver can be configured to transmit the vector quantity at 120 Hz or greater in one aspect, or 60 Hz or greater in another aspect.
One previously mentioned method to synchronize the visual and force data from different wireless or wired sensors is an event in the data file itself that creates a spike in both data streams, such as bouncing on the toe with the leg extended fully behind. A second, and more preferred method, is to synchronize the wireless transmitters without the need to create an action on the part of the subject. One method to do this is the TPSN protocol, or a Timing-Sync Protocol for Sensor Networks, consisting of a level discovery phase and a synchronization phase, given that two communications between the levels and nodes are present. With one root node assigned in the network, a time signal can be sent from Node A to Node B in a level discovery phase, but subsequently all nodes can keep track of time, with Node A sending a time-stamped packet to Node B, B to C, C to D, and so on to N nodes to complete the next work. The time-stamped packet of each node's time can then be sent backwards up the chain to double the difference between each one, or create a time-stamped loop which details the lag or clock difference between each node. In this manner, the four main delay times (send time, access time, propagation time, and receive time) are accounted for during the synchronization process.
For full-body, multi-limb segment gait analysis, a 3D vector of the Ground Reaction Force (GRF) can be used for accurate Inverse Dynamics. The force and moment measurement system, in one embodiment, comprises an insole sensor with wireless capability, and at least one body-worn computer able to wirelessly receive information from the sensor. In another embodiment, the sensor can be placed on the outside of a shoe for ease of use, and to enable cleaner signal generation to describe the foot-floor interaction. The force sensor, in a preferred embodiment, can sense plantar pressure force, anterior/posterior shear force, and medio/lateral shear force. As well, the preferred sensor may sense if a torque, or moment, is exerted between the foot and floor in the plane of the ground. As well, the preferred sensor may have sensing at multiple separate locations along its length to enable sensing the complete foot-floor interaction when only a part of the foot is on the floor, such as during normal ambulation at Initial Contact (IC) when only the heel is on the ground, or at Heel Off (HO) when only the toes are on the ground.
The sensor can easily be mounted on left or right human feet independently, and can record and transmit in real time a 3-axis force of ground contact, and a single axis moment of ground contact, to a nearby Windows desktop computer. The Center of Pressure (CoP) of the 3-axis Ground Reaction Force (GRF) can be located within the area of the sensing element to enable high-quality inverse dynamics, either by sensing element, or by row and column number. Vertical force, Anterior-Posterior shear force, and Lateral shear force are all needed when any partial section of the foot is contacting the ground. The sensor can be an integrated unit that senses vertical pressure, shear, and moment, or a dual layer approach that utilizes separate sensing elements or types. Minimizing the instep cross talk of ground force-induced shear to substrate bending shear will be desired.
An insole pressure sensor, such as those from TekScan or others, can be used, but such sensors are capable of creating a force vector of one dimension, reporting only pressure, or a force normal (perpendicular) to the sensing mat. A plurality of pressure sensors can be placed on the forward section of the insole, and a separate plurality of sensors can be placed on the heel portion of the sensing device, such that a single Center of Pressure can be created for the Ground Reaction Force (GRF) no matter what partial section of the foot is loaded when in contact with the ground. Sensors in the heel can sense load at Initial Contact (IC), and sensors under the toes can sense ground loading during Heel Off (HO). The sensor can measure loading from the occurrence of Heel Strike (HS), through the entirety of stance phase, and until the instant of Toe Off (TO).
The pressure sensors may include materials that deflect in some manner, and such deflection can affect the materials' electrical resistance, and thus, cause a change in voltage.
A second layer of sensing, either integrated with the first or separate, that is capable of quantifying the anterior/posterior and medio/lateral shear forces associated with the foot can be used to create a fully accurate 3D ground reaction force vector. A single shear sensor located at the mid-foot can add accuracy to the 3D nature of the ground reaction force. Further, an even more accurate sensor can have shear sensing at the heel and at the forefoot (metatarsal heads), so that shear is accurately quantifiable during loading periods when the foot is not fully on the ground, between heel strike and foot flat (FF), and again between foot flat and toe off (TO). At heel strike, a shear sensor located at the mid-foot would not be loaded, and therefore would not sense shear forces. As well, a mid-foot location for a shear sensor would not be loaded from terminal stance to toe off, as the heel comes off the ground.
The pressure and shear sensors can be connected to the ankle worn electronics via flexible conductors. The sensor electronics may include, at least, an integrated circuit as a main processor, at least one analog-to-digital converter to receive force signals, a removable, rechargeable battery, and a wireless component to enable communication with the computer.
The sensor can include an A to D converter, microprocessor, and storage device in the form of memory. The electronics can provide for filtering of the signal, amplification, and conversion to digital format. The sensor can include various electronic components for signal conditioning, temperature compensation, analog-to-digital conversion, digital-to-analog conversion and a serial communication bus. Additionally, sensor calibration data can be stored in an onboard memory. The memory can be any volatile or non-volatile memory, such as, but not limited to, PROM, EEPROM, Flash, DRAM, eDRAM, and PROM. The memory can hold sensor calibration data.
The kinetic data can be transmitted wirelessly at a minimum of 60 Hz in one aspect, and at a minimum of 120 Hz in another aspect. For an exemplary embodiment involving sensing the left and right feet, each wireless transmitter can send a time-stamped 7-axis vector quantity (identifiable as left or right) to the receiving computer with less than a half-second delay, resulting in a 7×1 vector (t, GRFx, GRFy, GRFx, Mz, CoPr, CoPc) for each foot. The time stamp of the ankle transmitting modules can be synchronized to the kinematic data on the receiving computer at least at 10× the rate of the data transfer. Thus, if the recording was at 100 Hz, or every 0.01 sec, the time stamp synchronization can be done at 1000 Hz, for an accuracy of 1 millisecond. The wireless transmitting system can be any one of a number of well-known short range radio frequency protocols, supported by commercially available hardware developed for the purpose, such as Zigby, Near Field Communication (NFC), Local Wireless Networks, or Bluetooth.
The sensors and electronics can last at least one year, five days per week usage case of three hours a day of walking. In addition, the sensor can only need to be calibrated once every hour without more than a 2% drift in zero values, or maximum range. The use of removable batteries is useful for patient freedom, and uninterrupted workflow. It is intended that the instrumented patient can be quantified for at least an hour without the need to recharge batteries. Removable batteries are a product benefit on the regulatory side, as the mobile user can never be attached to a “mains-connected” charger. The removable batteries can be 3.7 V Li batteries in the 750 mAhr range.
As indicated above, the system 10 and/or computer 26 can have one or more (fixed or base) processors 28 configured to temporally synchronize the kinetic data from the sensors 30 (or 58) and the kinematic data from the video cameras 18 together. Thus, the kinematic and kinetic data can be viewed and analyzed together. In addition, the kinematic and kinetic data can be saved together as a data file. Furthermore, the kinematic and kinetic data can be presented together in one or more reports to quantify and/or qualify a person's gait.
A method 100 (
In addition, the method 100 also includes obtaining 164 3D kinematic data as the person walks with at least one markerless video camera 18 that video records the person walking without markers. In one aspect, the method can include arranging 178 a plurality of video cameras 18 at a predetermined spatial volume 22. The plurality of video cameras 18 can be calibrated 192 by recording an initial predetermined pose of the person, such as a T-pose in
Furthermore, the method 100 includes temporally synchronizing 206 the kinetic data and the kinematic data together. In one aspect, synchronizing the data can include causing 220 a shock load scenario to create an impulse spike in both the kinetic data and the kinematic data. In one aspect, a method for proper synchronization of kinetic and kinematic data can include a standing calibration test where the person under test undergoes a shock-loading scenario that creates an impulse-like spike in both kinetic and kinematic data. In one aspect, the person can extend a leg behind him or her while quietly standing, and stubbing a toe on the ground to create a synchronization event. In addition, the initial calibration of the kinetic system can involve standing on a linked scale 150, weighing the person or patient, asking for the left foot to be fully extended behind the person, and capturing what the sensors in proximity to the foot and the vertical scale record at the same time for calibration by the computer. Previous knee and hip studies (E. J. Alexander, Gait & Posture, 1997) reported between 11% (knee) and 18% (hip) error in calculated moment with a 1-frame shift of 120 Hz kinetic and kinematic data. Thus, there is a need for proper temporal synchronization between the two data streams. This temporal synchronization can be established either with on-board electronics on the ankle worn devices and the receiving computer, or with actions by the subject within the recording before the investigated activity, such as tapping one toe on the ground while the leg is extended behind.
In addition to temporal synchronization error, another problem is orienting the in-shoe sensed ground reaction force, nominally aligned with the Foot Segment Coordinate System (FSCS), to the Global Lab Space Coordinate System (GLSCS). In one aspect, as the foot can only produce force when on the ground, the 3D force vector reported by the in-shoe sensing could be aligned with the lab space (GLSCS) by assuming that each axis is parallel when force is produced. While that is a good assumption for the majority of the foot flat gait cycle, and the majority of time a GRF could be produced, this ignores the force generation during Initial Contact (IC) and Heel Off (HO). There is force generation during this time when the heel is in contact with the ground, before the foot segment coordinate system is in alignment with the GLSCS, which happens at foot flat. Common foot angles at heel strike can range from 10 to 15 deg, resulting in a source of error in the reported GRF if foot (FSCS), and lab space (GLSCS) alignment is assumed. This transitory response period can be minimal, and forces can be low, so error can be quantified and determined to be acceptable or not. In another aspect, the method 100 can include orienting 234 a foot segment coordinate system of either of the pair of force sensors 30 or 58 and the patient's feet, with a foot segment coordinate system of the lab space of the video cameras 18. In one aspect, orienting the coordinate systems can include aligning a foot space vertical force component vector reported by one of the pair of sensors 30 or 58 with a lab space vertical axis created by the video camera, and potentially error correcting by assuming the force vector and vertical axis are parallel during flat foot while ignoring force generation during initial contact and heel off. In another aspect, orienting the coordinate systems can include rotating a force vector reported by one of the pair of sensors 30 or 58 using Euler Conversion matrices to the lab space coordinate system of the video camera.
In one aspect, the person 14 or patient can be an amputee with a lower limb prosthesis (such as a prosthetic foot, leg, knee, and/or ankle) on one or both legs. Thus, in addition to temporal synchronization error, and orienting coordinate systems, it may be necessary to quickly and accurately adjust segment mass properties based both on visual blob segment distribution, and choice of prosthetic componentry. One such way would be to stand quietly on a single scale 150 while wearing a pressure/vertical force monitoring system or sensors 30 or 58. A separate method would be to manually enter the prosthetic component descriptions to utilize previously stored segment mass properties of known devices. This would enable both calibrating the kinetic sensors 30 and or 58, and realigning the mass distribution tables for accurate representation of the recorded individual. If a standard leg segment would represent 17% of body weight on average, not only does an amputated leg have a lower percentage of total body mass, all other segments have a slightly higher percentage to bring the total to 100%. For example, a 230-pound person would fit a standard anthropomorphic table (Table 1). If that same person became a left-knee disarticulation amputee subsequently weighing 211-pounds and who kept the same limb mass distribution, and was then fitted with a 2.5-pound prosthetic knee and a half-pound prosthetic foot, then that person could have a mass distribution as found in the second table (Table 2).
With reference to the tables, it will be realized that not only did the proportions of segment mass change on the limbs that were amputated, but every segment mass changed its proportion of the total when multiple segments were replaced with lighter componentry. Thus, the method can include adjusting 248 segment mass properties based on both visual blob segment distribution and choice of prosthesis. The adjusting can be done by realigning mass distribution between a standard mass distribution of a non-amputee and mass distribution of the amputee with the prosthesis.
As described above, the computer 26 and/or processors 28 thereof, and associated computer programs, can render the 2D video from the video cameras 18 as 3D stick figures. The method 100 can further include rendering 262 the video recording of the person walking into a three-dimensional stick figure (represented at 266) with a plurality of segments 270 representative of a skeletal model (represented at 266) of the person. As described above, the three-dimensional stick
The method can further include analyzing the kinetic data and the kinematic data, and performing a common Inverse Dynamics Analysis to yield the force and torque in each joint segment during the recording. The method 100 can include performing 288 Intersegmental Dynamics Analysis (IDA) to a computer model based on the kinematic, the kinetic and segment data by applying joint moments to determine subsequent acceleration of the computer model. In addition, the method 100 can include performing 302 Acceleration Analysis (AA) to compute a contribution of net joint moments to forward propulsion of the person. In addition, the method 100 can include performing 302 Acceleration Analysis (AA) to compute a contribution of net joint moments to vertical support of the person.
Inverse Dynamics (ID) can be more accurate when starting at the force plate and working up, not starting at the edges of the kinematic chain and working down. The aim of biomechanical analysis can be to know what the muscles are doing: the timing of their contractions, the amount of force generated (or moment of force about a joint), and the power of the contraction—whether it is concentric or eccentric. These quantities can be derived from the kinematics using the laws of physics, specifically the Newton-Euler equations:
Newton (linear): F=ma(Force=mass×linear acceleration)
Euler (angular): M=Ia(Moment=mass moment of inertia×angular acceleration)
These equations describe the behavior of a mathematical model of the limb called a link-segment model, and the process used to derive the joint moments at each joint is known as Inverse Dynamics, so-called because we work back from the kinematics to derive the kinetics responsible for the motion (as shown in
In looking at the human biomechanics, the effect of a joint moment through the kinetic chain cannot be determined simply from looking at a time history of joint moments. Conventional biomechanics, and specifically Inverse Dynamics, can indicate what total force and moment is generated at each joint, but cannot quantify the relationship between those loads and their effect on motion. The current limitation of Inverse Dynamics data, as calculated by gait labs, is the production of net, or summed, joint moments and reaction forces for each segment. These net outputs are the result of a sum of all active inputs, many of which cancel out in summation. This summation does not impart a full picture of what the neuromuscular control systems of the body are coordinating. To quantify this coupling among all movement generators in the model, meaning 3D moments at each joint, requires new and sophisticated analysis methodologies.
In order to understand what the neuro-muscular system must administer, instead of the summed result of activation commands, it is essential to quantify how each joint torque will accelerate all segments in the connected kinetic chain. This perspective can be clarified by using Intersegmental Dynamics Analysis (IDA), where joint forces and torques are applied and subsequent accelerations are calculated. The insight gained from performing acceleration analysis, and the reason it is worth the difficulty, is the comprehensive view of the multiple effectors that contribute to the motion of any one joint.
In one aspect, the system and method can objectively quantify how the leg muscles contribute to motion and how those strategies may differ when weakness or amputation is present. The strategies used to compensate for muscle weakness and optimize movement patterns cannot be objectively and accurately determined using conventional biomechanical analyses. Motion capture data are often used to assess these strategies, but lack correlation between recorded joint torques and motion they produce. An analyst must speculate on this relationship based on anatomic classification of muscles—which is known to be inaccurate. For example, the gastrocnemius attaches to the back of the knee joint but has been shown to cause knee extension in some cases. Acceleration Analysis can be used to provide a map of the effect of torque-spanning multiple joints during any activities for which a complete data set is captured.
To perform Intersegmental Dynamics Analysis, joint moments from a gait lab trial were then applied separately to a computer model which could determine the subsequent accelerations of the model that was posed based on the kinematic data. Each load produces a unique acceleration on every segment, for every time frame. The acceleration of every segment in the model due to an individual load is hereafter termed the component acceleration (CA) due to that load. The sum of all component accelerations should be equal to the observed acceleration of the body as seen in the gait lab.
Points of interaction with the environment, such as where the feet touch the ground, are replaced with joints. Such replacement is necessary to account for the interaction between the body and the environment. Both feet (and all other segments subsequently) are constrained in a manner consistent with actual segment acceleration during the patient's motion. A force is created at this joint by model acceleration, and this component ground reaction force is an important check on model validity.
Amputees are also capable of adapting to a large variety of geometric configurations and alignments of prosthetic components. There are many choices in prosthetic componentry, but few scientifically-based guidelines for their prescription. Optimal prosthetic alignment depends heavily on clinical judgment. Consequently, considerable variability and sub-optimality exists in amputee functional outcomes. Despite a wide prosthetic knowledge base (research publications, manuals and textbooks) there is scant information on the relationship between prosthetic alignment and amputee function.
Muscle forces acting at a joint accelerate all other segments in the body. Acceleration Analysis (AA) can quantify how much acceleration is produced by each muscle group at any joint in the body. Power redistribution between segments and muscles play a major role in body support and progression, and the precise quantification of the contribution of individual muscles to support and progression resolve many conflicting explanations of how gait occurs. Similarly, AA can help to objectively quantify compensatory mechanisms in amputee gait and in doing so, elucidate the alignment-function relationship.
Acceleration Analysis (AA) can be used to compute the contribution of net joint moments to forward propulsion. This is defined by how all individual joint moments work to accelerate the Center of Mass of the subject in translation in Global Lab Space. Vertical accelerations are defined to be those contributing to support, or resisting the gravity vector. Horizontal Accelerations are those that contribute to propulsion, either moving the body forward, or braking to slow the body down.
The method 100 can include saving 316 the kinetic data and the kinematic data together as a data file. In addition, the analysis performed on the kinetic and/or kinematic data can be saved together with the data. Furthermore, the method 100 can include displaying 330 a video recording of the person walking, and/or a rendering of the video recording of the person walking as a three-dimensional stick figure with a plurality of segments representative of a skeletal model of the person, along with data. The data can include the analysis from the IDA and/or AA overlaid with the video or rendering. The data can include the kinetic data measured by either of the sensors 30 or 58 overlaid with the video or rendering. Furthermore, the method 100 can include reporting (indicated at 338 and 340 in
A sit to stand motion can be used to test subject capability, and can be expanded into a test that involves standing from a seated position, and walking several paces, referred to as the Timed Up and Go (TUG) Test. Alternatively, an “L Test” involves multiple turns both to the left and to the right, and two transfers. The L Test starts with a seated subject, asks that they stand, walk two meters forward, turn 90 degrees to the left, walk three meters forward, turn 180 degrees, walk back the three meters just traversed, turn 90 degrees to the right, walk the two meters back to the chair, turn 180 degrees, and sit down. During this test, in order to capture all data the subject exerts on the environment, the scale that was previously used during initial calibration could be placed on the seat to capture the weight exerted while seated. This third data stream also provides data to quantify the timing of the event of “seat-off”, and can also be used to align the time stamp on the kinetic and kinematic data.
An example of a variable cadence report and/or analysis for an exemplary patient is shown in
For example, a sample report could contain an evaluation of the most common metric for amputee care, the ability to vary cadence. Ability to vary cadence is the leading determination for the level of care and devices that an amputee could be eligible for, and separates the amputee population into K0 through K4 levels. Cadence as a term, however, has not been well defined by the industry. One common quantification of cadence is the ability to vary the numbers of steps taken per minute, such as saying a person can walk on level ground between 66 steps per minute and 111 steps minute. In this case, the measured quantity has units of steps per minute (steps/min). Another way to look at varying cadence is to look at the change in the length of one stride to the next, or one step to the next. In this case, a physiatrist might want to quantify that a patient can take short steps in confined spaces, such as a kitchen, and also be able to cover ground while walking the mall. The units on this measure would be the variation on length of a step, or meters per step (meters/step). The last measure that can be determined as cadence is the ability to vary the amount of ground covered in a given amount of time, also known as walking speed. For this example, a physical therapist might want to know if the person can walk a comfortable self-selected speed, but also cross an intersection when a pedestrian light has changed. The units on this measure would be length per time, or meters/min.
As cadence is defined to be different measures for different medical professionals, the system can generate a report to display many different information sets. For example, the system and method of the present invention can produce a graph as shown in
An example of L Test data reduction and reporting for an exemplary patient is shown in
Detectable events can be seen on the 2D video feed, and displayed in isolation to the subject and practitioner after being identified by algorithm from the 3D skeleton data. 3D composite data can be played back with the skeleton model overlaid to show both the practitioner and the subject the significance of patterns found in the recorded gait. One such gait deviation that could be useful to coach an amputee subject is shoulder drop, common when the prosthetic limb length is too short and needs to be adjusted. Other such events that are discernible with software applications are Hip Hiking (common when a prosthesis is too long and the amputee rotates the pelvis in the frontal plane to gain ground clearance for the foot), crushing the heel (common when a prosthetic foot is too soft for a person who feels like he or she is stepping into a hole with each IC [Initial Contact]).
The reporting and saving can allow for reporting for auditing/payment measures. The system and method allows a patient report to be generated easily and quickly with many different outcome measures. Typical gait reporting for both sides could be seconds per step left and right as a bilateral column chart at fast, medium and slow walk, which would instantly show stride time/length symmetry. The system and method can report quantified measures of activities of daily living, capability of variable cadence, verification of beneficiary receipt, etc. For example, the Medicare proof of delivery must be a signed and dated delivery slip which includes: beneficiary's name, delivery address, sufficiently detailed description to identify the item(s) being delivered, quantity delivered, date delivered, the date of delivery as entered by the beneficiary, designee or the supplier, and the beneficiary (or designee) signature.
Furthermore, the method 100 can also include creating a patient history file comprising patient information (indicated at 344 in
The system and method provide several improvements of current available methodologies which are immediately apparent. One improvement is the ease of set up for the system, which enables capture techniques of a modern gait lab without the requirement of placing markers on the subject in anatomically relevant positions, or the training needed to do so. Another improvement is that the subject motion is unconstrained, such as with solutions in modern gait labs that mount force sensing plates in the floor, but demand that the subject walk precisely over those parts of the lab. Force plate systems inherently constrain the motion of the subject, and thus limit the activities that can be studied. Another improvement is that all the components of the system stay within the offices of the investigator, without the need to send parts home to monitor a subject, which raises subject compliance issues, identification of the person vs. the collected data to prevent fraud, or the possibility of lost or broken componentry when used outside of a standard and controllable environment when the complete system stays within the clinic. Parts are not lost or missing when it comes time to conduct a trial or measure the next patient.
In one aspect, the system or movement/gait lab can be provided as a kit with the video cameras and sensors together. In another aspect, the computer and scale can be provided as well. The various components (video cameras, sensors, computer and/or scale) can be provided together in a case to facilitate transportation/mobility, and protect the components during transport. The case can be provided with a handle, an extendable handle, and/or wheels to facilitate transport. In another aspect, one or more battery chargers can be provided. In another aspect, a template can be provided that can be disposed on the floor, or marked on the floor, to define the spatial volume and/or outline a path corresponding to various tests.
The computer and associated programs and/or applications can be operated by the practitioner, who then interacts with the subject. The computer applications can receive inputs, such as physical parameters specific to the subject, receive information via the sensors 30 or 58 worn by the subject, and use that information to calculate and quantify the actions of the subject for a particular physical activity. The patient information can include the forces collected with the sensors, and the physical position of the body and limb segments during activity. Such a computer can comprise at least one USB 3.0 data card for every two cameras, and may comprise at least one active USB extension cable per camera. The system can include one or more computers 26. While shown as a single stand-alone computer, the computer can be more than one computer, distributed locally or remotely to one another. The computer can store algorithms for performing comparisons of the information collected from the patient to model information. The applications running the one or more computers may be described in the context of computer-executable instructions, such as program modules being executed by the host computer. Generally described, program modules can include routines, programs, applications, objects, components, data structures, and the like that perform tasks or implement particular abstract data types.
The applications can be stored on the computer 26 or in a remote server to which the computer is communicatively linked to, such as through the Internet. In one embodiment, for example, one computer may be local or a remote server. Furthermore, the functions performed by the computer may be distributed or shared among more than one computers that are all communicatively linked, such as via local area networks (LAN), wide area networks (WAN), and/or the Internet. Any communication over the Internet or other network may be encrypted for security.
The computer may be any one of a variety of devices including, but not limited to, personal computing devices, server-based computing devices, mini- and mainframe computers, laptops, or other electronic devices having some type of memory. More than one computer can be used in place of a single computer. The computer 26 can include a processor, a memory, a computer-readable medium drive (e.g., disk drive, a hard drive, CD-ROM/DVD-ROM, etc.), that are all communicatively connected to each other by a communication bus. The memory can comprises Random Access Memory (“RAM”), Read-Only Memory (“ROM”), flash memory, and the like. The host computer can include a display and one or more user input devices, such as a mouse, keyboard, etc.
The memory can store program code that provides a gait analysis application and a step detection application. The gait analysis application can include computer-executable instructions that, when executed by the processor, applies an algorithm to receive, display, and process input, including moment and force data. The step and phase detection application can apply an algorithm to a set of moment and axial force data to differentiate each step. Further, the step and phase detection application can establish if the subject is either in stance or swing phase of a gait cycle.
The computer application can enable all the quantification reporting herein.
While the forgoing examples are illustrative of the principles of the present invention in one or more particular applications, it will be apparent to those of ordinary skill in the art that numerous modifications in form, usage and details of implementation can be made without the exercise of inventive faculty, and without departing from the principles and concepts of the invention. Accordingly, it is not intended that the invention be limited, except as by the claims set forth below.
Number | Name | Date | Kind |
---|---|---|---|
3973332 | Slomski | Aug 1976 | A |
4112463 | Kamin | Sep 1978 | A |
4128316 | Zanoni | Dec 1978 | A |
4136387 | Sullivan et al. | Jan 1979 | A |
4137566 | Haas et al. | Jan 1979 | A |
4146924 | Birk et al. | Mar 1979 | A |
4163941 | Linn | Aug 1979 | A |
4172661 | Marcus et al. | Oct 1979 | A |
4197855 | Lewin | Apr 1980 | A |
4236180 | Cayzac | Nov 1980 | A |
4238828 | Hay et al. | Dec 1980 | A |
4281342 | Ueda et al. | Jul 1981 | A |
4307608 | Useldinger et al. | Dec 1981 | A |
4375674 | Thornton | Mar 1983 | A |
4416293 | Anderson et al. | Nov 1983 | A |
4667513 | Konno | May 1987 | A |
4813436 | Au | Mar 1989 | A |
4836033 | Seitz | Jun 1989 | A |
4843568 | Krueger et al. | Jun 1989 | A |
4882677 | Curran | Nov 1989 | A |
4971069 | Gracovetsky | Nov 1990 | A |
5033291 | Podoloff et al. | Jul 1991 | A |
5086404 | Claussen | Feb 1992 | A |
5203346 | Fuhr et al. | Apr 1993 | A |
5209240 | Jain et al. | May 1993 | A |
5337757 | Jain et al. | Aug 1994 | A |
5408873 | Schmidt et al. | Apr 1995 | A |
5441047 | David et al. | Aug 1995 | A |
5462065 | Cusimano | Oct 1995 | A |
5482048 | Johnson | Jan 1996 | A |
5490221 | Ransford et al. | Feb 1996 | A |
5524645 | Wills | Jun 1996 | A |
5657426 | Waters et al. | Aug 1997 | A |
5748199 | Palm | May 1998 | A |
5802220 | Black et al. | Sep 1998 | A |
5808540 | Wheeler et al. | Sep 1998 | A |
5818959 | Webb et al. | Oct 1998 | A |
5852669 | Eleftheriadis et al. | Dec 1998 | A |
5855818 | Gan et al. | Jan 1999 | A |
5987154 | Gibbon et al. | Nov 1999 | A |
5989700 | Krivopal | Nov 1999 | A |
6014133 | Yamakado et al. | Jan 2000 | A |
6032542 | Warnick et al. | Mar 2000 | A |
6047078 | Kang | Apr 2000 | A |
6056671 | Marmer | May 2000 | A |
6081273 | Weng et al. | Jun 2000 | A |
6122846 | Gray et al. | Sep 2000 | A |
6195921 | Truong | Mar 2001 | B1 |
6216545 | Taylor | Apr 2001 | B1 |
6231527 | Sol | May 2001 | B1 |
6301964 | Fyfe et al. | Oct 2001 | B1 |
6514219 | Guimond et al. | Feb 2003 | B1 |
6543299 | Taylor | Apr 2003 | B2 |
6552729 | Di Bernardo et al. | Apr 2003 | B1 |
6580812 | Harrington | Jun 2003 | B1 |
6674877 | Jojic et al. | Jan 2004 | B1 |
6697502 | Luo | Feb 2004 | B2 |
6744914 | Rubbert et al. | Jun 2004 | B1 |
6821257 | Jolley | Nov 2004 | B1 |
6836744 | Asphahani et al. | Dec 2004 | B1 |
6904801 | Bridges et al. | Jun 2005 | B1 |
6931938 | Knirck et al. | Aug 2005 | B2 |
7006683 | Brand | Feb 2006 | B2 |
7027642 | Rubbert et al. | Apr 2006 | B2 |
7074168 | Farnes et al. | Jul 2006 | B1 |
7156792 | Gibson-Horn | Jan 2007 | B2 |
7179234 | Nashner | Feb 2007 | B2 |
7197179 | Rubbert et al. | Mar 2007 | B2 |
7201063 | Taylor | Apr 2007 | B2 |
7219033 | Kolen | May 2007 | B2 |
7239718 | Park et al. | Jul 2007 | B2 |
7317836 | Fujimura et al. | Jan 2008 | B2 |
7379584 | Rubbert et al. | May 2008 | B2 |
7420472 | Tran | Sep 2008 | B2 |
7421101 | Georgescu et al. | Sep 2008 | B2 |
7447337 | Zhang et al. | Nov 2008 | B2 |
7457457 | Ives et al. | Nov 2008 | B2 |
7492268 | Ferguson et al. | Feb 2009 | B2 |
7502498 | Wen et al. | Mar 2009 | B2 |
7526071 | Drapeau | Apr 2009 | B2 |
7532744 | Sharoni et al. | May 2009 | B2 |
7539532 | Tran | May 2009 | B2 |
7539533 | Tran | May 2009 | B2 |
7558622 | Tran | Jul 2009 | B2 |
7623944 | Dariush | Nov 2009 | B2 |
7628074 | Vannucci et al. | Dec 2009 | B2 |
7657084 | Ives et al. | Feb 2010 | B2 |
7684896 | Dariush | Mar 2010 | B2 |
7706602 | Nakashima | Apr 2010 | B2 |
7708673 | Gibson-Horn | May 2010 | B2 |
7716005 | Shoureshi et al. | May 2010 | B2 |
7770473 | Von Lilienfeld-Toal et al. | Aug 2010 | B2 |
7794370 | Tackett | Sep 2010 | B2 |
7857771 | Alwan et al. | Dec 2010 | B2 |
7907774 | Parr et al. | Mar 2011 | B2 |
7980141 | Connor et al. | Jul 2011 | B2 |
7980856 | Grabiner et al. | Jul 2011 | B2 |
8102390 | Tapang | Jan 2012 | B2 |
8111165 | Ortega et al. | Feb 2012 | B2 |
8122772 | Clausen | Feb 2012 | B2 |
8126736 | Anderson et al. | Feb 2012 | B2 |
8139067 | Anguelov et al. | Mar 2012 | B2 |
8139822 | Selner | Mar 2012 | B2 |
8165844 | Luinge et al. | Apr 2012 | B2 |
8180714 | Corazza et al. | May 2012 | B2 |
8215186 | Macomber et al. | Jul 2012 | B2 |
8215773 | Gibson-Horn et al. | Jul 2012 | B2 |
8265730 | Alexander | Sep 2012 | B2 |
8323188 | Tran | Dec 2012 | B2 |
8323189 | Tran et al. | Dec 2012 | B2 |
8328718 | Tran | Dec 2012 | B2 |
8393229 | Tao et al. | Mar 2013 | B2 |
8409297 | Boone et al. | Apr 2013 | B2 |
8421642 | McIntosh | Apr 2013 | B1 |
8425415 | Tran | Apr 2013 | B2 |
8449471 | Tran | May 2013 | B2 |
8452080 | Engedal | May 2013 | B2 |
8461988 | Tran | Jun 2013 | B2 |
8467574 | Mathe et al. | Jun 2013 | B2 |
8471848 | Tschesnok | Jun 2013 | B2 |
8475368 | Tran et al. | Jul 2013 | B2 |
8500636 | Tran | Aug 2013 | B2 |
8503720 | Shotton et al. | Aug 2013 | B2 |
8525673 | Tran | Sep 2013 | B2 |
8525687 | Tran | Sep 2013 | B2 |
8527217 | Moodie | Sep 2013 | B2 |
8531291 | Tran | Sep 2013 | B2 |
8551026 | Alwan et al. | Oct 2013 | B2 |
8565479 | Gurman et al. | Oct 2013 | B2 |
8622747 | Chu et al. | Jan 2014 | B2 |
8652038 | Tran et al. | Feb 2014 | B2 |
8659596 | Corazza et al. | Feb 2014 | B2 |
8661915 | Taylor | Mar 2014 | B2 |
8668386 | Morton et al. | Mar 2014 | B2 |
8684900 | Tran | Apr 2014 | B2 |
8684922 | Tran | Apr 2014 | B2 |
8685093 | Anderson et al. | Apr 2014 | B2 |
8687765 | Kotowski et al. | Apr 2014 | B2 |
8698888 | Bonnet | Apr 2014 | B2 |
8708903 | Tran | Apr 2014 | B2 |
8727978 | Tran et al. | May 2014 | B2 |
8737767 | Hodgins et al. | May 2014 | B2 |
8747313 | Tran et al. | Jun 2014 | B2 |
8747336 | Tran | Jun 2014 | B2 |
8749115 | Pas et al. | Jun 2014 | B2 |
8750971 | Tran | Jun 2014 | B2 |
8758272 | Kubiak et al. | Jun 2014 | B2 |
8758273 | Kubiak et al. | Jun 2014 | B2 |
8764651 | Tran | Jul 2014 | B2 |
8784502 | Macomber et al. | Jul 2014 | B2 |
8800386 | Taylor | Aug 2014 | B2 |
8845494 | Whitall et al. | Sep 2014 | B2 |
8870795 | Kim et al. | Oct 2014 | B2 |
8904876 | Taylor et al. | Dec 2014 | B2 |
9140717 | Perkins | Sep 2015 | B2 |
20030164829 | Bregler et al. | Sep 2003 | A1 |
20030169907 | Edwards et al. | Sep 2003 | A1 |
20030208116 | Liang et al. | Nov 2003 | A1 |
20030215130 | Nakamura et al. | Nov 2003 | A1 |
20040021660 | Ng-Thow-Hing et al. | Feb 2004 | A1 |
20040155962 | Marks | Aug 2004 | A1 |
20040186358 | Chernow | Sep 2004 | A1 |
20040259690 | Frykman et al. | Dec 2004 | A1 |
20050000298 | Pfeifer et al. | Jan 2005 | A1 |
20050078178 | Brown et al. | Apr 2005 | A1 |
20050184878 | Grold et al. | Aug 2005 | A1 |
20050265583 | Covell | Dec 2005 | A1 |
20060245618 | Boregowda et al. | Nov 2006 | A1 |
20060267978 | Litke et al. | Nov 2006 | A1 |
20070104351 | Yang et al. | May 2007 | A1 |
20070204687 | Haselhurst et al. | Sep 2007 | A1 |
20070250287 | Spector | Oct 2007 | A1 |
20070275830 | Lee | Nov 2007 | A1 |
20080191864 | Wolfson | Aug 2008 | A1 |
20080221487 | Zohar | Sep 2008 | A1 |
20090031825 | Kishida et al. | Feb 2009 | A1 |
20090213123 | Crow | Aug 2009 | A1 |
20090252423 | Zhu et al. | Oct 2009 | A1 |
20100023345 | Schottlander | Jan 2010 | A1 |
20100076563 | Otto | Mar 2010 | A1 |
20100162832 | Brauers | Jul 2010 | A1 |
20100278405 | Kakadiaris et al. | Nov 2010 | A1 |
20110004126 | Einav et al. | Jan 2011 | A1 |
20110112397 | Shen | May 2011 | A1 |
20110221457 | Takahashi et al. | Sep 2011 | A1 |
20110249865 | Lee et al. | Oct 2011 | A1 |
20110288448 | Sanders et al. | Nov 2011 | A1 |
20110292036 | Sali et al. | Dec 2011 | A1 |
20120004579 | Luo | Jan 2012 | A1 |
20120119904 | Coleman Boone et al. | May 2012 | A1 |
20120137795 | Selner | Jun 2012 | A1 |
20120191017 | Huang | Jul 2012 | A1 |
20120253234 | Yang et al. | Oct 2012 | A1 |
20120255160 | Boone et al. | Oct 2012 | A1 |
20120274442 | Mottram | Nov 2012 | A1 |
20120276999 | Kalpaxis et al. | Nov 2012 | A1 |
20130123665 | Mariani et al. | May 2013 | A1 |
20130123667 | Komatireddy et al. | May 2013 | A1 |
20130190135 | Pryor | Jul 2013 | A1 |
20140039657 | Spector | Feb 2014 | A1 |
20140090488 | Taylor et al. | Apr 2014 | A1 |
20140090489 | Taylor et al. | Apr 2014 | A1 |
20140182391 | Huang et al. | Jul 2014 | A1 |
20140230563 | Huang | Aug 2014 | A1 |
20140296997 | Herr | Oct 2014 | A1 |
20140343460 | Evans, III et al. | Nov 2014 | A1 |
20150038878 | Pusch | Feb 2015 | A1 |
20150213193 | Apte | Jul 2015 | A1 |
Number | Date | Country |
---|---|---|
101341515 | Oct 2011 | CN |
102460134 | Jan 2014 | CN |
10-201001789 | Oct 2011 | DE |
0742597 | Nov 1996 | EP |
1714612 | Oct 2006 | EP |
1935379 | Jun 2008 | EP |
2042847 | Apr 2009 | EP |
2484281 | Aug 2012 | EP |
2674913 | Jul 2014 | EP |
2776175 | Sep 1999 | FR |
2003337071 | Nov 2003 | JP |
2010122018 | Jun 2010 | JP |
WO 9415530 | Jul 1994 | WO |
WO 2004008095 | Jan 2004 | WO |
WO 2004066821 | Aug 2004 | WO |
WO 2004092744 | Oct 2004 | WO |
WO 2005003902 | Jan 2005 | WO |
WO 2006078538 | Jul 2006 | WO |
WO 2006078566 | Jul 2006 | WO |
WO 2006119186 | Nov 2006 | WO |
WO 2007072246 | Jun 2007 | WO |
WO 2008042903 | Apr 2008 | WO |
WO 2008102308 | Aug 2008 | WO |
WO 2008103181 | Aug 2008 | WO |
WO 2008109567 | Nov 2008 | WO |
WO 2009070782 | Jun 2009 | WO |
WO 2009089204 | Jul 2009 | WO |
WO 2009120270 | Oct 2009 | WO |
WO 2010135620 | Nov 2010 | WO |
WO 2011026001 | Mar 2011 | WO |
WO 2011053839 | May 2011 | WO |
WO 2012014714 | Feb 2012 | WO |
WO 2012060578 | May 2012 | WO |
WO 2012171967 | Dec 2012 | WO |
WO 2013022890 | Feb 2013 | WO |
WO 2013132129 | Sep 2013 | WO |
WO 2013186010 | Dec 2013 | WO |
WO 2014068269 | May 2014 | WO |
WO 2014100045 | Jun 2014 | WO |
WO 2014135187 | Sep 2014 | WO |
Entry |
---|
Alexander et al: “Effect of Synchronization Error Between Force and Position Data in Moment Calculations” (Gait & Posture, vol. 5, Issue 2, Apr. 1997, p. 177). |
Ayyappa; “Normal Human Locomotion, Part 1: Basic Concepts and Terminology.” Journal of Prosthetics and Orthotics (1997); vol. 9, No. 1; 10 pages. |
Chesnin et al.: “Comparison of an in-shoe Pressure Measurement Device to a Force Plate: CoP Measurement Validity;” Gait & Posture, vol. 12, Issue 2, Oct. 1, 2000, pp. 128-133. |
Ganeriwal et al.; “Timing-Sync Protocol for Sensor Networks;” The First ACM Conference on Embedded Networked Sensor Systems (SenSys), University of California (2003); p. 138-149 (12 pages). |
Kepple et al.: “Relative Contributions of the Lower Extremity Joint Moments Forward Progression and Support During Gait;” Gait & Posture 6 (1997); pp. 1-8. |
Martin et al.: “Volumetric Descriptions of Objects from Multiple Views,” IEEE Transactions on Pattern Analysis and Machine Intelligence, Mar. 1983; vol. Pami-5, No. 2; pp. 150-158. |
Mori et al.: “Insole-Type Simultaneous Measurement System of Plantar Pressure and Shear Force During Gait for Diabetic Patients;” Journal of Robotics and Mechatronics, vol. 24, No. 5; University of Toyko, 2012; pp. 766-772. |
Neptune et al.; “Muscle force redistributes segmental power for body progression during walking;” Gait & Posture (19), 2004; pp. 194-205. |
Niem: “Robust and Fast Modeling of 3D Natural Objects from Multiple Views;” Proceedings ‘Image and Video Processing II’; 2004, 10 pages. |
Riley et al.: “Propulsive adaptation to changing gait speed;” Journal of Biomechanics 34 (2001); pp. 197-202. |
Roche; “Time Synchronization in Wireless Networks;” Washington University in St. Louis; 2006; 12 pages; http://www.cs.wustl.edu/˜jain/cse57406/ftp/time—sync/index.html. |
Samson et al; “A Portable System for Foot Biomechanical Analysis During Gait;” Gait & Posture vol. 40 (2014); pp. 420-428. |
Siegel: “Using induced accelerations to understand knee stability during gait of individuals with muscle weakness;” Gait & Posture 23 (2006); pp. 435-440. |
Steele et al.: “Acceleration Analysis to Quantify Compensation in Amputee Gait: The Effect of Prosthetic Foot Position in a Trans-Tibial Ampute;” Mar. 2007 Procedings of American Academy of Orthotists and Prosthetists. |
Talaty: “Computer Modeling and Simulation in Gait Analysis;” PM&R: State of the Art Reviews; (2002). |
Van Den Bogert et al.; “A Weighted Least-Squares Method for Inverse Dynamic Analysis;” Computer Methods in Biomechanics and Biomedical Engineering, vol. 11, No. 1; pp. 3-9; MSL Academic Endeavors (2008). |
Wassom et al: “Vertical Jump Measurement Techniques Using Markerless 3d Motion Capture and Inertial Sensors” Medicine and Science in Sports and Exercise vol. 45. No. 5. (2013). |
Zajac; “Determining Muscle's Force and Action in Multi-Articular Movement;” Exercise in Sport and Science Reviews (1989); vol. 17, pp. 187-230. |
Number | Date | Country | |
---|---|---|---|
20160370854 A1 | Dec 2016 | US |